Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/logs/pitr-pxb-8-4.log Warning: version difference between client (1.36) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.36) and server (1.32) exceeds the supported minor version skew of +/-1 + PXCO_FEATURE_GATES=XtrabackupSidecar=true + GTID_PATTERN='[A-F0-9a-f]{8}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{12}:[0-9]+' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + main + create_infra pitr-pxb-25367 + local ns=pitr-pxb-25367 + '[' -n pxc-operator ']' + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl get pxc --all-namespaces -o wide + kubectl patch pxc -n pitr-pxb-7501 pitr --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/pitr patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.yDY6owtlCK ++ mktemp + local LAST_ERR=/tmp/tmp.FsBQ4DkcXC + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.yDY6owtlCK perconaxtradbcluster.pxc.percona.com "pitr" deleted from pitr-pxb-7501 namespace + cat /tmp/tmp.FsBQ4DkcXC + rm /tmp/tmp.yDY6owtlCK /tmp/tmp.FsBQ4DkcXC + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.efUnncCyje ++ mktemp + local LAST_ERR=/tmp/tmp.W9CQasQrna + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.efUnncCyje perconaxtradbclusterbackup.pxc.percona.com "on-pitr-minio" deleted from pitr-pxb-7501 namespace perconaxtradbclusterbackup.pxc.percona.com "on-pitr-minio-2" deleted from pitr-pxb-7501 namespace perconaxtradbclusterbackup.pxc.percona.com "on-pitr-minio-3" deleted from pitr-pxb-7501 namespace + cat /tmp/tmp.W9CQasQrna + rm /tmp/tmp.efUnncCyje /tmp/tmp.W9CQasQrna + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.UROzhbLpZ5 ++ mktemp + local LAST_ERR=/tmp/tmp.Nlp7qYmAWX + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.UROzhbLpZ5 perconaxtradbclusterrestore.pxc.percona.com "restore-on-pitr-minio" deleted from pitr-pxb-7501 namespace + cat /tmp/tmp.Nlp7qYmAWX + rm /tmp/tmp.UROzhbLpZ5 /tmp/tmp.Nlp7qYmAWX + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ sed s/NAMESPACE// ++ awk '-F ' '{print $2}' ++ tail -n1 ++ helm list --all-namespaces --filter chaos-mesh + local chaos_mesh_ns= + '[' -n '' ']' ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get MutatingWebhookConfiguration + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl get clusterrolebinding ++ grep chaos-mesh + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get clusterrole + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + '[' -n '' ']' + desc 'cleaned up old namespaces pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pxc-operator + grep -E -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + xargs kubectl delete ns + kubectl_bin get ns ++ mktemp + local LAST_OUT=/tmp/tmp.vYiF6QAOxw ++ mktemp + local LAST_ERR=/tmp/tmp.2iCOGNXvQO + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + awk '{print$1}' ++ mktemp + local LAST_OUT=/tmp/tmp.AlIjVtQmj1 ++ mktemp + local LAST_ERR=/tmp/tmp.h9DobKhUw1 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.vYiF6QAOxw + cat /tmp/tmp.2iCOGNXvQO + rm /tmp/tmp.vYiF6QAOxw /tmp/tmp.2iCOGNXvQO + return 0 namespace "cert-manager" deleted namespace "pitr-pxb-7501" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.AlIjVtQmj1 namespace "pxc-operator" deleted + cat /tmp/tmp.h9DobKhUw1 + rm /tmp/tmp.AlIjVtQmj1 /tmp/tmp.h9DobKhUw1 + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + echo -n 'waiting for namespace/pxc-operator to be deleted' waiting for namespace/pxc-operator to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pxc-operator" not found + desc 'create namespace pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.frqVzQQjg8 ++ mktemp + local LAST_ERR=/tmp/tmp.m79IlK9kST + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.frqVzQQjg8 namespace/pxc-operator created + cat /tmp/tmp.m79IlK9kST + rm /tmp/tmp.frqVzQQjg8 /tmp/tmp.m79IlK9kST + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.qQqNbjAGM5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.0PoXNYy74W ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qQqNbjAGM5 ++ cat /tmp/tmp.0PoXNYy74W ++ rm /tmp/tmp.qQqNbjAGM5 /tmp/tmp.0PoXNYy74W ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2433-9bec550e-7-cluster12 --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.0MVhNOX3JR ++ mktemp + local LAST_ERR=/tmp/tmp.1iiUjYLMul + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2433-9bec550e-7-cluster12 --namespace=pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.0MVhNOX3JR Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2433-9bec550e-7-cluster12" modified. + cat /tmp/tmp.1iiUjYLMul + rm /tmp/tmp.0MVhNOX3JR /tmp/tmp.1iiUjYLMul + return 0 + deploy_operator + desc 'start PXC operator' + set +o xtrace ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.JfiQhsN8gV ++ mktemp + local LAST_ERR=/tmp/tmp.Dv0ENwfWw7 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.JfiQhsN8gV customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied + cat /tmp/tmp.Dv0ENwfWw7 + rm /tmp/tmp.JfiQhsN8gV /tmp/tmp.Dv0ENwfWw7 + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + sed -e 's^namespace: .*^namespace: pxc-operator^' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/deploy/cw-rbac.yaml + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.SckKCcAUvr ++ mktemp + local LAST_ERR=/tmp/tmp.8hVzJ4ZZZW + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.SckKCcAUvr clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.8hVzJ4ZZZW + rm /tmp/tmp.SckKCcAUvr /tmp/tmp.8hVzJ4ZZZW + return 0 + kubectl_bin apply -f - + sed -e 's^failureThreshold: .*^failureThreshold: 10^' + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-2433-9bec550e^' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/deploy/cw-operator.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.M1q6EkMxeY ++ mktemp + local LAST_ERR=/tmp/tmp.ZZ7wIesMBS + local exit_status=0 ++ seq 0 2 + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "DISABLE_TELEMETRY").value) = "true"' - + for i in '$(seq 0 2)' + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "LOG_LEVEL").value) = "VERBOSE"' - + set +e + kubectl apply -f - + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "PXCO_FEATURE_GATES").value) = "XtrabackupSidecar=true"' - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.M1q6EkMxeY deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.ZZ7wIesMBS + rm /tmp/tmp.M1q6EkMxeY /tmp/tmp.ZZ7wIesMBS + return 0 + sleep 10 + kubectl_bin wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s ++ mktemp + local LAST_OUT=/tmp/tmp.IsC4PBJqYv ++ mktemp + local LAST_ERR=/tmp/tmp.bF6ln41pvn + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.IsC4PBJqYv pod/percona-xtradb-cluster-operator-6fc4ff78b7-96wkz condition met + cat /tmp/tmp.bF6ln41pvn + rm /tmp/tmp.IsC4PBJqYv /tmp/tmp.bF6ln41pvn + return 0 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator --field-selector=status.phase=Running -o json -n pxc-operator +++ mktemp ++ head -1 ++ jq -r '.items[] | select(.metadata.deletionTimestamp == null) | .metadata.name' ++ local LAST_OUT=/tmp/tmp.rgCwzNdCkq +++ mktemp ++ local LAST_ERR=/tmp/tmp.JTf6WJt4kM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator --field-selector=status.phase=Running -o json -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.rgCwzNdCkq ++ cat /tmp/tmp.JTf6WJt4kM ++ rm /tmp/tmp.rgCwzNdCkq /tmp/tmp.JTf6WJt4kM ++ return 0 + wait_pod percona-xtradb-cluster-operator-6fc4ff78b7-96wkz 480 pxc-operator + local pod=percona-xtradb-cluster-operator-6fc4ff78b7-96wkz + local max_retry=480 + local ns=pxc-operator ++ echo percona-xtradb-cluster-operator-6fc4ff78b7-96wkz ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-6fc4ff78b7-96wkz condition met waiting for pod/percona-xtradb-cluster-operator-6fc4ff78b7-96wkz to become Ready.Ok + sleep 3 + create_namespace pitr-pxb-25367 + local namespace=pitr-pxb-25367 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ awk '{print $1}' ++ grep chaos-mesh.org + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + '[' -n '' ']' + desc 'cleaned up old namespaces pitr-pxb-25367' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pitr-pxb-25367 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pitr-pxb-25367 + awk '{print$1}' ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.rpqCRxxLLK + local LAST_OUT=/tmp/tmp.QCj29ncXew ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.DsJPDRQRQn + local exit_status=0 + local LAST_ERR=/tmp/tmp.RTofnzcXVL + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pitr-pxb-25367 + xargs kubectl delete ns + grep -E -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pitr-pxb-25367 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.rpqCRxxLLK + cat /tmp/tmp.DsJPDRQRQn + rm /tmp/tmp.rpqCRxxLLK /tmp/tmp.DsJPDRQRQn + return 0 error: resource(s) were provided, but no name was specified + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pitr-pxb-25367 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.QCj29ncXew + cat /tmp/tmp.RTofnzcXVL Error from server (NotFound): namespaces "pitr-pxb-25367" not found + rm /tmp/tmp.QCj29ncXew /tmp/tmp.RTofnzcXVL + return 1 + : + wait_for_delete namespace/pitr-pxb-25367 + local res=namespace/pitr-pxb-25367 + echo -n 'waiting for namespace/pitr-pxb-25367 to be deleted' waiting for namespace/pitr-pxb-25367 to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pitr-pxb-25367" not found + desc 'create namespace pitr-pxb-25367' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pitr-pxb-25367 ----------------------------------------------------------------------------------- + kubectl_bin create namespace pitr-pxb-25367 ++ mktemp + local LAST_OUT=/tmp/tmp.cjXXHrTlI8 ++ mktemp + local LAST_ERR=/tmp/tmp.9wtGBl3at1 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pitr-pxb-25367 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.cjXXHrTlI8 namespace/pitr-pxb-25367 created + cat /tmp/tmp.9wtGBl3at1 + rm /tmp/tmp.cjXXHrTlI8 /tmp/tmp.9wtGBl3at1 + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.5rn0mHzX38 +++ mktemp ++ local LAST_ERR=/tmp/tmp.VooluiaaDz ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.5rn0mHzX38 ++ cat /tmp/tmp.VooluiaaDz ++ rm /tmp/tmp.5rn0mHzX38 /tmp/tmp.VooluiaaDz ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2433-9bec550e-7-cluster12 --namespace=pitr-pxb-25367 ++ mktemp + local LAST_OUT=/tmp/tmp.LXweB7cQcs ++ mktemp + local LAST_ERR=/tmp/tmp.M7d51uckwU + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2433-9bec550e-7-cluster12 --namespace=pitr-pxb-25367 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.LXweB7cQcs Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2433-9bec550e-7-cluster12" modified. + cat /tmp/tmp.M7d51uckwU + rm /tmp/tmp.LXweB7cQcs /tmp/tmp.M7d51uckwU + return 0 + apply_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.WUpP5cXoOD ++ mktemp + local LAST_ERR=/tmp/tmp.ULdVU9btqP + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.WUpP5cXoOD secret/minio-secret created secret/aws-s3-secret created secret/do-spaces-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.ULdVU9btqP + rm /tmp/tmp.WUpP5cXoOD /tmp/tmp.ULdVU9btqP + return 0 + deploy_cert_manager + desc 'deploy cert manager' + set +o xtrace ----------------------------------------------------------------------------------- deploy cert manager ----------------------------------------------------------------------------------- + kubectl_bin create namespace cert-manager ++ mktemp + local LAST_OUT=/tmp/tmp.3QsnbO6Vuq ++ mktemp + local LAST_ERR=/tmp/tmp.lIPsJPzZ72 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace cert-manager + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.3QsnbO6Vuq namespace/cert-manager created + cat /tmp/tmp.lIPsJPzZ72 + rm /tmp/tmp.3QsnbO6Vuq /tmp/tmp.lIPsJPzZ72 + return 0 + kubectl_bin label namespace cert-manager certmanager.k8s.io/disable-validation=true ++ mktemp + local LAST_OUT=/tmp/tmp.qwGc6EpbWp ++ mktemp + local LAST_ERR=/tmp/tmp.8DhiKdWSO9 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl label namespace cert-manager certmanager.k8s.io/disable-validation=true + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.qwGc6EpbWp namespace/cert-manager labeled + cat /tmp/tmp.8DhiKdWSO9 + rm /tmp/tmp.qwGc6EpbWp /tmp/tmp.8DhiKdWSO9 + return 0 + kubectl_bin apply -f https://github.com/jetstack/cert-manager/releases/download/v1.19.2/cert-manager.yaml --validate=false ++ mktemp + local LAST_OUT=/tmp/tmp.Vw3L7Momr1 ++ mktemp + local LAST_ERR=/tmp/tmp.PItFQbL8bN + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.19.2/cert-manager.yaml --validate=false + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Vw3L7Momr1 namespace/cert-manager configured customresourcedefinition.apiextensions.k8s.io/challenges.acme.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/orders.acme.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/certificaterequests.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/certificates.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/clusterissuers.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/issuers.cert-manager.io unchanged serviceaccount/cert-manager-cainjector created serviceaccount/cert-manager created serviceaccount/cert-manager-webhook created clusterrole.rbac.authorization.k8s.io/cert-manager-cainjector unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-issuers unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificates unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-orders unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-challenges unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-cluster-view unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-view unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-edit unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-cainjector unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-issuers unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificates unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-orders unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-challenges unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews unchanged role.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection unchanged role.rbac.authorization.k8s.io/cert-manager:leaderelection unchanged role.rbac.authorization.k8s.io/cert-manager-tokenrequest created role.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created rolebinding.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection unchanged rolebinding.rbac.authorization.k8s.io/cert-manager:leaderelection unchanged rolebinding.rbac.authorization.k8s.io/cert-manager-tokenrequest created rolebinding.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created service/cert-manager-cainjector created service/cert-manager created service/cert-manager-webhook created deployment.apps/cert-manager-cainjector created deployment.apps/cert-manager created deployment.apps/cert-manager-webhook created mutatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook configured validatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook configured + cat /tmp/tmp.PItFQbL8bN Warning: resource namespaces/cert-manager is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. + rm /tmp/tmp.Vw3L7Momr1 /tmp/tmp.PItFQbL8bN + return 0 + '[' '' == 4.10 ']' + sleep 70 + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/issuer.yml ++ mktemp + local LAST_OUT=/tmp/tmp.NbDXQlsY7p ++ mktemp + local LAST_ERR=/tmp/tmp.O0tWcDX0Ah + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/issuer.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.NbDXQlsY7p issuer.cert-manager.io/selfsigning-issuer created + cat /tmp/tmp.O0tWcDX0Ah + rm /tmp/tmp.NbDXQlsY7p /tmp/tmp.O0tWcDX0Ah + return 0 + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/cert.yml ++ mktemp + local LAST_OUT=/tmp/tmp.rbAtlMZqUS ++ mktemp + local LAST_ERR=/tmp/tmp.lpJ8BXDhYn + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/cert.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.rbAtlMZqUS certificate.cert-manager.io/tls-minio created + cat /tmp/tmp.lpJ8BXDhYn Warning: spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. + rm /tmp/tmp.rbAtlMZqUS /tmp/tmp.lpJ8BXDhYn + return 0 + sleep_with_log 25 + local d=25 + log 'sleeping for 25 seconds' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T16:51:09+0000]' sleeping for 25 seconds [2026-04-29T16:51:09+0000] sleeping for 25 seconds + sleep 25 + start_minio tls-minio + deploy_helm pitr-pxb-25367 + helm repo add hashicorp https://helm.releases.hashicorp.com "hashicorp" already exists with the same configuration, skipping + helm repo add minio https://charts.min.io/ "minio" already exists with the same configuration, skipping + helm repo update Hang tight while we grab the latest from your chart repositories... ...Successfully got an update from the "chaos-mesh" chart repository ...Successfully got an update from the "minio" chart repository ...Successfully got an update from the "percona" chart repository ...Successfully got an update from the "hashicorp" chart repository Update Complete. ⎈Happy Helming!⎈ + local cert_secret=tls-minio + local endpoint=http://minio-service:9000 + minio_args=(--version $MINIO_VER --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set "users[0].accessKey=some-access-key" --set "users[0].secretKey=some-secret-key" --set "users[0].policy=consoleAdmin" --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G) + local minio_args + [[ -n tls-minio ]] + endpoint=https://minio-service:9000 + minio_args+=(--set tls.enabled=true --set tls.certSecret="$cert_secret" --set tls.publicCrt=tls.crt --set tls.privateKey=tls.key) + desc 'install Minio' + set +o xtrace ----------------------------------------------------------------------------------- install Minio ----------------------------------------------------------------------------------- + helm uninstall minio-service Error: uninstall: Release not loaded: minio-service: release: not found + : + retry 10 60 helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G --set tls.enabled=true --set tls.certSecret=tls-minio --set tls.publicCrt=tls.crt --set tls.privateKey=tls.key minio/minio + local max=10 + local delay=60 + shift 2 + local n=1 + helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G --set tls.enabled=true --set tls.certSecret=tls-minio --set tls.publicCrt=tls.crt --set tls.privateKey=tls.key minio/minio NAME: minio-service LAST DEPLOYED: Wed Apr 29 16:51:37 2026 NAMESPACE: pitr-pxb-25367 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: MinIO can be accessed via port 9000 on the following DNS name from within your cluster: minio-service.pitr-pxb-25367.cluster.local To access MinIO from localhost, run the below commands: 1. export POD_NAME=$(kubectl get pods --namespace pitr-pxb-25367 -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}") 2. kubectl port-forward $POD_NAME 9000 --namespace pitr-pxb-25367 Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/ You can now access MinIO server on http://localhost:9000. Follow the below steps to connect to MinIO server with mc client: 1. Download the MinIO mc client - https://min.io/docs/minio/linux/reference/minio-mc.html#quickstart 2. export MC_HOST_minio-service-local=http://$(kubectl get secret --namespace pitr-pxb-25367 minio-service -o jsonpath="{.data.rootUser}" | base64 --decode):$(kubectl get secret --namespace pitr-pxb-25367 minio-service -o jsonpath="{.data.rootPassword}" | base64 --decode)@localhost:9000 3. mc ls minio-service-local + sleep 30 ++ kubectl_bin get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PFG7Agvdhf +++ mktemp ++ local LAST_ERR=/tmp/tmp.djgxf70X7A ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PFG7Agvdhf ++ cat /tmp/tmp.djgxf70X7A ++ rm /tmp/tmp.PFG7Agvdhf /tmp/tmp.djgxf70X7A ++ return 0 + MINIO_POD=minio-service-757b74c96c-r6mw5 + wait_pod minio-service-757b74c96c-r6mw5 + local pod=minio-service-757b74c96c-r6mw5 + local max_retry=480 + local ns= ++ echo minio-service-757b74c96c-r6mw5 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/minio-service-757b74c96c-r6mw5 condition met waiting for pod/minio-service-757b74c96c-r6mw5 to become Ready.Ok + kubectl_bin run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- /usr/bin/env AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url https://minio-service:9000 --no-verify-ssl s3 mb s3://operator-testing ++ mktemp + local LAST_OUT=/tmp/tmp.CvtUoxqlw1 ++ mktemp + local LAST_ERR=/tmp/tmp.mgN0DuGKyq + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- /usr/bin/env AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url https://minio-service:9000 --no-verify-ssl s3 mb s3://operator-testing + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.CvtUoxqlw1 /usr/lib/python2.7/site-packages/botocore/vendored/requests/packages/urllib3/connectionpool.py:768: InsecureRequestWarning: Unverified HTTPS request is being made. Adding certificate verification is strongly advised. See: https://urllib3.readthedocs.org/en/latest/security.html InsecureRequestWarning) make_bucket: operator-testing pod "aws-cli" deleted from pitr-pxb-25367 namespace + cat /tmp/tmp.mgN0DuGKyq All commands and output from this session will be recorded in container logs, including credentials and sensitive information passed through the command prompt. If you don't see a command prompt, try pressing enter. + rm /tmp/tmp.CvtUoxqlw1 /tmp/tmp.mgN0DuGKyq + return 0 + cluster=pitr + spinup_pxc pitr /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/pitr.yml + local cluster=pitr + local config=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/pitr.yml + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/secrets.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml + local port=3306 + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.QbwJsZdjWP ++ mktemp + local LAST_ERR=/tmp/tmp.dOnZhRQKeC + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.QbwJsZdjWP secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created + cat /tmp/tmp.dOnZhRQKeC + rm /tmp/tmp.QbwJsZdjWP /tmp/tmp.dOnZhRQKeC + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml + local pvc_name= + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml '' + kubectl_bin apply -f - + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml + local pvc_name= + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.4#' + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.4#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/fluentbit:main-logcollector#' ++ mktemp + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.4-backup#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.pitr-pxb-25367~ + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml + local LAST_OUT=/tmp/tmp.xTe4buR4Zy ++ mktemp + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2433-9bec550e#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + local LAST_ERR=/tmp/tmp.ZACr4szTkb + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.xTe4buR4Zy deployment.apps/pxc-client created + cat /tmp/tmp.ZACr4szTkb + rm /tmp/tmp.xTe4buR4Zy /tmp/tmp.ZACr4szTkb + return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/pitr.yml + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/pitr.yml + local pvc_name= + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/pitr.yml '' + kubectl_bin apply -f - + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/pitr.yml + local pvc_name= ++ mktemp + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.4#' + local LAST_OUT=/tmp/tmp.o8WFqvQEPt ++ mktemp + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.4#' + local LAST_ERR=/tmp/tmp.G5kVqtxSTK + local exit_status=0 + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/fluentbit:main-logcollector#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.pitr-pxb-25367~ + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.4-backup#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2433-9bec550e#' ++ seq 0 2 + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/pitr.yml + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.o8WFqvQEPt perconaxtradbcluster.pxc.percona.com/pitr created + cat /tmp/tmp.G5kVqtxSTK + rm /tmp/tmp.o8WFqvQEPt /tmp/tmp.G5kVqtxSTK + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- ++ get_proxy pitr ++ local target_cluster=pitr +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.19etoVq6Lr ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ipt16JCcpG +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.19etoVq6Lr +++ cat /tmp/tmp.ipt16JCcpG +++ rm /tmp/tmp.19etoVq6Lr /tmp/tmp.ipt16JCcpG +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.AO5tOs4IZu ++++ mktemp +++ local LAST_ERR=/tmp/tmp.5MZLF0Ovdp +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.AO5tOs4IZu +++ cat /tmp/tmp.5MZLF0Ovdp +++ rm /tmp/tmp.AO5tOs4IZu /tmp/tmp.5MZLF0Ovdp +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo pitr-proxysql ++ return + local proxy=pitr-proxysql + kubectl_bin wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n pitr-pxb-25367 ++ mktemp + local LAST_OUT=/tmp/tmp.xqU1mBO1b5 ++ mktemp + local LAST_ERR=/tmp/tmp.Wb0OjKl2Wz + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n pitr-pxb-25367 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n pitr-pxb-25367 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n pitr-pxb-25367 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.xqU1mBO1b5 + cat /tmp/tmp.Wb0OjKl2Wz error: no matching resources found + rm /tmp/tmp.xqU1mBO1b5 /tmp/tmp.Wb0OjKl2Wz + return 1 + true + wait_for_running pitr-proxysql 1 + local name=pitr-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod pitr-proxysql-0 480 + local pod=pitr-proxysql-0 + local max_retry=480 + local ns= ++ echo pitr-proxysql-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=proxysql + set +o xtrace Error from server (NotFound): pods "pitr-proxysql-0" not found waiting for pod/pitr-proxysql-0 to become Ready...........Ok + wait_for_running pitr-pxc 3 + local name=pitr-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod pitr-pxc-0 480 + local pod=pitr-pxc-0 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pitr-pxc-0 + local container=pxc + set +o xtrace pod/pitr-pxc-0 condition met waiting for pod/pitr-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod pitr-pxc-1 480 + local pod=pitr-pxc-1 + local max_retry=480 + local ns= ++ echo pitr-pxc-1 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/pitr-pxc-1 condition met waiting for pod/pitr-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod pitr-pxc-2 480 + local pod=pitr-pxc-2 + local max_retry=480 + local ns= ++ echo pitr-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/pitr-pxc-2 condition met waiting for pod/pitr-pxc-2 to become Ready.Ok + sleep 10 ++ kubectl get pxc pitr -o 'jsonpath={.spec.secretsName}' + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ base64 --decode ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uo6fEne7yW +++ mktemp ++ local LAST_ERR=/tmp/tmp.JJOkQErwam ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.uo6fEne7yW ++ cat /tmp/tmp.JJOkQErwam ++ rm /tmp/tmp.uo6fEne7yW /tmp/tmp.JJOkQErwam ++ return 0 + local root_pass=root_password + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' '-h pitr-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' + local 'uri=-h pitr-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GT9eFMvKB4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.1u7IFLoG2q ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GT9eFMvKB4 ++ cat /tmp/tmp.1u7IFLoG2q ++ rm /tmp/tmp.GT9eFMvKB4 /tmp/tmp.1u7IFLoG2q ++ return 0 + client_pod=pxc-client-56fd5498cd-72vmw + wait_pod pxc-client-56fd5498cd-72vmw + local pod=pxc-client-56fd5498cd-72vmw + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-72vmw ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-72vmw condition met waiting for pod/pxc-client-56fd5498cd-72vmw to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h pitr-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h pitr-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KQ1DGkH1gv +++ mktemp ++ local LAST_ERR=/tmp/tmp.7hVKSm5IEv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KQ1DGkH1gv ++ cat /tmp/tmp.7hVKSm5IEv ++ rm /tmp/tmp.KQ1DGkH1gv /tmp/tmp.7hVKSm5IEv ++ return 0 + client_pod=pxc-client-56fd5498cd-72vmw + wait_pod pxc-client-56fd5498cd-72vmw + local pod=pxc-client-56fd5498cd-72vmw + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-56fd5498cd-72vmw + local container= + set +o xtrace pod/pxc-client-56fd5498cd-72vmw condition met waiting for pod/pxc-client-56fd5498cd-72vmw to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h pitr-pxc-0.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-1-84.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h pitr-pxc-0.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vc6UlYDZuu +++ mktemp ++ local LAST_ERR=/tmp/tmp.HUPp3BZnUd ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.vc6UlYDZuu ++ cat /tmp/tmp.HUPp3BZnUd ++ rm /tmp/tmp.vc6UlYDZuu /tmp/tmp.HUPp3BZnUd ++ return 0 + client_pod=pxc-client-56fd5498cd-72vmw + wait_pod pxc-client-56fd5498cd-72vmw + local pod=pxc-client-56fd5498cd-72vmw + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-72vmw ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-72vmw condition met waiting for pod/pxc-client-56fd5498cd-72vmw to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.CsOYL2N7Rv/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.CsOYL2N7Rv/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-1.sql /tmp/tmp.CsOYL2N7Rv/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h pitr-pxc-1.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-1-84.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h pitr-pxc-1.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZXCI4DLdK2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.IRJt5tRV9a ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ZXCI4DLdK2 ++ cat /tmp/tmp.IRJt5tRV9a ++ rm /tmp/tmp.ZXCI4DLdK2 /tmp/tmp.IRJt5tRV9a ++ return 0 + client_pod=pxc-client-56fd5498cd-72vmw + wait_pod pxc-client-56fd5498cd-72vmw + local pod=pxc-client-56fd5498cd-72vmw + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-72vmw ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-72vmw condition met waiting for pod/pxc-client-56fd5498cd-72vmw to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.CsOYL2N7Rv/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.CsOYL2N7Rv/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-1.sql /tmp/tmp.CsOYL2N7Rv/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h pitr-pxc-2.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-1-84.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h pitr-pxc-2.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vrmcFFk6jK +++ mktemp ++ local LAST_ERR=/tmp/tmp.wVjMUIOr1B ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.vrmcFFk6jK ++ cat /tmp/tmp.wVjMUIOr1B ++ rm /tmp/tmp.vrmcFFk6jK /tmp/tmp.wVjMUIOr1B ++ return 0 + client_pod=pxc-client-56fd5498cd-72vmw + wait_pod pxc-client-56fd5498cd-72vmw + local pod=pxc-client-56fd5498cd-72vmw + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-72vmw ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-72vmw condition met waiting for pod/pxc-client-56fd5498cd-72vmw to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.CsOYL2N7Rv/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.CsOYL2N7Rv/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-1.sql /tmp/tmp.CsOYL2N7Rv/select-1.sql + is_keyring_plugin_in_use pitr + local cluster=pitr + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + kubectl exec pitr-pxc-0 -c pxc -- ls /var/lib/mysql/mysqld.my + return 1 + run_backup pitr on-pitr-minio + local cluster=pitr + local backup=on-pitr-minio + local name= + '[' -z ']' + name=on-pitr-minio + log 'run backup pxc-backup/on-pitr-minio' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T16:58:33+0000]' run backup pxc-backup/on-pitr-minio [2026-04-29T16:58:33+0000] run backup pxc-backup/on-pitr-minio + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/on-pitr-minio.yml + yq eval '.metadata.name="on-pitr-minio"' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.EMuRj9w0JV ++ mktemp + local LAST_ERR=/tmp/tmp.aXawSWC0Ha + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.EMuRj9w0JV perconaxtradbclusterbackup.pxc.percona.com/on-pitr-minio created + cat /tmp/tmp.aXawSWC0Ha + rm /tmp/tmp.EMuRj9w0JV /tmp/tmp.aXawSWC0Ha + return 0 + wait_backup on-pitr-minio + local backup=on-pitr-minio + local status=Succeeded + set +o xtrace waiting for pxc-backup/on-pitr-minio to reach Succeeded state..............Succeeded + write_test_data pitr + local cluster=pitr + local config= + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/secrets.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/conf/client.yml ++ get_proxy pitr ++ local target_cluster=pitr +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.aT0F0IZLyt ++++ mktemp +++ local LAST_ERR=/tmp/tmp.11xceMD5la +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.aT0F0IZLyt +++ cat /tmp/tmp.11xceMD5la +++ rm /tmp/tmp.aT0F0IZLyt /tmp/tmp.11xceMD5la +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.j0lUqNJioO ++++ mktemp +++ local LAST_ERR=/tmp/tmp.KfPNbpVFPs +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.j0lUqNJioO +++ cat /tmp/tmp.KfPNbpVFPs +++ rm /tmp/tmp.j0lUqNJioO /tmp/tmp.KfPNbpVFPs +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo pitr-proxysql ++ return + local proxy=pitr-proxysql + desc 'write test data' + set +o xtrace ----------------------------------------------------------------------------------- write test data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS test; use test; CREATE TABLE IF NOT EXISTS test (id int PRIMARY KEY) ;' '-h pitr-proxysql -uroot -proot_password' + local 'command=CREATE DATABASE IF NOT EXISTS test; use test; CREATE TABLE IF NOT EXISTS test (id int PRIMARY KEY) ;' + local 'uri=-h pitr-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gDns6vPst6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.77noI9OmLf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gDns6vPst6 ++ cat /tmp/tmp.77noI9OmLf ++ rm /tmp/tmp.gDns6vPst6 /tmp/tmp.77noI9OmLf ++ return 0 + client_pod=pxc-client-56fd5498cd-72vmw + wait_pod pxc-client-56fd5498cd-72vmw + local pod=pxc-client-56fd5498cd-72vmw + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-72vmw ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-72vmw condition met waiting for pod/pxc-client-56fd5498cd-72vmw to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT test.test (id) VALUES (100500); INSERT test.test (id) VALUES (100501); INSERT test.test (id) VALUES (100502);' '-h pitr-proxysql -uroot -proot_password' + local 'command=INSERT test.test (id) VALUES (100500); INSERT test.test (id) VALUES (100501); INSERT test.test (id) VALUES (100502);' + local 'uri=-h pitr-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ojjcxHzCz5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.UwK6d1i4uB ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ojjcxHzCz5 ++ cat /tmp/tmp.UwK6d1i4uB ++ rm /tmp/tmp.ojjcxHzCz5 /tmp/tmp.UwK6d1i4uB ++ return 0 + client_pod=pxc-client-56fd5498cd-72vmw + wait_pod pxc-client-56fd5498cd-72vmw + local pod=pxc-client-56fd5498cd-72vmw + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-72vmw ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-72vmw condition met waiting for pod/pxc-client-56fd5498cd-72vmw to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep_with_log 30 + local d=30 + log 'sleeping for 30 seconds' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T16:59:18+0000]' sleeping for 30 seconds [2026-04-29T16:59:18+0000] sleeping for 30 seconds + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-2 'SELECT * from test.test;' '-h pitr-pxc-0.pitr-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-2-84.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'SELECT * from test.test;' '-h pitr-pxc-0.pitr-pxc -uroot -proot_password' + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uiXMRs9JOn +++ mktemp ++ local LAST_ERR=/tmp/tmp.XlFQ08zilC ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.uiXMRs9JOn ++ cat /tmp/tmp.XlFQ08zilC ++ rm /tmp/tmp.uiXMRs9JOn /tmp/tmp.XlFQ08zilC ++ return 0 + client_pod=pxc-client-56fd5498cd-72vmw + wait_pod pxc-client-56fd5498cd-72vmw + local pod=pxc-client-56fd5498cd-72vmw + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-56fd5498cd-72vmw + local container= + set +o xtrace pod/pxc-client-56fd5498cd-72vmw condition met waiting for pod/pxc-client-56fd5498cd-72vmw to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.CsOYL2N7Rv/select-2.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.CsOYL2N7Rv/select-2.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-2.sql /tmp/tmp.CsOYL2N7Rv/select-2.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-2 'SELECT * from test.test;' '-h pitr-pxc-1.pitr-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-2-84.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'SELECT * from test.test;' '-h pitr-pxc-1.pitr-pxc -uroot -proot_password' + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.z6NIm6bvYL +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZFKRwiIjLl ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.z6NIm6bvYL ++ cat /tmp/tmp.ZFKRwiIjLl ++ rm /tmp/tmp.z6NIm6bvYL /tmp/tmp.ZFKRwiIjLl ++ return 0 + client_pod=pxc-client-56fd5498cd-72vmw + wait_pod pxc-client-56fd5498cd-72vmw + local pod=pxc-client-56fd5498cd-72vmw + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-56fd5498cd-72vmw + local container= + set +o xtrace pod/pxc-client-56fd5498cd-72vmw condition met waiting for pod/pxc-client-56fd5498cd-72vmw to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.CsOYL2N7Rv/select-2.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.CsOYL2N7Rv/select-2.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-2.sql /tmp/tmp.CsOYL2N7Rv/select-2.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-2 'SELECT * from test.test;' '-h pitr-pxc-2.pitr-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-2-84.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 5\.7 ]] + run_mysql 'SELECT * from test.test;' '-h pitr-pxc-2.pitr-pxc -uroot -proot_password' + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DW3kFvpjkS +++ mktemp ++ local LAST_ERR=/tmp/tmp.fZn15x2Iq4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DW3kFvpjkS ++ cat /tmp/tmp.fZn15x2Iq4 ++ rm /tmp/tmp.DW3kFvpjkS /tmp/tmp.fZn15x2Iq4 ++ return 0 + client_pod=pxc-client-56fd5498cd-72vmw + wait_pod pxc-client-56fd5498cd-72vmw + local pod=pxc-client-56fd5498cd-72vmw + local max_retry=480 + local ns= ++ echo pxc-client-56fd5498cd-72vmw ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-56fd5498cd-72vmw condition met waiting for pod/pxc-client-56fd5498cd-72vmw to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.CsOYL2N7Rv/select-2.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.CsOYL2N7Rv/select-2.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/select-2.sql /tmp/tmp.CsOYL2N7Rv/select-2.sql ++ is_keyring_plugin_in_use pitr ++ local cluster=pitr ++ [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] ++ kubectl exec pitr-pxc-0 -c pxc -- ls /var/lib/mysql/mysqld.my ++ return 1 + '[' '' ']' + desc 'patching xtrabackup password' + set +o xtrace ----------------------------------------------------------------------------------- patching xtrabackup password ----------------------------------------------------------------------------------- ++ echo -n pass1 ++ base64 + patch_secret my-cluster-secrets xtrabackup cGFzczE= + local secret=my-cluster-secrets + local key=xtrabackup + local value=cGFzczE= + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczE="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.bhJBlMdXrP ++ mktemp + local LAST_ERR=/tmp/tmp.fhnD6tFoG8 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczE="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.bhJBlMdXrP secret/my-cluster-secrets patched + cat /tmp/tmp.fhnD6tFoG8 + rm /tmp/tmp.bhJBlMdXrP /tmp/tmp.fhnD6tFoG8 + return 0 + wait_cluster_consistency pitr 3 2 + local cluster_name=pitr + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/pitr to be ready' waiting for pxc/pitr to be ready++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RVzA35WkF3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZlBgUQFRv3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.RVzA35WkF3 ++ cat /tmp/tmp.ZlBgUQFRv3 ++ rm /tmp/tmp.RVzA35WkF3 /tmp/tmp.ZlBgUQFRv3 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kayFPraTYY +++ mktemp ++ local LAST_ERR=/tmp/tmp.PBzX72xY5U ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kayFPraTYY ++ cat /tmp/tmp.PBzX72xY5U ++ rm /tmp/tmp.kayFPraTYY /tmp/tmp.PBzX72xY5U ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LsRbXSLYLM +++ mktemp ++ local LAST_ERR=/tmp/tmp.DcHnkmltkI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.LsRbXSLYLM ++ cat /tmp/tmp.DcHnkmltkI ++ rm /tmp/tmp.LsRbXSLYLM /tmp/tmp.DcHnkmltkI ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.38cQG0PZqY +++ mktemp ++ local LAST_ERR=/tmp/tmp.aRHCEXwXcY ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.38cQG0PZqY ++ cat /tmp/tmp.aRHCEXwXcY ++ rm /tmp/tmp.38cQG0PZqY /tmp/tmp.aRHCEXwXcY ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Y4lyUY3Xsy +++ mktemp ++ local LAST_ERR=/tmp/tmp.32nvgOIcxw ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Y4lyUY3Xsy ++ cat /tmp/tmp.32nvgOIcxw ++ rm /tmp/tmp.Y4lyUY3Xsy /tmp/tmp.32nvgOIcxw ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ekTY4PxsOb +++ mktemp ++ local LAST_ERR=/tmp/tmp.fRoh2bgtDZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ekTY4PxsOb ++ cat /tmp/tmp.fRoh2bgtDZ ++ rm /tmp/tmp.ekTY4PxsOb /tmp/tmp.fRoh2bgtDZ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MnH9bmvRwA +++ mktemp ++ local LAST_ERR=/tmp/tmp.6926cqnQvt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MnH9bmvRwA ++ cat /tmp/tmp.6926cqnQvt ++ rm /tmp/tmp.MnH9bmvRwA /tmp/tmp.6926cqnQvt ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ogGY6kjrdI +++ mktemp ++ local LAST_ERR=/tmp/tmp.FNmkDzbUiW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ogGY6kjrdI ++ cat /tmp/tmp.FNmkDzbUiW ++ rm /tmp/tmp.ogGY6kjrdI /tmp/tmp.FNmkDzbUiW ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gAhv9o0LTk +++ mktemp ++ local LAST_ERR=/tmp/tmp.glRHTzJmt5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gAhv9o0LTk ++ cat /tmp/tmp.glRHTzJmt5 ++ rm /tmp/tmp.gAhv9o0LTk /tmp/tmp.glRHTzJmt5 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FyB41HRfcn +++ mktemp ++ local LAST_ERR=/tmp/tmp.0tNKbuOWuM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.FyB41HRfcn ++ cat /tmp/tmp.0tNKbuOWuM ++ rm /tmp/tmp.FyB41HRfcn /tmp/tmp.0tNKbuOWuM ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GiTDtJVRRx +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ihr7RD1trf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GiTDtJVRRx ++ cat /tmp/tmp.Ihr7RD1trf ++ rm /tmp/tmp.GiTDtJVRRx /tmp/tmp.Ihr7RD1trf ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8ZOzVc2U3s +++ mktemp ++ local LAST_ERR=/tmp/tmp.oGh17bNMPp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8ZOzVc2U3s ++ cat /tmp/tmp.oGh17bNMPp ++ rm /tmp/tmp.8ZOzVc2U3s /tmp/tmp.oGh17bNMPp ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NbNZgExH6B +++ mktemp ++ local LAST_ERR=/tmp/tmp.DNuGQjNRb6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NbNZgExH6B ++ cat /tmp/tmp.DNuGQjNRb6 ++ rm /tmp/tmp.NbNZgExH6B /tmp/tmp.DNuGQjNRb6 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.txOiDNhFHQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.d5eHLHbaqA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.txOiDNhFHQ ++ cat /tmp/tmp.d5eHLHbaqA ++ rm /tmp/tmp.txOiDNhFHQ /tmp/tmp.d5eHLHbaqA ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qGQ6gyPLBH +++ mktemp ++ local LAST_ERR=/tmp/tmp.s9UEOAg5Fm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qGQ6gyPLBH ++ cat /tmp/tmp.s9UEOAg5Fm ++ rm /tmp/tmp.qGQ6gyPLBH /tmp/tmp.s9UEOAg5Fm ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.To96ZQ9sod +++ mktemp ++ local LAST_ERR=/tmp/tmp.G3pgWZZr5N ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.To96ZQ9sod ++ cat /tmp/tmp.G3pgWZZr5N ++ rm /tmp/tmp.To96ZQ9sod /tmp/tmp.G3pgWZZr5N ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9Q1bh1brCX +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ph3RN5FNfa ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9Q1bh1brCX ++ cat /tmp/tmp.Ph3RN5FNfa ++ rm /tmp/tmp.9Q1bh1brCX /tmp/tmp.Ph3RN5FNfa ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.w5iYNQNr0j +++ mktemp ++ local LAST_ERR=/tmp/tmp.BqIV6a2ZQI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.w5iYNQNr0j ++ cat /tmp/tmp.BqIV6a2ZQI ++ rm /tmp/tmp.w5iYNQNr0j /tmp/tmp.BqIV6a2ZQI ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Oz7w6xvz7W +++ mktemp ++ local LAST_ERR=/tmp/tmp.sDWzKF34Nb ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Oz7w6xvz7W ++ cat /tmp/tmp.sDWzKF34Nb ++ rm /tmp/tmp.Oz7w6xvz7W /tmp/tmp.sDWzKF34Nb ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rqIGx553Yy +++ mktemp ++ local LAST_ERR=/tmp/tmp.vCSfm3lC04 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.rqIGx553Yy ++ cat /tmp/tmp.vCSfm3lC04 ++ rm /tmp/tmp.rqIGx553Yy /tmp/tmp.vCSfm3lC04 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.C218uoVhUZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.NYRJvH2Sb5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.C218uoVhUZ ++ cat /tmp/tmp.NYRJvH2Sb5 ++ rm /tmp/tmp.C218uoVhUZ /tmp/tmp.NYRJvH2Sb5 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SStey5kK0y +++ mktemp ++ local LAST_ERR=/tmp/tmp.RNh6pNuHZl ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.SStey5kK0y ++ cat /tmp/tmp.RNh6pNuHZl ++ rm /tmp/tmp.SStey5kK0y /tmp/tmp.RNh6pNuHZl ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4RsPZK8umM +++ mktemp ++ local LAST_ERR=/tmp/tmp.yiF16L77UD ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.4RsPZK8umM ++ cat /tmp/tmp.yiF16L77UD ++ rm /tmp/tmp.4RsPZK8umM /tmp/tmp.yiF16L77UD ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.e82XNgvG29 +++ mktemp ++ local LAST_ERR=/tmp/tmp.WkZMc09056 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.e82XNgvG29 ++ cat /tmp/tmp.WkZMc09056 ++ rm /tmp/tmp.e82XNgvG29 /tmp/tmp.WkZMc09056 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sNj2cX2qFf +++ mktemp ++ local LAST_ERR=/tmp/tmp.peNBNpKvac ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.sNj2cX2qFf ++ cat /tmp/tmp.peNBNpKvac ++ rm /tmp/tmp.sNj2cX2qFf /tmp/tmp.peNBNpKvac ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gfh0zUbGvn +++ mktemp ++ local LAST_ERR=/tmp/tmp.SQaL1CCz3P ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gfh0zUbGvn ++ cat /tmp/tmp.SQaL1CCz3P ++ rm /tmp/tmp.gfh0zUbGvn /tmp/tmp.SQaL1CCz3P ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine pitr +++ local cluster_name=pitr ++++ get_proxy pitr ++++ local target_cluster=pitr +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.MFTGqtHs1r ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.lwDPbAGeHp +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.MFTGqtHs1r +++++ cat /tmp/tmp.lwDPbAGeHp +++++ rm /tmp/tmp.MFTGqtHs1r /tmp/tmp.lwDPbAGeHp +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.RfPRJntIJi ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.RFIm0soBMX +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.RfPRJntIJi +++++ cat /tmp/tmp.RFIm0soBMX +++++ rm /tmp/tmp.RfPRJntIJi /tmp/tmp.RFIm0soBMX +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo pitr-proxysql ++++ return +++ local cluster_proxy=pitr-proxysql +++ echo proxysql ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8onK0KJIhc +++ mktemp ++ local LAST_ERR=/tmp/tmp.jF6WLSSZ3x ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8onK0KJIhc ++ cat /tmp/tmp.jF6WLSSZ3x ++ rm /tmp/tmp.8onK0KJIhc /tmp/tmp.jF6WLSSZ3x ++ return 0 + [[ 2 == \2 ]] + echo + desc 'patching xtrabackup password' + set +o xtrace ----------------------------------------------------------------------------------- patching xtrabackup password ----------------------------------------------------------------------------------- ++ base64 ++ echo -n pass2 + patch_secret my-cluster-secrets xtrabackup cGFzczI= + local secret=my-cluster-secrets + local key=xtrabackup + local value=cGFzczI= + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczI="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.2ixX5YwFQN ++ mktemp + local LAST_ERR=/tmp/tmp.5valjeOOb0 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczI="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.2ixX5YwFQN secret/my-cluster-secrets patched + cat /tmp/tmp.5valjeOOb0 + rm /tmp/tmp.2ixX5YwFQN /tmp/tmp.5valjeOOb0 + return 0 + wait_cluster_consistency pitr 3 2 + local cluster_name=pitr + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/pitr to be ready' waiting for pxc/pitr to be ready++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HYtOjH9DLz +++ mktemp ++ local LAST_ERR=/tmp/tmp.ieq9hfLQ1h ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.HYtOjH9DLz ++ cat /tmp/tmp.ieq9hfLQ1h ++ rm /tmp/tmp.HYtOjH9DLz /tmp/tmp.ieq9hfLQ1h ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WmbTYNEi7G +++ mktemp ++ local LAST_ERR=/tmp/tmp.ypesnQiBbM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.WmbTYNEi7G ++ cat /tmp/tmp.ypesnQiBbM ++ rm /tmp/tmp.WmbTYNEi7G /tmp/tmp.ypesnQiBbM ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1TNfkc91dd +++ mktemp ++ local LAST_ERR=/tmp/tmp.s5QYbnhoR5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1TNfkc91dd ++ cat /tmp/tmp.s5QYbnhoR5 ++ rm /tmp/tmp.1TNfkc91dd /tmp/tmp.s5QYbnhoR5 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IagdXtHpid +++ mktemp ++ local LAST_ERR=/tmp/tmp.OhcfT1qYqx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.IagdXtHpid ++ cat /tmp/tmp.OhcfT1qYqx ++ rm /tmp/tmp.IagdXtHpid /tmp/tmp.OhcfT1qYqx ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nRjLerqh8R +++ mktemp ++ local LAST_ERR=/tmp/tmp.sV1lLX0T00 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.nRjLerqh8R ++ cat /tmp/tmp.sV1lLX0T00 ++ rm /tmp/tmp.nRjLerqh8R /tmp/tmp.sV1lLX0T00 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tzWd4LtKRg +++ mktemp ++ local LAST_ERR=/tmp/tmp.rRuEJS9P1m ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.tzWd4LtKRg ++ cat /tmp/tmp.rRuEJS9P1m ++ rm /tmp/tmp.tzWd4LtKRg /tmp/tmp.rRuEJS9P1m ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.G95PIMSZ5j +++ mktemp ++ local LAST_ERR=/tmp/tmp.q5bMs4eNts ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.G95PIMSZ5j ++ cat /tmp/tmp.q5bMs4eNts ++ rm /tmp/tmp.G95PIMSZ5j /tmp/tmp.q5bMs4eNts ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gvi43pDzpm +++ mktemp ++ local LAST_ERR=/tmp/tmp.BuPdfKtnKW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gvi43pDzpm ++ cat /tmp/tmp.BuPdfKtnKW ++ rm /tmp/tmp.gvi43pDzpm /tmp/tmp.BuPdfKtnKW ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.X2UCCVjpPW +++ mktemp ++ local LAST_ERR=/tmp/tmp.QUYE2YCbZK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.X2UCCVjpPW ++ cat /tmp/tmp.QUYE2YCbZK ++ rm /tmp/tmp.X2UCCVjpPW /tmp/tmp.QUYE2YCbZK ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QePVHPiWXT +++ mktemp ++ local LAST_ERR=/tmp/tmp.cIaTKa1Nhg ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.QePVHPiWXT ++ cat /tmp/tmp.cIaTKa1Nhg ++ rm /tmp/tmp.QePVHPiWXT /tmp/tmp.cIaTKa1Nhg ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DTWSncPzF1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.K1lZTScgur ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DTWSncPzF1 ++ cat /tmp/tmp.K1lZTScgur ++ rm /tmp/tmp.DTWSncPzF1 /tmp/tmp.K1lZTScgur ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jNj86295Em +++ mktemp ++ local LAST_ERR=/tmp/tmp.XwyqpQ3iD8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jNj86295Em ++ cat /tmp/tmp.XwyqpQ3iD8 ++ rm /tmp/tmp.jNj86295Em /tmp/tmp.XwyqpQ3iD8 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oSVQdEKDp0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.iE34tY5CB4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.oSVQdEKDp0 ++ cat /tmp/tmp.iE34tY5CB4 ++ rm /tmp/tmp.oSVQdEKDp0 /tmp/tmp.iE34tY5CB4 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JyclUgXzWL +++ mktemp ++ local LAST_ERR=/tmp/tmp.9KwUrr56wt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.JyclUgXzWL ++ cat /tmp/tmp.9KwUrr56wt ++ rm /tmp/tmp.JyclUgXzWL /tmp/tmp.9KwUrr56wt ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uIsNJJqBcp +++ mktemp ++ local LAST_ERR=/tmp/tmp.bNXHDYYqLm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.uIsNJJqBcp ++ cat /tmp/tmp.bNXHDYYqLm ++ rm /tmp/tmp.uIsNJJqBcp /tmp/tmp.bNXHDYYqLm ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.D85kb3XOwq +++ mktemp ++ local LAST_ERR=/tmp/tmp.gzUKkCSPNH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.D85kb3XOwq ++ cat /tmp/tmp.gzUKkCSPNH ++ rm /tmp/tmp.D85kb3XOwq /tmp/tmp.gzUKkCSPNH ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.o0VItrltwy +++ mktemp ++ local LAST_ERR=/tmp/tmp.83MqpXlJXn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.o0VItrltwy ++ cat /tmp/tmp.83MqpXlJXn ++ rm /tmp/tmp.o0VItrltwy /tmp/tmp.83MqpXlJXn ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PQc49k0q0H +++ mktemp ++ local LAST_ERR=/tmp/tmp.MpBLKrU40s ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PQc49k0q0H ++ cat /tmp/tmp.MpBLKrU40s ++ rm /tmp/tmp.PQc49k0q0H /tmp/tmp.MpBLKrU40s ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8oZaC34Ubi +++ mktemp ++ local LAST_ERR=/tmp/tmp.cHoUFGrNck ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8oZaC34Ubi ++ cat /tmp/tmp.cHoUFGrNck ++ rm /tmp/tmp.8oZaC34Ubi /tmp/tmp.cHoUFGrNck ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sgeKtlyIGw +++ mktemp ++ local LAST_ERR=/tmp/tmp.AAhViFFejq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.sgeKtlyIGw ++ cat /tmp/tmp.AAhViFFejq ++ rm /tmp/tmp.sgeKtlyIGw /tmp/tmp.AAhViFFejq ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SP91J4oXKP +++ mktemp ++ local LAST_ERR=/tmp/tmp.HsABREpPUp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.SP91J4oXKP ++ cat /tmp/tmp.HsABREpPUp ++ rm /tmp/tmp.SP91J4oXKP /tmp/tmp.HsABREpPUp ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yBDGL4bsGd +++ mktemp ++ local LAST_ERR=/tmp/tmp.R7qPVmpKyO ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.yBDGL4bsGd ++ cat /tmp/tmp.R7qPVmpKyO ++ rm /tmp/tmp.yBDGL4bsGd /tmp/tmp.R7qPVmpKyO ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HFKFmpReTA +++ mktemp ++ local LAST_ERR=/tmp/tmp.dIi4P0W93T ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.HFKFmpReTA ++ cat /tmp/tmp.dIi4P0W93T ++ rm /tmp/tmp.HFKFmpReTA /tmp/tmp.dIi4P0W93T ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YiSnn8hUR0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.dV0r4F75dt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.YiSnn8hUR0 ++ cat /tmp/tmp.dV0r4F75dt ++ rm /tmp/tmp.YiSnn8hUR0 /tmp/tmp.dV0r4F75dt ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XRnXiiSZck +++ mktemp ++ local LAST_ERR=/tmp/tmp.0Y2g8qbSMz ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XRnXiiSZck ++ cat /tmp/tmp.0Y2g8qbSMz ++ rm /tmp/tmp.XRnXiiSZck /tmp/tmp.0Y2g8qbSMz ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 24 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vRvuQIpPsg +++ mktemp ++ local LAST_ERR=/tmp/tmp.kpg0MmkmtF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.vRvuQIpPsg ++ cat /tmp/tmp.kpg0MmkmtF ++ rm /tmp/tmp.vRvuQIpPsg /tmp/tmp.kpg0MmkmtF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 25 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ubEoR736i6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Y5thi0p6XD ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ubEoR736i6 ++ cat /tmp/tmp.Y5thi0p6XD ++ rm /tmp/tmp.ubEoR736i6 /tmp/tmp.Y5thi0p6XD ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZaPDMjfoon +++ mktemp ++ local LAST_ERR=/tmp/tmp.6yy1lDaRmd ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ZaPDMjfoon ++ cat /tmp/tmp.6yy1lDaRmd ++ rm /tmp/tmp.ZaPDMjfoon /tmp/tmp.6yy1lDaRmd ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine pitr +++ local cluster_name=pitr ++++ get_proxy pitr ++++ local target_cluster=pitr +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.zDsezS6giO ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.zTS9I4oc1W +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.zDsezS6giO +++++ cat /tmp/tmp.zTS9I4oc1W +++++ rm /tmp/tmp.zDsezS6giO /tmp/tmp.zTS9I4oc1W +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.HWnbAnhBkS ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.Vb3d6yHwaj +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.HWnbAnhBkS +++++ cat /tmp/tmp.Vb3d6yHwaj +++++ rm /tmp/tmp.HWnbAnhBkS /tmp/tmp.Vb3d6yHwaj +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo pitr-proxysql ++++ return +++ local cluster_proxy=pitr-proxysql +++ echo proxysql ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0fPrGm2ceG +++ mktemp ++ local LAST_ERR=/tmp/tmp.Eb6f4wZ1sa ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.0fPrGm2ceG ++ cat /tmp/tmp.Eb6f4wZ1sa ++ rm /tmp/tmp.0fPrGm2ceG /tmp/tmp.Eb6f4wZ1sa ++ return 0 + [[ 2 == \2 ]] + echo + desc 'patching xtrabackup password' + set +o xtrace ----------------------------------------------------------------------------------- patching xtrabackup password ----------------------------------------------------------------------------------- ++ echo -n pass3 ++ base64 + patch_secret my-cluster-secrets xtrabackup cGFzczM= + local secret=my-cluster-secrets + local key=xtrabackup + local value=cGFzczM= + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczM="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.FvL2thSxAH ++ mktemp + local LAST_ERR=/tmp/tmp.VbMVmC7led + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczM="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.FvL2thSxAH secret/my-cluster-secrets patched + cat /tmp/tmp.VbMVmC7led + rm /tmp/tmp.FvL2thSxAH /tmp/tmp.VbMVmC7led + return 0 + wait_cluster_consistency pitr 3 2 + local cluster_name=pitr + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/pitr to be ready' waiting for pxc/pitr to be ready++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.exyl1dLIEM +++ mktemp ++ local LAST_ERR=/tmp/tmp.ak5hidm1Zk ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.exyl1dLIEM ++ cat /tmp/tmp.ak5hidm1Zk ++ rm /tmp/tmp.exyl1dLIEM /tmp/tmp.ak5hidm1Zk ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.giYF8kGruw +++ mktemp ++ local LAST_ERR=/tmp/tmp.lr3RvisuHj ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.giYF8kGruw ++ cat /tmp/tmp.lr3RvisuHj ++ rm /tmp/tmp.giYF8kGruw /tmp/tmp.lr3RvisuHj ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gxUhyYj1Yc +++ mktemp ++ local LAST_ERR=/tmp/tmp.fM3t8vU1R9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gxUhyYj1Yc ++ cat /tmp/tmp.fM3t8vU1R9 ++ rm /tmp/tmp.gxUhyYj1Yc /tmp/tmp.fM3t8vU1R9 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GlT5OOoVH4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.QezlSpVw8j ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GlT5OOoVH4 ++ cat /tmp/tmp.QezlSpVw8j ++ rm /tmp/tmp.GlT5OOoVH4 /tmp/tmp.QezlSpVw8j ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.o49CwA40SE +++ mktemp ++ local LAST_ERR=/tmp/tmp.27qqpg4piY ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.o49CwA40SE ++ cat /tmp/tmp.27qqpg4piY ++ rm /tmp/tmp.o49CwA40SE /tmp/tmp.27qqpg4piY ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eTZ3R4NNd9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.k0xsrQFwyR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.eTZ3R4NNd9 ++ cat /tmp/tmp.k0xsrQFwyR ++ rm /tmp/tmp.eTZ3R4NNd9 /tmp/tmp.k0xsrQFwyR ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tIvaSvzmcF +++ mktemp ++ local LAST_ERR=/tmp/tmp.qyPEA0iXc7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.tIvaSvzmcF ++ cat /tmp/tmp.qyPEA0iXc7 ++ rm /tmp/tmp.tIvaSvzmcF /tmp/tmp.qyPEA0iXc7 ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RjNt0lWz6U +++ mktemp ++ local LAST_ERR=/tmp/tmp.RvxOUK8LTr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.RjNt0lWz6U ++ cat /tmp/tmp.RvxOUK8LTr ++ rm /tmp/tmp.RjNt0lWz6U /tmp/tmp.RvxOUK8LTr ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BCW9ZafKdZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.8IXgWWBtNy ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.BCW9ZafKdZ ++ cat /tmp/tmp.8IXgWWBtNy ++ rm /tmp/tmp.BCW9ZafKdZ /tmp/tmp.8IXgWWBtNy ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qQU0kOyeAj +++ mktemp ++ local LAST_ERR=/tmp/tmp.hh77Rmw26O ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qQU0kOyeAj ++ cat /tmp/tmp.hh77Rmw26O ++ rm /tmp/tmp.qQU0kOyeAj /tmp/tmp.hh77Rmw26O ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qGATxml1uw +++ mktemp ++ local LAST_ERR=/tmp/tmp.kCKZ8rDiqE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qGATxml1uw ++ cat /tmp/tmp.kCKZ8rDiqE ++ rm /tmp/tmp.qGATxml1uw /tmp/tmp.kCKZ8rDiqE ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QI1UKyc0mk +++ mktemp ++ local LAST_ERR=/tmp/tmp.8hZ2DFRCss ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.QI1UKyc0mk ++ cat /tmp/tmp.8hZ2DFRCss ++ rm /tmp/tmp.QI1UKyc0mk /tmp/tmp.8hZ2DFRCss ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.iA576UQtmu +++ mktemp ++ local LAST_ERR=/tmp/tmp.04xxEA6oPF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.iA576UQtmu ++ cat /tmp/tmp.04xxEA6oPF ++ rm /tmp/tmp.iA576UQtmu /tmp/tmp.04xxEA6oPF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.segGKuMLSm +++ mktemp ++ local LAST_ERR=/tmp/tmp.IDp1vMxnHF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.segGKuMLSm ++ cat /tmp/tmp.IDp1vMxnHF ++ rm /tmp/tmp.segGKuMLSm /tmp/tmp.IDp1vMxnHF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.T5uSr2KXFQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.vwHwpFWr0m ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.T5uSr2KXFQ ++ cat /tmp/tmp.vwHwpFWr0m ++ rm /tmp/tmp.T5uSr2KXFQ /tmp/tmp.vwHwpFWr0m ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.24M9jmUoaj +++ mktemp ++ local LAST_ERR=/tmp/tmp.Qk702MgmuC ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.24M9jmUoaj ++ cat /tmp/tmp.Qk702MgmuC ++ rm /tmp/tmp.24M9jmUoaj /tmp/tmp.Qk702MgmuC ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ydnczqQWOn +++ mktemp ++ local LAST_ERR=/tmp/tmp.RY7H8felw6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ydnczqQWOn ++ cat /tmp/tmp.RY7H8felw6 ++ rm /tmp/tmp.ydnczqQWOn /tmp/tmp.RY7H8felw6 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.om4uLzOToW +++ mktemp ++ local LAST_ERR=/tmp/tmp.tJ0CBrE2i5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.om4uLzOToW ++ cat /tmp/tmp.tJ0CBrE2i5 ++ rm /tmp/tmp.om4uLzOToW /tmp/tmp.tJ0CBrE2i5 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Rrg4H6iiyh +++ mktemp ++ local LAST_ERR=/tmp/tmp.qFKxMl5eJh ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Rrg4H6iiyh ++ cat /tmp/tmp.qFKxMl5eJh ++ rm /tmp/tmp.Rrg4H6iiyh /tmp/tmp.qFKxMl5eJh ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OusgvRGz6L +++ mktemp ++ local LAST_ERR=/tmp/tmp.7dPj43ZUQG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.OusgvRGz6L ++ cat /tmp/tmp.7dPj43ZUQG ++ rm /tmp/tmp.OusgvRGz6L /tmp/tmp.7dPj43ZUQG ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.93vz9TGZhO +++ mktemp ++ local LAST_ERR=/tmp/tmp.H4Mjg5eibu ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.93vz9TGZhO ++ cat /tmp/tmp.H4Mjg5eibu ++ rm /tmp/tmp.93vz9TGZhO /tmp/tmp.H4Mjg5eibu ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AmSiyqVgk3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.9KJa6ywEvq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.AmSiyqVgk3 ++ cat /tmp/tmp.9KJa6ywEvq ++ rm /tmp/tmp.AmSiyqVgk3 /tmp/tmp.9KJa6ywEvq ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8VYreGxJ52 +++ mktemp ++ local LAST_ERR=/tmp/tmp.nlAayQPAwf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8VYreGxJ52 ++ cat /tmp/tmp.nlAayQPAwf ++ rm /tmp/tmp.8VYreGxJ52 /tmp/tmp.nlAayQPAwf ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bMp9LmcdKt +++ mktemp ++ local LAST_ERR=/tmp/tmp.Y6g4KylKqb ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bMp9LmcdKt ++ cat /tmp/tmp.Y6g4KylKqb ++ rm /tmp/tmp.bMp9LmcdKt /tmp/tmp.Y6g4KylKqb ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IccgDGcDEY +++ mktemp ++ local LAST_ERR=/tmp/tmp.tuhcm3KOTF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.IccgDGcDEY ++ cat /tmp/tmp.tuhcm3KOTF ++ rm /tmp/tmp.IccgDGcDEY /tmp/tmp.tuhcm3KOTF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 24 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OLn8o0tRjR +++ mktemp ++ local LAST_ERR=/tmp/tmp.bJ3lQxdOKd ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.OLn8o0tRjR ++ cat /tmp/tmp.bJ3lQxdOKd ++ rm /tmp/tmp.OLn8o0tRjR /tmp/tmp.bJ3lQxdOKd ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 25 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dHDNLytnje +++ mktemp ++ local LAST_ERR=/tmp/tmp.iZEWyGFgCC ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dHDNLytnje ++ cat /tmp/tmp.iZEWyGFgCC ++ rm /tmp/tmp.dHDNLytnje /tmp/tmp.iZEWyGFgCC ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fd35yEsws8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.u3ZSWampYp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.fd35yEsws8 ++ cat /tmp/tmp.u3ZSWampYp ++ rm /tmp/tmp.fd35yEsws8 /tmp/tmp.u3ZSWampYp ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine pitr +++ local cluster_name=pitr ++++ get_proxy pitr ++++ local target_cluster=pitr +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.vtZNlePW0g ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.kUET60QvKY +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.vtZNlePW0g +++++ cat /tmp/tmp.kUET60QvKY +++++ rm /tmp/tmp.vtZNlePW0g /tmp/tmp.kUET60QvKY +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.fVwvzCAQcZ ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.n6Mjhkuhgd +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.fVwvzCAQcZ +++++ cat /tmp/tmp.n6Mjhkuhgd +++++ rm /tmp/tmp.fVwvzCAQcZ /tmp/tmp.n6Mjhkuhgd +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo pitr-proxysql ++++ return +++ local cluster_proxy=pitr-proxysql +++ echo proxysql ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qlvsc8U8kZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.PMmcYpUVtP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qlvsc8U8kZ ++ cat /tmp/tmp.PMmcYpUVtP ++ rm /tmp/tmp.qlvsc8U8kZ /tmp/tmp.PMmcYpUVtP ++ return 0 + [[ 2 == \2 ]] + echo + compare_kubectl secret/pitr-mysql-init + local resource=secret/pitr-mysql-init + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init.yml + local new_result=/tmp/tmp.CsOYL2N7Rv/secret_pitr-mysql-init.yml + desc 'compare secret/pitr-mysql-init-' + set +o xtrace ----------------------------------------------------------------------------------- compare secret/pitr-mysql-init- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.4 =~ 8\.4 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init-84.yml ']' + version_gt 1.33 ++ bc -l ++ echo '1.32 >= 1.33' + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ bc -l ++ echo '1.32 >= 1.29' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init-k129.yml ']' + version_gt 1.27 ++ echo '1.32 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init-k127.yml ']' + version_gt 1.24 ++ echo '1.32 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init-k124.yml ']' + version_gt 1.22 ++ bc -l ++ echo '1.32 >= 1.22' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init-k122.yml ']' + version_gt 1.21 ++ echo '1.32 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init-oc.yml ']' + version_gt 1.29 ++ echo '1.32 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init-aks.yml ']' + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.spec.template.spec.containers[].env[] | select(.name == "XTRABACKUP_ENABLED")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.metadata.annotations."kubernetes.digitalocean.com/load-balancer-id") | del(.metadata.annotations."service.beta.kubernetes.io/do-loadbalancer-type") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("pitr-pxb-25367", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + kubectl_bin get -o yaml secret/pitr-mysql-init ++ mktemp + local LAST_OUT=/tmp/tmp.FK0ukjfpuW ++ mktemp + local LAST_ERR=/tmp/tmp.b0oqPP7Cjz + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml secret/pitr-mysql-init + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.FK0ukjfpuW + cat /tmp/tmp.b0oqPP7Cjz + rm /tmp/tmp.FK0ukjfpuW /tmp/tmp.b0oqPP7Cjz + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/compare/secret_pitr-mysql-init.yml /tmp/tmp.CsOYL2N7Rv/secret_pitr-mysql-init.yml + log 'compare_kubectl: secret/pitr-mysql-init OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T17:10:00+0000]' compare_kubectl: secret/pitr-mysql-init OK [2026-04-29T17:10:00+0000] compare_kubectl: secret/pitr-mysql-init OK + mysql_host=pitr-pxc-0.pitr-pxc.pitr-pxb-25367 ++ get_gtid_executed pitr-pxc-0.pitr-pxc.pitr-pxb-25367 ++ local host=pitr-pxc-0.pitr-pxc.pitr-pxb-25367 +++ /usr/bin/sed 's/\([a-f0-9-]\{36\}\):[0-9]*-\([0-9]*\).*/\1:\2/' +++ run_mysql 'SELECT @@gtid_executed;' '-h pitr-pxc-0.pitr-pxc.pitr-pxb-25367 -uroot -proot_password' +++ local 'command=SELECT @@gtid_executed;' +++ local 'uri=-h pitr-pxc-0.pitr-pxc.pitr-pxb-25367 -uroot -proot_password' ++++ get_client_pod ++++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.HtKikdiGrF +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.8BvC6EnV5m ++++ local exit_status=0 +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ set +e ++++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 ']' ++++ break ++++ cat /tmp/tmp.HtKikdiGrF ++++ cat /tmp/tmp.8BvC6EnV5m ++++ rm /tmp/tmp.HtKikdiGrF /tmp/tmp.8BvC6EnV5m ++++ return 0 +++ client_pod=pxc-client-56fd5498cd-72vmw +++ wait_pod pxc-client-56fd5498cd-72vmw +++ local pod=pxc-client-56fd5498cd-72vmw +++ local max_retry=480 +++ local ns= ++++ echo pxc-client-56fd5498cd-72vmw ++++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++++ grep -E '^(pxc|proxysql)$' +++ local container= +++ set +o xtrace pod/pxc-client-56fd5498cd-72vmw condition met waiting for pod/pxc-client-56fd5498cd-72vmw to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok +++ set +o xtrace ++ local gtid=f399b36f-43eb-11f1-a745-ea1a4d500ba9:26 ++ [[ ! f399b36f-43eb-11f1-a745-ea1a4d500ba9:26 =~ [A-F0-9a-f]{8}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{12}:[0-9]+ ]] ++ echo f399b36f-43eb-11f1-a745-ea1a4d500ba9:26 + gtid=f399b36f-43eb-11f1-a745-ea1a4d500ba9:26 + log '[CASE 1]: gtid_executed: f399b36f-43eb-11f1-a745-ea1a4d500ba9:26' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-29T17:10:12+0000]' '[CASE' '1]:' gtid_executed: f399b36f-43eb-11f1-a745-ea1a4d500ba9:26 [2026-04-29T17:10:12+0000] [CASE 1]: gtid_executed: f399b36f-43eb-11f1-a745-ea1a4d500ba9:26 + write_data_for_pitr pitr 100503 3 + local cluster=pitr + local start=100503 + local rows=3 ++ get_proxy pitr ++ local target_cluster=pitr +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.gpKKVwbbqo ++++ mktemp +++ local LAST_ERR=/tmp/tmp.XXpsf28990 +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.gpKKVwbbqo +++ cat /tmp/tmp.XXpsf28990 +++ rm /tmp/tmp.gpKKVwbbqo /tmp/tmp.XXpsf28990 +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Cxdtdcv6Ug ++++ mktemp +++ local LAST_ERR=/tmp/tmp.W66sTMOG45 +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.Cxdtdcv6Ug +++ cat /tmp/tmp.W66sTMOG45 +++ rm /tmp/tmp.Cxdtdcv6Ug /tmp/tmp.W66sTMOG45 +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo pitr-proxysql ++ return + local proxy=pitr-proxysql + local sql= + (( i=0 )) + (( i/f399b36f-43eb-11f1-a745-ea1a4d500ba9:26/g' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.pitr-pxb-25367~ + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2433/e2e-tests/pitr-pxb/conf/restore-on-pitr-minio-gtid.yaml + /usr/bin/sed -e 's///g' + /usr/bin/sed -e 's///g' ++ mktemp + local LAST_OUT=/tmp/tmp.NffKLtIltC ++ mktemp + local LAST_ERR=/tmp/tmp.nEZBUtn93D + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.NffKLtIltC perconaxtradbclusterrestore.pxc.percona.com/restore-on-pitr-minio-gtid created + cat /tmp/tmp.nEZBUtn93D + rm /tmp/tmp.NffKLtIltC /tmp/tmp.nEZBUtn93D + return 0 + wait_backup_restore restore-on-pitr-minio-gtid 'Stopping Cluster' + local backup_name=restore-on-pitr-minio-gtid + local 'target_state=Stopping Cluster' + local wait_time=720 + set +o xtrace waiting for pxc-restore/restore-on-pitr-minio-gtid to reach Stopping Cluster state 2026-04-29T17:11:52 pxc-restore/restore-on-pitr-minio-gtid state: Stopping Cluster + wait_for_delete pod/pitr-proxysql-0 + local res=pod/pitr-proxysql-0 + echo -n 'waiting for pod/pitr-proxysql-0 to be deleted' waiting for pod/pitr-proxysql-0 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-0" not found + wait_for_delete pod/pitr-proxysql-1 + local res=pod/pitr-proxysql-1 + echo -n 'waiting for pod/pitr-proxysql-1 to be deleted' waiting for pod/pitr-proxysql-1 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-1" not found + wait_for_delete pod/pitr-pxc-2 + local res=pod/pitr-pxc-2 + echo -n 'waiting for pod/pitr-pxc-2 to be deleted' waiting for pod/pitr-pxc-2 to be deleted+ set +o xtrace ..Error from server (NotFound): pods "pitr-pxc-2" not found + wait_for_delete pod/pitr-pxc-1 + local res=pod/pitr-pxc-1 + echo -n 'waiting for pod/pitr-pxc-1 to be deleted' waiting for pod/pitr-pxc-1 to be deleted+ set +o xtrace ...Error from server (NotFound): pods "pitr-pxc-1" not found + wait_for_delete pod/pitr-pxc-0 + local res=pod/pitr-pxc-0 + echo -n 'waiting for pod/pitr-pxc-0 to be deleted' waiting for pod/pitr-pxc-0 to be deleted+ set +o xtrace ....Error from server (NotFound): pods "pitr-pxc-0" not found + wait_backup_restore restore-on-pitr-minio-gtid 'Point-in-time recovering' + local backup_name=restore-on-pitr-minio-gtid + local 'target_state=Point-in-time recovering' + local wait_time=720 + set +o xtrace waiting for pxc-restore/restore-on-pitr-minio-gtid to reach Point-in-time recovering state 2026-04-29T17:12:33 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:12:36 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:12:39 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:12:42 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:12:44 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:12:47 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:12:50 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:12:54 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:12:57 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:01 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:04 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:08 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:12 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:16 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:20 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:24 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:27 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:31 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:34 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:37 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:40 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:43 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:46 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:49 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:53 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:13:57 pxc-restore/restore-on-pitr-minio-gtid state: Restoring 2026-04-29T17:14:00 pxc-restore/restore-on-pitr-minio-gtid state: Point-in-time recovering + wait_for_delete pod/pitr-proxysql-0 + local res=pod/pitr-proxysql-0 + echo -n 'waiting for pod/pitr-proxysql-0 to be deleted' waiting for pod/pitr-proxysql-0 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-0" not found + wait_for_delete pod/pitr-proxysql-1 + local res=pod/pitr-proxysql-1 + echo -n 'waiting for pod/pitr-proxysql-1 to be deleted' waiting for pod/pitr-proxysql-1 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-proxysql-1" not found + wait_for_delete pod/pitr-pxc-2 + local res=pod/pitr-pxc-2 + echo -n 'waiting for pod/pitr-pxc-2 to be deleted' waiting for pod/pitr-pxc-2 to be deleted+ set +o xtrace Error from server (NotFound): pods "pitr-pxc-2" not found + wait_for_delete pod/pitr-pxc-1 + local res=pod/pitr-pxc-1 + echo -n 'waiting for pod/pitr-pxc-1 to be deleted' waiting for pod/pitr-pxc-1 to be deleted+ set +o xtrace ........................................................................................................................2026-04-29T16:48:56.213Z INFO setup Runs on {"platform": "kubernetes", "version": "v1.32.13-gke.1362000"} 2026-04-29T16:48:56.213Z INFO setup Manager starting up {"gitCommit": "9bec550e38b53ea8a42441eabf05aff8b718c288", "gitBranch": "PR-2433-9bec550e", "buildTime": "2026-04-29T14:21:52Z", "goVersion": "go1.25.9", "os": "linux", "arch": "amd64"} 2026-04-29T16:48:56.213Z INFO setup Feature gates {"PXCO_FEATURE_GATES": "XtrabackupSidecar=true", "enabled": "XtrabackupSidecar=true"} 2026-04-29T16:48:56.216Z INFO setup Registering Components. 2026-04-29T16:48:57.064Z INFO controller-runtime.webhook Registering webhook {"path": "/validate-percona-xtradbcluster"} 2026-04-29T16:48:57.064Z INFO setup Starting the Cmd. 2026-04-29T16:48:57.064Z INFO controller-runtime.metrics Starting metrics server 2026-04-29T16:48:57.064Z INFO controller-runtime.metrics Serving metrics server {"bindAddress": ":8080", "secure": false} 2026-04-29T16:48:57.064Z INFO starting server {"name": "health probe", "addr": "[::]:8081"} 2026-04-29T16:48:57.065Z INFO controller-runtime.webhook Starting webhook server 2026-04-29T16:48:57.065Z INFO controller-runtime.certwatcher Updated current TLS certificate {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key"} 2026-04-29T16:48:57.065Z INFO controller-runtime.webhook Serving webhook server {"host": "", "port": 9443} 2026-04-29T16:48:57.065Z INFO controller-runtime.certwatcher Starting certificate poll+watcher {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key", "interval": "10s"} 2026-04-29T16:48:57.165Z INFO Attempting to acquire leader lease... {"lock": "pxc-operator/08db1feb.percona.com"} 2026-04-29T16:48:57.214Z INFO Successfully acquired lease {"lock": "pxc-operator/08db1feb.percona.com"} 2026-04-29T16:48:57.215Z DEBUG events percona-xtradb-cluster-operator-6fc4ff78b7-96wkz_2261d2a0-839d-474d-bc40-77b4a21062c8 became leader {"type": "Normal", "object": {"kind":"Lease","namespace":"pxc-operator","name":"08db1feb.percona.com","uid":"45e51e4d-2818-440a-973b-04a34d70b3f7","apiVersion":"coordination.k8s.io/v1","resourceVersion":"1777481337204303009"}, "reason": "LeaderElection"} 2026-04-29T16:48:57.215Z INFO Starting EventSource {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "source": "kind source: *v1.Secret"} 2026-04-29T16:48:57.216Z INFO Starting EventSource {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "source": "kind source: *v1.PerconaXtraDBClusterRestore"} 2026-04-29T16:48:57.216Z INFO Starting EventSource {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "source": "kind source: *v1.PerconaXtraDBCluster"} 2026-04-29T16:48:57.216Z INFO Starting EventSource {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "source": "kind source: *v1.PerconaXtraDBClusterBackup"} 2026-04-29T16:48:57.316Z INFO Starting Controller {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore"} 2026-04-29T16:48:57.316Z INFO Starting workers {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "worker count": 1} 2026-04-29T16:48:57.316Z INFO Starting Controller {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster"} 2026-04-29T16:48:57.316Z INFO Starting workers {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "worker count": 1} 2026-04-29T16:48:57.317Z INFO Starting Controller {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup"} 2026-04-29T16:48:57.317Z INFO Starting workers {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "worker count": 1} 2026-04-29T16:52:59.869Z INFO Set CR version {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "ffd9c29d-fc30-49e3-8a82-0213bbbb41bf", "version": "1.20.0"} 2026-04-29T16:53:00.182Z INFO spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "ffd9c29d-fc30-49e3-8a82-0213bbbb41bf"} 2026-04-29T16:53:03.230Z INFO spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "ffd9c29d-fc30-49e3-8a82-0213bbbb41bf"} 2026-04-29T16:53:03.263Z INFO spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "ffd9c29d-fc30-49e3-8a82-0213bbbb41bf"} 2026-04-29T16:53:06.363Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "ffd9c29d-fc30-49e3-8a82-0213bbbb41bf", "object": "auto-pitr-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2026-04-29T16:53:06.382Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "ffd9c29d-fc30-49e3-8a82-0213bbbb41bf", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2026-04-29T16:53:06.500Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "ffd9c29d-fc30-49e3-8a82-0213bbbb41bf", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2026-04-29T16:53:06.550Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "ffd9c29d-fc30-49e3-8a82-0213bbbb41bf", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2026-04-29T16:53:06.611Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "ffd9c29d-fc30-49e3-8a82-0213bbbb41bf", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-29T16:53:06.654Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "ffd9c29d-fc30-49e3-8a82-0213bbbb41bf", "object": "pitr-pxc-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-29T16:53:06.766Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "ffd9c29d-fc30-49e3-8a82-0213bbbb41bf", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-29T16:53:06.853Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "ffd9c29d-fc30-49e3-8a82-0213bbbb41bf", "object": "pitr-proxysql-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2026-04-29T16:53:07.656Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "8f4bf401-9076-4c07-9be2-746e194c318c", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2026-04-29T16:53:07.678Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "8f4bf401-9076-4c07-9be2-746e194c318c", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2026-04-29T16:54:29.017Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "8bb218ec-47a6-4d45-b57d-3dd1a49dac71", "user": "operator"} 2026-04-29T16:54:29.058Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "8bb218ec-47a6-4d45-b57d-3dd1a49dac71", "user": "monitor"} 2026-04-29T16:54:29.127Z INFO User monitor: granted privileges {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "8bb218ec-47a6-4d45-b57d-3dd1a49dac71"} 2026-04-29T16:54:29.166Z INFO monitor user privileges granted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "8bb218ec-47a6-4d45-b57d-3dd1a49dac71"} 2026-04-29T16:54:29.214Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "8bb218ec-47a6-4d45-b57d-3dd1a49dac71", "user": "xtrabackup"} 2026-04-29T16:54:29.270Z INFO User xtrabackup: granted privileges {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "8bb218ec-47a6-4d45-b57d-3dd1a49dac71"} 2026-04-29T16:54:29.310Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "8bb218ec-47a6-4d45-b57d-3dd1a49dac71", "user": "replication"} 2026-04-29T16:54:29.319Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "8bb218ec-47a6-4d45-b57d-3dd1a49dac71", "err": "get primary pxc pod: not found"} 2026-04-29T16:54:34.435Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "5d39e393-f724-4c71-a991-c549eff07a5b", "err": "get primary pxc pod: not found"} 2026-04-29T16:54:39.536Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "85d0c5bf-2e22-4891-a8c9-d265eee8984a", "err": "get primary pxc pod: not found"} 2026-04-29T16:54:44.661Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "58d3b194-8a72-4a8e-ac91-fb17eb323a40", "err": "get primary pxc pod: not found"} 2026-04-29T16:57:05.343Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "f4612b19-2f20-4416-9a91-4b6dd5378301", "user": "root"} 2026-04-29T16:57:05.392Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "f4612b19-2f20-4416-9a91-4b6dd5378301", "object": "pitr-pitr", "kind": "&TypeMeta{Kind:,APIVersion:,}"} 2026-04-29T16:57:05.455Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "f4612b19-2f20-4416-9a91-4b6dd5378301", "object": "pitr-pitr", "kind": "&TypeMeta{Kind:Deployment,APIVersion:apps/v1,}"} 2026-04-29T16:57:05.506Z INFO update PXC version (fetched from db) {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "f4612b19-2f20-4416-9a91-4b6dd5378301", "new version": "8.4.7-7.1"} 2026-04-29T16:57:06.048Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "0ddaa133-107d-49fe-9f6b-d1a2ccbf740d", "object": "pitr-pitr", "kind": "&TypeMeta{Kind:,APIVersion:,}", "hashChanged": false, "metaChanged": true} &v1.Service{ TypeMeta: v1.TypeMeta{ - Kind: "Service", + Kind: "", - APIVersion: "v1", + APIVersion: "", }, ObjectMeta: v1.ObjectMeta{ ... // 2 identical fields Namespace: "pitr-pxb-25367", SelfLink: "", - UID: "53a5d369-2d40-4911-bf34-fd8892ef9774", + UID: "", - ResourceVersion: "1777481825448383002", + ResourceVersion: "", Generation: 0, - CreationTimestamp: v1.Time{Time: s"2026-04-29 16:57:05 +0000 UTC"}, + CreationTimestamp: v1.Time{}, DeletionTimestamp: nil, DeletionGracePeriodSeconds: nil, Labels: nil, Annotations: map[string]string{ - "cloud.google.com/neg": `{"ingress":true}`, "percona.com/last-config-hash": "eyJwb3J0cyI6W3sibmFtZSI6Imh0dHAiLCJwb3J0Ijo4MDgwLCJ0YXJnZXRQb3J0"..., }, OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "20559ea9-e17b-447e-a2cb-6b4a334d5aaa", ...}}, Finalizers: nil, - ManagedFields: []v1.ManagedFieldsEntry{ - { - Manager: "percona-xtradb-cluster-operator", - Operation: "Update", - APIVersion: "v1", - Time: s"2026-04-29 16:57:05 +0000 UTC", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., - }, - }, + ManagedFields: nil, }, Spec: v1.ServiceSpec{ Ports: []v1.ServicePort{ { Name: "http", - Protocol: "TCP", + Protocol: "", AppProtocol: nil, Port: 8080, TargetPort: intstr.IntOrString{ Type: 0, - IntVal: 8080, + IntVal: 0, StrVal: "", }, NodePort: 0, }, }, Selector: {"app.kubernetes.io/component": "pitr", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, ClusterIP: "34.118.229.122", - ClusterIPs: []string{"34.118.229.122"}, + ClusterIPs: nil, Type: "ClusterIP", ExternalIPs: nil, - SessionAffinity: "None", + SessionAffinity: "", LoadBalancerIP: "", LoadBalancerSourceRanges: nil, ... // 3 identical fields PublishNotReadyAddresses: false, SessionAffinityConfig: nil, - IPFamilies: []v1.IPFamily{"IPv4"}, + IPFamilies: nil, - IPFamilyPolicy: &"SingleStack", + IPFamilyPolicy: nil, AllocateLoadBalancerNodePorts: nil, LoadBalancerClass: nil, - InternalTrafficPolicy: &"Cluster", + InternalTrafficPolicy: nil, TrafficDistribution: nil, }, Status: {}, } 2026-04-29T16:57:07.409Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "f4612b19-2f20-4416-9a91-4b6dd5378301"} 2026-04-29T16:57:13.437Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "178a2d5d-4085-4c98-9d1c-6767c5bf560c"} 2026-04-29T16:57:18.852Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "a2b1c4bb-7c98-4ae2-b4a1-6a2cd3441bcb"} 2026-04-29T16:57:24.053Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "0c9a3bb2-f01a-453d-80eb-8ba2c1d8b36e"} 2026-04-29T16:57:29.341Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "0a90f9b6-2ef1-49cb-8b22-b1cbd6382aba"} 2026-04-29T16:57:34.717Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "1149a1db-05df-4fea-848a-ee51f423768b"} 2026-04-29T16:57:39.757Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "4d82ba81-8bb7-488e-8152-440f31497178"} 2026-04-29T16:57:45.138Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "b065188b-8fb1-437d-907e-c345401e4b68"} 2026-04-29T16:57:50.267Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "50c860d9-dca5-4c36-a5e1-b94633183ba1"} 2026-04-29T16:57:55.858Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "3bbccd40-3ac7-4716-865a-a4c77c93fff5"} 2026-04-29T16:58:01.110Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "6ccee154-6701-43dc-8072-e2d95b171e45"} 2026-04-29T16:58:06.468Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "c5f4006f-92ca-4cf7-8068-4513c388a191"} 2026-04-29T16:58:11.544Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "8d22ef94-988b-46f1-ae33-1bc2089d0c9c"} 2026-04-29T16:58:16.843Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "bcf9d01e-335e-4710-9396-388580ce79e0"} 2026-04-29T16:58:21.950Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "cd08cd56-6290-4854-9ebe-6e5459536b19"} 2026-04-29T16:58:27.524Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "38e5ea15-d179-4ea6-b791-d09057fd05ca"} 2026-04-29T16:58:32.733Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "6754b599-ce22-47de-b469-884367369506"} 2026-04-29T16:58:35.261Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "on-pitr-minio", "reconcileID": "3a350974-9bb1-460a-a154-a8a1988aac1d", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:58:35.305Z INFO Created a new backup job {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "on-pitr-minio", "reconcileID": "3a350974-9bb1-460a-a154-a8a1988aac1d", "namespace": "pitr-pxb-25367", "name": "xb-on-pitr-minio"} 2026-04-29T16:58:38.109Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "e53206a6-c988-4d8c-94a1-1830ece846a8"} 2026-04-29T16:58:40.306Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "on-pitr-minio", "reconcileID": "3bc86283-c41a-46bc-a0d8-5df1df7d2dd5", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:58:40.392Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "on-pitr-minio", "reconcileID": "e0a1077d-23c5-48f6-aba0-7916416eaa7a", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:58:43.155Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "867b9947-ac55-4104-8a8e-cf9ba99efa24"} 2026-04-29T16:58:45.438Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "on-pitr-minio", "reconcileID": "25d9d81e-2322-46da-8208-948e0296a482", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:58:45.517Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "on-pitr-minio", "reconcileID": "f690e53f-e290-4d79-baad-50e8a4c8f377", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:58:48.652Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "d092bd23-556f-4624-8ce9-d8ed1c202b53"} 2026-04-29T16:58:50.577Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "on-pitr-minio", "reconcileID": "3e7f2713-d316-48a0-9891-0fd5232f8043", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:58:50.656Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "on-pitr-minio", "reconcileID": "af88b54e-bc2a-4ce8-b4ea-843f0ad4b841", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:58:53.834Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "1879d6ab-e8a2-402d-989a-413cecd09c04"} 2026-04-29T16:58:55.701Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "on-pitr-minio", "reconcileID": "5ba1f362-c5d9-479c-8520-e15ef8429a70", "cluster": "pitr", "storage": "minio", "allowed": true} 2026-04-29T16:58:55.757Z INFO Backup succeeded {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "on-pitr-minio", "reconcileID": "5ba1f362-c5d9-479c-8520-e15ef8429a70", "job": "xb-on-pitr-minio"} 2026-04-29T16:58:55.758Z DEBUG Removing binlog gap file from binlog collector {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "on-pitr-minio", "reconcileID": "5ba1f362-c5d9-479c-8520-e15ef8429a70", "job": "xb-on-pitr-minio", "pod": "pitr-pitr-6f99f5b5fc-lsbh2"} 2026-04-29T16:58:57.176Z DEBUG Removing binlog timeline file from binlog collector {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "on-pitr-minio", "reconcileID": "5ba1f362-c5d9-479c-8520-e15ef8429a70", "job": "xb-on-pitr-minio", "pod": "pitr-pitr-6f99f5b5fc-lsbh2"} 2026-04-29T16:58:57.257Z DEBUG Removing mysql-init secret {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "PerconaXtraDBClusterBackup": {"name":"on-pitr-minio","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "on-pitr-minio", "reconcileID": "5ba1f362-c5d9-479c-8520-e15ef8429a70", "job": "xb-on-pitr-minio", "secret": "pitr-mysql-init"} 2026-04-29T16:58:59.323Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "b0506ed1-5973-4c44-b56b-e02ca68c9ef1"} 2026-04-29T16:59:05.041Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "382f73d5-791b-4c0c-b5c7-f35a289027ef"} 2026-04-29T16:59:10.853Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "c427dd75-af4d-4dd7-a4d3-f9eca586e46f"} 2026-04-29T16:59:16.717Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "e3a78eb6-da96-4016-9161-c849458921a4"} 2026-04-29T16:59:22.548Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "62f98c6d-a9dc-46c6-b668-bdf03b462435"} 2026-04-29T16:59:28.433Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "1f86823e-6253-4849-be53-d6e849ec52d5"} 2026-04-29T16:59:34.330Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "2344b49d-f4bb-4794-85e6-c9e5f8fee7b8"} 2026-04-29T16:59:39.823Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "b0f8faea-d221-4d0f-b091-a5df01bc9235"} 2026-04-29T16:59:45.662Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "649d0e60-b3e5-4489-8075-6b43cc9928fc"} 2026-04-29T16:59:51.324Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "3d72e945-38e4-40d3-b31a-d37a1c6450e1"} 2026-04-29T16:59:57.543Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "2d3a50c4-5618-4cb7-9936-1cddb267785e"} 2026-04-29T17:00:03.239Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "d5932f76-580a-4f8f-83fd-2c9c53f025b3"} 2026-04-29T17:00:07.714Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "4c31950e-4694-4eb1-b5ba-6c7c0dfcae25", "latest": "2026-04-29 17:00:01 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T17:00:09.151Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "4c31950e-4694-4eb1-b5ba-6c7c0dfcae25"} 2026-04-29T17:00:14.564Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "e2ce3cb4-97e7-4a05-9b11-cc3e69d327a8"} 2026-04-29T17:00:20.664Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "aad2718a-d50f-48dd-a0a4-9ba1fcac238d"} 2026-04-29T17:00:25.919Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "fce910c9-b3b9-48b9-956c-894736f274f8", "user": "xtrabackup"} 2026-04-29T17:00:25.942Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "fce910c9-b3b9-48b9-956c-894736f274f8", "user": "xtrabackup"} 2026-04-29T17:00:25.985Z INFO MySQL init secret created {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "fce910c9-b3b9-48b9-956c-894736f274f8", "secret": "pitr-mysql-init", "user": "xtrabackup"} 2026-04-29T17:00:26.068Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "fce910c9-b3b9-48b9-956c-894736f274f8", "user": "xtrabackup"} 2026-04-29T17:00:26.085Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "fce910c9-b3b9-48b9-956c-894736f274f8", "user": "xtrabackup"} 2026-04-29T17:00:26.089Z INFO PXC pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "fce910c9-b3b9-48b9-956c-894736f274f8", "last-applied-secret": "0b63f10185e0f666a3056d0177462973afcc77b1a6d51bcab5495434c31de258"} 2026-04-29T17:00:26.092Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "fce910c9-b3b9-48b9-956c-894736f274f8", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} &v1.StatefulSet{ TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"}, ObjectMeta: v1.ObjectMeta{ ... // 2 identical fields Namespace: "pitr-pxb-25367", SelfLink: "", - UID: "37c2fe66-113b-4ba7-a0d8-55e1f62aecb7", + UID: "", - ResourceVersion: "1777481819621855023", + ResourceVersion: "", - Generation: 1, + Generation: 0, - CreationTimestamp: v1.Time{Time: s"2026-04-29 16:53:06 +0000 UTC"}, + CreationTimestamp: v1.Time{}, DeletionTimestamp: nil, DeletionGracePeriodSeconds: nil, Labels: nil, - Annotations: map[string]string{ - "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImNjYzgwZmM5NTllMDNjODc0MDFmYWViN2E2YWRlYjc5IiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJmZTEyZDEwNGJhNDg3MjA4ZjhjNGIzNTRkYmYzMDRjZCIsInBlcmNvbmEuY29tL3NzbC1pbnRl"..., - }, + Annotations: map[string]string{ + "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsImxhc3QtYXBwbGllZC1zZWNyZXQiOiIwYjYzZjEwMTg1ZTBmNjY2YTMwNTZkMDE3NzQ2Mjk3M2FmY2M3N2IxYTZkNTFiY2FiNTQ5NTQzNGMzMWRlMjU4IiwicGVyY29uYS5jb20vY29uZmlndXJhdGlvbi1oYXNoIjoiY2NjODBmYzk1OWUwM2M4NzQwMWZhZWI3"..., + }, OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "20559ea9-e17b-447e-a2cb-6b4a334d5aaa", ...}}, Finalizers: nil, - ManagedFields: []v1.ManagedFieldsEntry{ - { - Manager: "percona-xtradb-cluster-operator", - Operation: "Update", - APIVersion: "apps/v1", - Time: s"2026-04-29 16:53:06 +0000 UTC", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., - }, - { - Manager: "kube-controller-manager", - Operation: "Update", - APIVersion: "apps/v1", - Time: s"2026-04-29 16:56:59 +0000 UTC", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., - Subresource: "status", - }, - }, + ManagedFields: nil, }, Spec: v1.StatefulSetSpec{ Replicas: &3, Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, Template: v1.PodTemplateSpec{ ObjectMeta: v1.ObjectMeta{ ... // 9 identical fields DeletionGracePeriodSeconds: nil, Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: map[string]string{ "kubectl.kubernetes.io/default-container": "pxc", + "last-applied-secret": "0b63f10185e0f666a3056d0177462973afcc77b1a6d51bcab5495434c31de258", "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "fe12d104ba487208f8c4b354dbf304cd", "percona.com/ssl-internal-hash": "1cc2c7503eff194d448bab70cf98617a", }, OwnerReferences: nil, Finalizers: nil, ManagedFields: nil, }, Spec: v1.PodSpec{ Volumes: []v1.Volume{ {Name: "tmp", VolumeSource: {EmptyDir: &{}}}, { Name: "config", VolumeSource: v1.VolumeSource{ ... // 16 identical fields FC: nil, AzureFile: nil, ConfigMap: &v1.ConfigMapVolumeSource{ LocalObjectReference: {Name: "pitr-pxc"}, Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, VsphereVolume: nil, Quobyte: nil, ... // 9 identical fields }, }, { Name: "ssl-internal", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-ssl-internal", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "ssl", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-ssl", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &false, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "auto-config", VolumeSource: v1.VolumeSource{ ... // 16 identical fields FC: nil, AzureFile: nil, ConfigMap: &v1.ConfigMapVolumeSource{ LocalObjectReference: {Name: "auto-pitr-pxc"}, Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, VsphereVolume: nil, Quobyte: nil, ... // 9 identical fields }, }, { Name: "vault-keyring-secret", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "some-name-vault", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "mysql-users-secret-file", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "internal-pitr", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &false, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "mysql-init-file", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-mysql-init", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, {Name: "bin", VolumeSource: {EmptyDir: &{}}}, {Name: "backup-logs", VolumeSource: {EmptyDir: &{}}}, }, InitContainers: []v1.Container{ { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, }, Containers: []v1.Container{ { ... // 3 identical fields Args: nil, WorkingDir: "", Ports: []v1.ContainerPort{ { Name: "grpc", HostPort: 0, ContainerPort: 6450, - Protocol: "TCP", + Protocol: "", HostIP: "", }, }, EnvFrom: nil, Env: []v1.EnvVar{ { Name: "POD_NAMESPACE", Value: "", ValueFrom: &v1.EnvVarSource{ FieldRef: &v1.ObjectFieldSelector{ - APIVersion: "v1", + APIVersion: "", FieldPath: "metadata.namespace", }, ResourceFieldRef: nil, ConfigMapKeyRef: nil, ... // 2 identical fields }, }, {Name: "XTRABACKUP_USER_PASS", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "my-cluster-secrets"}, Key: "xtrabackup"}}}, {Name: "VAULT_KEYRING_PATH", Value: "/etc/mysql/vault-keyring-secret/keyring_vault.conf"}, }, Resources: {Limits: {s"cpu": {i: {...}, s: "200m", Format: "DecimalSI"}, s"memory": {i: {...}, Format: "BinarySI"}}, Requests: {s"cpu": {i: {...}, s: "200m", Format: "DecimalSI"}, s"memory": {i: {...}, Format: "BinarySI"}}}, ResizePolicy: nil, ... // 6 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, { ... // 3 identical fields Args: {"mysqld"}, WorkingDir: "", Ports: []v1.ContainerPort{ { Name: "mysql", HostPort: 0, ContainerPort: 3306, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "sst", HostPort: 0, ContainerPort: 4444, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "write-set", HostPort: 0, ContainerPort: 4567, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "ist", HostPort: 0, ContainerPort: 4568, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "mysql-admin", HostPort: 0, ContainerPort: 33062, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "mysqlx", HostPort: 0, ContainerPort: 33060, - Protocol: "TCP", + Protocol: "", HostIP: "", }, }, EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}}, Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...}, ... // 4 identical fields VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...}, VolumeDevices: nil, LivenessProbe: &v1.Probe{ ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}}, InitialDelaySeconds: 300, TimeoutSeconds: 5, - PeriodSeconds: 10, + PeriodSeconds: 0, SuccessThreshold: 1, FailureThreshold: 3, TerminationGracePeriodSeconds: nil, }, ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...}, StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, }, EphemeralContainers: nil, - RestartPolicy: "Always", + RestartPolicy: "", TerminationGracePeriodSeconds: &600, ActiveDeadlineSeconds: nil, - DNSPolicy: "ClusterFirst", + DNSPolicy: "", NodeSelector: nil, ServiceAccountName: "default", - DeprecatedServiceAccount: "default", + DeprecatedServiceAccount: "", AutomountServiceAccountToken: nil, NodeName: "", ... // 7 identical fields Subdomain: "", Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, - SchedulerName: "default-scheduler", + SchedulerName: "", Tolerations: nil, HostAliases: nil, ... // 6 identical fields PreemptionPolicy: nil, Overhead: nil, - TopologySpreadConstraints: nil, + TopologySpreadConstraints: []v1.TopologySpreadConstraint{}, SetHostnameAsFQDN: nil, OS: nil, ... // 6 identical fields }, }, VolumeClaimTemplates: []v1.PersistentVolumeClaim{ { TypeMeta: {}, ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, Spec: v1.PersistentVolumeClaimSpec{ ... // 3 identical fields VolumeName: "", StorageClassName: nil, - VolumeMode: &"Filesystem", + VolumeMode: nil, DataSource: nil, DataSourceRef: nil, VolumeAttributesClassName: nil, }, Status: v1.PersistentVolumeClaimStatus{ - Phase: "Pending", + Phase: "", AccessModes: nil, Capacity: nil, ... // 5 identical fields }, }, }, ServiceName: "pitr-pxc", - PodManagementPolicy: "OrderedReady", + PodManagementPolicy: "", UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, - RevisionHistoryLimit: &10, + RevisionHistoryLimit: nil, MinReadySeconds: 0, - PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", + PersistentVolumeClaimRetentionPolicy: nil, Ordinals: nil, }, Status: v1.StatefulSetStatus{ - ObservedGeneration: 1, + ObservedGeneration: 0, - Replicas: 3, + Replicas: 0, - ReadyReplicas: 3, + ReadyReplicas: 0, - CurrentReplicas: 3, + CurrentReplicas: 0, - UpdatedReplicas: 3, + UpdatedReplicas: 0, - CurrentRevision: "pitr-pxc-79f8874bc7", + CurrentRevision: "", - UpdateRevision: "pitr-pxc-79f8874bc7", + UpdateRevision: "", - CollisionCount: &0, + CollisionCount: nil, Conditions: nil, - AvailableReplicas: 3, + AvailableReplicas: 0, }, } 2026-04-29T17:00:26.901Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "0cb42c39-f921-4fc2-92cd-69a87818cca8"} 2026-04-29T17:01:02.802Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "de741854-6b63-485b-a4c0-8eed25150fa1", "latest": "2026-04-29 17:01:01 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T17:01:32.041Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "79c66102-04f8-4625-ba42-cbdbb710c35e", "err": "failed to ensure cluster readonly status: connect to pod pitr-pxc-1: dial tcp: lookup pitr-pxc-1.pitr-pxc.pitr-pxb-25367 on 34.118.224.10:53: no such host"} [mysql] 2026/04/29 17:02:24 packets.go:58 read tcp 10.79.82.53:36896->10.79.81.49:33062: read: connection reset by peer 2026-04-29T17:02:24.492Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "1dd050ab-7923-41fa-940d-e606999ceec8", "err": "failed to connect to pod pitr-pxc-0: invalid connection"} 2026-04-29T17:02:30.041Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "62530bbe-2753-428f-aeda-fe18d864c552", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-25367.svc.cluster.local"} 2026-04-29T17:02:35.743Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "7b99b956-3068-4885-999d-5139eaa48547", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-25367.svc.cluster.local"} 2026-04-29T17:02:41.514Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "04569979-318f-481d-9278-2aeaa8c26b26", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-25367.svc.cluster.local"} 2026-04-29T17:02:47.244Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "f178b985-2f28-47ca-af93-59652718c3a9", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-25367.svc.cluster.local"} 2026-04-29T17:02:53.006Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "4de1608d-e2f4-4f16-b68b-284f82a6bb44", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-25367.svc.cluster.local"} 2026-04-29T17:02:58.552Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "3bdc6493-d010-4412-ae7e-d1d2b6c1dab1", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-25367.svc.cluster.local"} 2026-04-29T17:02:59.001Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "3bdc6493-d010-4412-ae7e-d1d2b6c1dab1", "latest": "2026-04-29 17:01:34 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T17:03:04.212Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "8239c791-a35d-407d-9276-cc3247c164e4", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-25367.svc.cluster.local"} 2026-04-29T17:03:09.931Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "37dca50a-b4f0-47dd-85ee-5fd4a95f4b44", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-25367.svc.cluster.local"} 2026-04-29T17:03:13.870Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "18328d8c-732a-4804-9a1f-e41a5c3a73a1"} 2026-04-29T17:03:19.064Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "357e9dbd-e177-4010-aee2-5e9a99a7699c"} 2026-04-29T17:03:24.957Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "d5cf62fc-85a6-45de-8e97-da65b0d8138d"} 2026-04-29T17:03:25.585Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "e166dc83-7f67-48c9-8cb1-a225e4a0bd38", "user": "xtrabackup"} 2026-04-29T17:03:25.605Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "e166dc83-7f67-48c9-8cb1-a225e4a0bd38", "user": "xtrabackup"} 2026-04-29T17:03:25.625Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "e166dc83-7f67-48c9-8cb1-a225e4a0bd38", "secret": "pitr-mysql-init", "user": "xtrabackup"} 2026-04-29T17:03:25.644Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "e166dc83-7f67-48c9-8cb1-a225e4a0bd38", "user": "xtrabackup"} 2026-04-29T17:03:25.661Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "e166dc83-7f67-48c9-8cb1-a225e4a0bd38", "user": "xtrabackup"} 2026-04-29T17:03:25.665Z INFO PXC pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "e166dc83-7f67-48c9-8cb1-a225e4a0bd38", "last-applied-secret": "224de9364d91d83b47a9fbccb8a8d94fcf29c1395d3e5f54543d4c345ec7095d"} 2026-04-29T17:03:25.669Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "e166dc83-7f67-48c9-8cb1-a225e4a0bd38", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} &v1.StatefulSet{ TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"}, ObjectMeta: v1.ObjectMeta{ ... // 2 identical fields Namespace: "pitr-pxb-25367", SelfLink: "", - UID: "37c2fe66-113b-4ba7-a0d8-55e1f62aecb7", + UID: "", - ResourceVersion: "1777482190296351023", + ResourceVersion: "", - Generation: 2, + Generation: 0, - CreationTimestamp: v1.Time{Time: s"2026-04-29 16:53:06 +0000 UTC"}, + CreationTimestamp: v1.Time{}, DeletionTimestamp: nil, DeletionGracePeriodSeconds: nil, Labels: nil, - Annotations: map[string]string{ - "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsImxhc3QtYXBwbGllZC1zZWNyZXQiOiIwYjYzZjEwMTg1ZTBmNjY2YTMwNTZkMDE3NzQ2Mjk3M2FmY2M3N2IxYTZkNTFiY2FiNTQ5NTQzNGMzMWRlMjU4IiwicGVyY29uYS5jb20vY29uZmlndXJhdGlvbi1oYXNoIjoiY2NjODBmYzk1OWUwM2M4NzQwMWZhZWI3"..., - }, + Annotations: map[string]string{ + "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsImxhc3QtYXBwbGllZC1zZWNyZXQiOiIyMjRkZTkzNjRkOTFkODNiNDdhOWZiY2NiOGE4ZDk0ZmNmMjljMTM5NWQzZTVmNTQ1NDNkNGMzNDVlYzcwOTVkIiwicGVyY29uYS5jb20vY29uZmlndXJhdGlvbi1oYXNoIjoiY2NjODBmYzk1OWUwM2M4NzQwMWZhZWI3"..., + }, OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "20559ea9-e17b-447e-a2cb-6b4a334d5aaa", ...}}, Finalizers: nil, - ManagedFields: []v1.ManagedFieldsEntry{ - { - Manager: "percona-xtradb-cluster-operator", - Operation: "Update", - APIVersion: "apps/v1", - Time: s"2026-04-29 17:00:26 +0000 UTC", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., - }, - { - Manager: "kube-controller-manager", - Operation: "Update", - APIVersion: "apps/v1", - Time: s"2026-04-29 17:03:10 +0000 UTC", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., - Subresource: "status", - }, - }, + ManagedFields: nil, }, Spec: v1.StatefulSetSpec{ Replicas: &3, Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, Template: v1.PodTemplateSpec{ ObjectMeta: v1.ObjectMeta{ ... // 9 identical fields DeletionGracePeriodSeconds: nil, Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: map[string]string{ "kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": strings.Join({ - "0b63f10185e0f666a3056d0177462973afcc77b1a6d51bcab5495434c31de258", + "224de9364d91d83b47a9fbccb8a8d94fcf29c1395d3e5f54543d4c345ec7095d", }, ""), "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "fe12d104ba487208f8c4b354dbf304cd", "percona.com/ssl-internal-hash": "1cc2c7503eff194d448bab70cf98617a", }, OwnerReferences: nil, Finalizers: nil, ManagedFields: nil, }, Spec: v1.PodSpec{ Volumes: []v1.Volume{ {Name: "tmp", VolumeSource: {EmptyDir: &{}}}, { Name: "config", VolumeSource: v1.VolumeSource{ ... // 16 identical fields FC: nil, AzureFile: nil, ConfigMap: &v1.ConfigMapVolumeSource{ LocalObjectReference: {Name: "pitr-pxc"}, Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, VsphereVolume: nil, Quobyte: nil, ... // 9 identical fields }, }, { Name: "ssl-internal", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-ssl-internal", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "ssl", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-ssl", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &false, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "auto-config", VolumeSource: v1.VolumeSource{ ... // 16 identical fields FC: nil, AzureFile: nil, ConfigMap: &v1.ConfigMapVolumeSource{ LocalObjectReference: {Name: "auto-pitr-pxc"}, Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, VsphereVolume: nil, Quobyte: nil, ... // 9 identical fields }, }, { Name: "vault-keyring-secret", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "some-name-vault", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "mysql-users-secret-file", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "internal-pitr", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &false, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "mysql-init-file", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-mysql-init", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, {Name: "bin", VolumeSource: {EmptyDir: &{}}}, {Name: "backup-logs", VolumeSource: {EmptyDir: &{}}}, }, InitContainers: []v1.Container{ { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, }, Containers: []v1.Container{ { ... // 3 identical fields Args: nil, WorkingDir: "", Ports: []v1.ContainerPort{ { Name: "grpc", HostPort: 0, ContainerPort: 6450, - Protocol: "TCP", + Protocol: "", HostIP: "", }, }, EnvFrom: nil, Env: []v1.EnvVar{ { Name: "POD_NAMESPACE", Value: "", ValueFrom: &v1.EnvVarSource{ FieldRef: &v1.ObjectFieldSelector{ - APIVersion: "v1", + APIVersion: "", FieldPath: "metadata.namespace", }, ResourceFieldRef: nil, ConfigMapKeyRef: nil, ... // 2 identical fields }, }, {Name: "XTRABACKUP_USER_PASS", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "my-cluster-secrets"}, Key: "xtrabackup"}}}, {Name: "VAULT_KEYRING_PATH", Value: "/etc/mysql/vault-keyring-secret/keyring_vault.conf"}, }, Resources: {Limits: {s"cpu": {i: {...}, s: "200m", Format: "DecimalSI"}, s"memory": {i: {...}, Format: "BinarySI"}}, Requests: {s"cpu": {i: {...}, s: "200m", Format: "DecimalSI"}, s"memory": {i: {...}, Format: "BinarySI"}}}, ResizePolicy: nil, ... // 6 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, { ... // 3 identical fields Args: {"mysqld"}, WorkingDir: "", Ports: []v1.ContainerPort{ { Name: "mysql", HostPort: 0, ContainerPort: 3306, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "sst", HostPort: 0, ContainerPort: 4444, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "write-set", HostPort: 0, ContainerPort: 4567, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "ist", HostPort: 0, ContainerPort: 4568, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "mysql-admin", HostPort: 0, ContainerPort: 33062, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "mysqlx", HostPort: 0, ContainerPort: 33060, - Protocol: "TCP", + Protocol: "", HostIP: "", }, }, EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}}, Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...}, ... // 4 identical fields VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...}, VolumeDevices: nil, LivenessProbe: &v1.Probe{ ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}}, InitialDelaySeconds: 300, TimeoutSeconds: 5, - PeriodSeconds: 10, + PeriodSeconds: 0, SuccessThreshold: 1, FailureThreshold: 3, TerminationGracePeriodSeconds: nil, }, ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...}, StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, }, EphemeralContainers: nil, - RestartPolicy: "Always", + RestartPolicy: "", TerminationGracePeriodSeconds: &600, ActiveDeadlineSeconds: nil, - DNSPolicy: "ClusterFirst", + DNSPolicy: "", NodeSelector: nil, ServiceAccountName: "default", - DeprecatedServiceAccount: "default", + DeprecatedServiceAccount: "", AutomountServiceAccountToken: nil, NodeName: "", ... // 7 identical fields Subdomain: "", Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, - SchedulerName: "default-scheduler", + SchedulerName: "", Tolerations: nil, HostAliases: nil, ... // 6 identical fields PreemptionPolicy: nil, Overhead: nil, - TopologySpreadConstraints: nil, + TopologySpreadConstraints: []v1.TopologySpreadConstraint{}, SetHostnameAsFQDN: nil, OS: nil, ... // 6 identical fields }, }, VolumeClaimTemplates: []v1.PersistentVolumeClaim{ { TypeMeta: {}, ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, Spec: v1.PersistentVolumeClaimSpec{ ... // 3 identical fields VolumeName: "", StorageClassName: nil, - VolumeMode: &"Filesystem", + VolumeMode: nil, DataSource: nil, DataSourceRef: nil, VolumeAttributesClassName: nil, }, Status: v1.PersistentVolumeClaimStatus{ - Phase: "Pending", + Phase: "", AccessModes: nil, Capacity: nil, ... // 5 identical fields }, }, }, ServiceName: "pitr-pxc", - PodManagementPolicy: "OrderedReady", + PodManagementPolicy: "", UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, - RevisionHistoryLimit: &10, + RevisionHistoryLimit: nil, MinReadySeconds: 0, - PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", + PersistentVolumeClaimRetentionPolicy: nil, Ordinals: nil, }, Status: v1.StatefulSetStatus{ - ObservedGeneration: 2, + ObservedGeneration: 0, - Replicas: 3, + Replicas: 0, - ReadyReplicas: 3, + ReadyReplicas: 0, - CurrentReplicas: 3, + CurrentReplicas: 0, - UpdatedReplicas: 3, + UpdatedReplicas: 0, - CurrentRevision: "pitr-pxc-b59b4f75b", + CurrentRevision: "", - UpdateRevision: "pitr-pxc-b59b4f75b", + UpdateRevision: "", - CollisionCount: &0, + CollisionCount: nil, Conditions: nil, - AvailableReplicas: 3, + AvailableReplicas: 0, }, } 2026-04-29T17:03:28.669Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "e166dc83-7f67-48c9-8cb1-a225e4a0bd38"} 2026-04-29T17:03:53.982Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "7a68bb82-1c78-43b8-b4dd-2ef62386b8a1", "error": "exec binlog collector pod pitr-pitr-6f99f5b5fc-lsbh2: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-6f99f5b5fc-lsbh2\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:03:56.326Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "8c3f4b96-f56d-4cef-ba54-8977806108d0", "error": "exec binlog collector pod pitr-pitr-6f99f5b5fc-lsbh2: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-6f99f5b5fc-lsbh2\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:03:59.144Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "c28f154f-52f0-494a-ad45-96d60b61cecd", "error": "exec binlog collector pod pitr-pitr-6f99f5b5fc-lsbh2: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-6f99f5b5fc-lsbh2\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:04:07.228Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "fb940810-9448-490a-8f9b-db5d2c73e2f6", "latest": "2026-04-29 17:04:02 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T17:04:30.218Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "84e990a7-eea1-44d8-8bdf-5a74d6278cf0", "err": "failed to ensure cluster readonly status: connect to pod pitr-pxc-1: dial tcp: lookup pitr-pxc-1.pitr-pxc.pitr-pxb-25367 on 34.118.224.10:53: no such host"} 2026-04-29T17:04:36.506Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "3f240171-8006-4189-b5fb-faa1f53803d3", "err": "failed to ensure cluster readonly status: connect to pod pitr-pxc-1: dial tcp: lookup pitr-pxc-1.pitr-pxc.pitr-pxb-25367 on 34.118.224.10:53: no such host"} 2026-04-29T17:05:35.585Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "ba77cc6e-7206-48a6-aab4-43c8b6a40a89", "err": "failed to connect to pod pitr-pxc-0: dial tcp: lookup pitr-pxc-0.pitr-pxc.pitr-pxb-25367 on 34.118.224.10:53: no such host"} 2026-04-29T17:05:41.255Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "ab13b9dd-1b37-48bf-a8d9-f4c4e0953c76", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-25367.svc.cluster.local"} 2026-04-29T17:05:47.076Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "24a89872-a6d6-40ff-b7f0-98c8709a37ea", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-25367.svc.cluster.local"} 2026-04-29T17:05:52.974Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "61d386c6-24f6-4162-b042-a7374c9bc538", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-25367.svc.cluster.local"} 2026-04-29T17:05:58.694Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "4bb55df1-bd0b-48ed-90f6-d3333076518d", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-25367.svc.cluster.local"} 2026-04-29T17:06:04.206Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "345352d7-447a-4e13-8e3c-e9cc351edb86", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-25367.svc.cluster.local"} 2026-04-29T17:06:09.738Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "cbb591ee-9e31-461d-99fb-07c4825e82aa", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-25367.svc.cluster.local"} 2026-04-29T17:06:15.256Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "8ba7db31-0305-4d5c-869d-aa9138b8cbd9", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-25367.svc.cluster.local"} 2026-04-29T17:06:20.961Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "7655c453-aebc-4214-9525-3d8f0d7f7b1b", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-25367.svc.cluster.local"} 2026-04-29T17:06:26.688Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "bedc1be6-8206-4677-aea4-85b4b968c46d", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-25367.svc.cluster.local"} 2026-04-29T17:06:36.225Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "56b577b9-cf86-4a0c-943a-f624229e9404"} 2026-04-29T17:06:41.452Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "5227b1e8-f491-43fb-a4e4-4324ba45c220"} 2026-04-29T17:06:44.804Z INFO Password changed, updating user {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "c4cf1eac-30af-45bf-a618-af386c8fd8f5", "user": "xtrabackup"} 2026-04-29T17:06:44.828Z INFO Password updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "c4cf1eac-30af-45bf-a618-af386c8fd8f5", "user": "xtrabackup"} 2026-04-29T17:06:44.853Z INFO MySQL init secret updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "c4cf1eac-30af-45bf-a618-af386c8fd8f5", "secret": "pitr-mysql-init", "user": "xtrabackup"} 2026-04-29T17:06:44.872Z INFO Internal secrets updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "c4cf1eac-30af-45bf-a618-af386c8fd8f5", "user": "xtrabackup"} 2026-04-29T17:06:44.891Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "c4cf1eac-30af-45bf-a618-af386c8fd8f5", "user": "xtrabackup"} 2026-04-29T17:06:44.894Z INFO PXC pods will be restarted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "c4cf1eac-30af-45bf-a618-af386c8fd8f5", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27"} 2026-04-29T17:06:44.897Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "c4cf1eac-30af-45bf-a618-af386c8fd8f5", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} &v1.StatefulSet{ TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"}, ObjectMeta: v1.ObjectMeta{ ... // 2 identical fields Namespace: "pitr-pxb-25367", SelfLink: "", - UID: "37c2fe66-113b-4ba7-a0d8-55e1f62aecb7", + UID: "", - ResourceVersion: "1777482387587551023", + ResourceVersion: "", - Generation: 3, + Generation: 0, - CreationTimestamp: v1.Time{Time: s"2026-04-29 16:53:06 +0000 UTC"}, + CreationTimestamp: v1.Time{}, DeletionTimestamp: nil, DeletionGracePeriodSeconds: nil, Labels: nil, - Annotations: map[string]string{ - "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsImxhc3QtYXBwbGllZC1zZWNyZXQiOiIyMjRkZTkzNjRkOTFkODNiNDdhOWZiY2NiOGE4ZDk0ZmNmMjljMTM5NWQzZTVmNTQ1NDNkNGMzNDVlYzcwOTVkIiwicGVyY29uYS5jb20vY29uZmlndXJhdGlvbi1oYXNoIjoiY2NjODBmYzk1OWUwM2M4NzQwMWZhZWI3"..., - }, + Annotations: map[string]string{ + "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn19LCJ0ZW1wbGF0ZSI6eyJtZXRhZGF0YSI6eyJsYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRlcy5pby9uYW1lIjoicGVyY29uYS14dHJhZGItY2x1c3RlciIsImFwcC5rdWJlcm5ldGVzLmlvL3BhcnQtb2YiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIn0sImFubm90YXRpb25zIjp7Imt1YmVjdGwua3ViZXJuZXRlcy5pby9kZWZhdWx0LWNvbnRhaW5lciI6InB4YyIsImxhc3QtYXBwbGllZC1zZWNyZXQiOiJkZjhjYmI5YmQ2NzM1MmMxMDIyOGY0ODcwNjkwZmU4MDIwYzE5MzNjNjkxMWY3Yjk0NmM0N2UyYjc2MDgwYjI3IiwicGVyY29uYS5jb20vY29uZmlndXJhdGlvbi1oYXNoIjoiY2NjODBmYzk1OWUwM2M4NzQwMWZhZWI3"..., + }, OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "20559ea9-e17b-447e-a2cb-6b4a334d5aaa", ...}}, Finalizers: nil, - ManagedFields: []v1.ManagedFieldsEntry{ - { - Manager: "percona-xtradb-cluster-operator", - Operation: "Update", - APIVersion: "apps/v1", - Time: s"2026-04-29 17:03:25 +0000 UTC", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., - }, - { - Manager: "kube-controller-manager", - Operation: "Update", - APIVersion: "apps/v1", - Time: s"2026-04-29 17:06:27 +0000 UTC", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., - Subresource: "status", - }, - }, + ManagedFields: nil, }, Spec: v1.StatefulSetSpec{ Replicas: &3, Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, Template: v1.PodTemplateSpec{ ObjectMeta: v1.ObjectMeta{ ... // 9 identical fields DeletionGracePeriodSeconds: nil, Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: map[string]string{ "kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": strings.Join({ - "224de9364d91d83b47a9fbccb8a8d94fcf29c1395d3e5f54543d4c345ec7095d", + "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27", }, ""), "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "fe12d104ba487208f8c4b354dbf304cd", "percona.com/ssl-internal-hash": "1cc2c7503eff194d448bab70cf98617a", }, OwnerReferences: nil, Finalizers: nil, ManagedFields: nil, }, Spec: v1.PodSpec{ Volumes: []v1.Volume{ {Name: "tmp", VolumeSource: {EmptyDir: &{}}}, { Name: "config", VolumeSource: v1.VolumeSource{ ... // 16 identical fields FC: nil, AzureFile: nil, ConfigMap: &v1.ConfigMapVolumeSource{ LocalObjectReference: {Name: "pitr-pxc"}, Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, VsphereVolume: nil, Quobyte: nil, ... // 9 identical fields }, }, { Name: "ssl-internal", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-ssl-internal", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "ssl", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-ssl", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &false, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "auto-config", VolumeSource: v1.VolumeSource{ ... // 16 identical fields FC: nil, AzureFile: nil, ConfigMap: &v1.ConfigMapVolumeSource{ LocalObjectReference: {Name: "auto-pitr-pxc"}, Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, VsphereVolume: nil, Quobyte: nil, ... // 9 identical fields }, }, { Name: "vault-keyring-secret", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "some-name-vault", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "mysql-users-secret-file", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "internal-pitr", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &false, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "mysql-init-file", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-mysql-init", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, {Name: "bin", VolumeSource: {EmptyDir: &{}}}, {Name: "backup-logs", VolumeSource: {EmptyDir: &{}}}, }, InitContainers: []v1.Container{ { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, }, Containers: []v1.Container{ { ... // 3 identical fields Args: nil, WorkingDir: "", Ports: []v1.ContainerPort{ { Name: "grpc", HostPort: 0, ContainerPort: 6450, - Protocol: "TCP", + Protocol: "", HostIP: "", }, }, EnvFrom: nil, Env: []v1.EnvVar{ { Name: "POD_NAMESPACE", Value: "", ValueFrom: &v1.EnvVarSource{ FieldRef: &v1.ObjectFieldSelector{ - APIVersion: "v1", + APIVersion: "", FieldPath: "metadata.namespace", }, ResourceFieldRef: nil, ConfigMapKeyRef: nil, ... // 2 identical fields }, }, {Name: "XTRABACKUP_USER_PASS", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "my-cluster-secrets"}, Key: "xtrabackup"}}}, {Name: "VAULT_KEYRING_PATH", Value: "/etc/mysql/vault-keyring-secret/keyring_vault.conf"}, }, Resources: {Limits: {s"cpu": {i: {...}, s: "200m", Format: "DecimalSI"}, s"memory": {i: {...}, Format: "BinarySI"}}, Requests: {s"cpu": {i: {...}, s: "200m", Format: "DecimalSI"}, s"memory": {i: {...}, Format: "BinarySI"}}}, ResizePolicy: nil, ... // 6 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, { ... // 3 identical fields Args: {"mysqld"}, WorkingDir: "", Ports: []v1.ContainerPort{ { Name: "mysql", HostPort: 0, ContainerPort: 3306, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "sst", HostPort: 0, ContainerPort: 4444, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "write-set", HostPort: 0, ContainerPort: 4567, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "ist", HostPort: 0, ContainerPort: 4568, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "mysql-admin", HostPort: 0, ContainerPort: 33062, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "mysqlx", HostPort: 0, ContainerPort: 33060, - Protocol: "TCP", + Protocol: "", HostIP: "", }, }, EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}}, Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...}, ... // 4 identical fields VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...}, VolumeDevices: nil, LivenessProbe: &v1.Probe{ ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}}, InitialDelaySeconds: 300, TimeoutSeconds: 5, - PeriodSeconds: 10, + PeriodSeconds: 0, SuccessThreshold: 1, FailureThreshold: 3, TerminationGracePeriodSeconds: nil, }, ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...}, StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, }, EphemeralContainers: nil, - RestartPolicy: "Always", + RestartPolicy: "", TerminationGracePeriodSeconds: &600, ActiveDeadlineSeconds: nil, - DNSPolicy: "ClusterFirst", + DNSPolicy: "", NodeSelector: nil, ServiceAccountName: "default", - DeprecatedServiceAccount: "default", + DeprecatedServiceAccount: "", AutomountServiceAccountToken: nil, NodeName: "", ... // 7 identical fields Subdomain: "", Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, - SchedulerName: "default-scheduler", + SchedulerName: "", Tolerations: nil, HostAliases: nil, ... // 6 identical fields PreemptionPolicy: nil, Overhead: nil, - TopologySpreadConstraints: nil, + TopologySpreadConstraints: []v1.TopologySpreadConstraint{}, SetHostnameAsFQDN: nil, OS: nil, ... // 6 identical fields }, }, VolumeClaimTemplates: []v1.PersistentVolumeClaim{ { TypeMeta: {}, ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, Spec: v1.PersistentVolumeClaimSpec{ ... // 3 identical fields VolumeName: "", StorageClassName: nil, - VolumeMode: &"Filesystem", + VolumeMode: nil, DataSource: nil, DataSourceRef: nil, VolumeAttributesClassName: nil, }, Status: v1.PersistentVolumeClaimStatus{ - Phase: "Pending", + Phase: "", AccessModes: nil, Capacity: nil, ... // 5 identical fields }, }, }, ServiceName: "pitr-pxc", - PodManagementPolicy: "OrderedReady", + PodManagementPolicy: "", UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, - RevisionHistoryLimit: &10, + RevisionHistoryLimit: nil, MinReadySeconds: 0, - PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", + PersistentVolumeClaimRetentionPolicy: nil, Ordinals: nil, }, Status: v1.StatefulSetStatus{ - ObservedGeneration: 3, + ObservedGeneration: 0, - Replicas: 3, + Replicas: 0, - ReadyReplicas: 3, + ReadyReplicas: 0, - CurrentReplicas: 3, + CurrentReplicas: 0, - UpdatedReplicas: 3, + UpdatedReplicas: 0, - CurrentRevision: "pitr-pxc-79558f96fd", + CurrentRevision: "", - UpdateRevision: "pitr-pxc-79558f96fd", + UpdateRevision: "", - CollisionCount: &0, + CollisionCount: nil, Conditions: nil, - AvailableReplicas: 3, + AvailableReplicas: 0, }, } 2026-04-29T17:06:47.504Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "c4cf1eac-30af-45bf-a618-af386c8fd8f5"} 2026-04-29T17:07:17.596Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "9ae9d928-bc6d-4df5-ad9d-e9c7d08a409f", "error": "exec binlog collector pod pitr-pitr-6f99f5b5fc-lsbh2: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-6f99f5b5fc-lsbh2\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:07:20.269Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "89f53f9f-5545-459d-a88e-38e4df9af566", "error": "exec binlog collector pod pitr-pitr-6f99f5b5fc-lsbh2: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-6f99f5b5fc-lsbh2\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:07:23.151Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "419b7a32-b6c5-47dd-a629-d4667522828c", "error": "exec binlog collector pod pitr-pitr-6f99f5b5fc-lsbh2: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-6f99f5b5fc-lsbh2\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:07:25.705Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "2707b56e-46f3-44be-986a-d5fcc768bcb9", "error": "exec binlog collector pod pitr-pitr-6f99f5b5fc-lsbh2: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-6f99f5b5fc-lsbh2\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:07:28.705Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "00bf6d07-7613-4a97-9c53-2f2f8c423ecf", "error": "exec binlog collector pod pitr-pitr-6f99f5b5fc-lsbh2: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-6f99f5b5fc-lsbh2\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:07:31.298Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "1e12977e-ce5a-4e99-ab5c-82bc56f799af", "error": "exec binlog collector pod pitr-pitr-6f99f5b5fc-lsbh2: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-6f99f5b5fc-lsbh2\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:07:34.029Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "d95ffc1b-7c07-4922-a8b7-0c042a7d4f29", "error": "exec binlog collector pod pitr-pitr-6f99f5b5fc-lsbh2: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-6f99f5b5fc-lsbh2\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:07:36.586Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "537418ee-e547-4667-ad3a-ba7de0161043", "error": "exec binlog collector pod pitr-pitr-6f99f5b5fc-lsbh2: failed to execute command in pod: unable to upgrade connection: container not found (\"pitr\")", "errorVerbose": "unable to upgrade connection: container not found (\"pitr\")\nfailed to execute command in pod\ngithub.com/percona/percona-xtradb-cluster-operator/clientcmd.(*Client).Exec\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/clientcmd/clientcmd.go:141\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:148\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nexec binlog collector pod pitr-pitr-6f99f5b5fc-lsbh2\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup.UpdatePITRTimeline\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/backup/pitr.go:150\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:464\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:222\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:479\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:495 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:438 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.23.1/pkg/internal/controller/controller.go:313 2026-04-29T17:07:47.227Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "edba2ba5-5170-4373-9e58-7d03b64e1c31", "latest": "2026-04-29 17:07:41 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T17:07:59.236Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "38bf4280-3c19-4d89-9711-f6cbf1dbeaca", "err": "failed to ensure cluster readonly status: connect to pod pitr-pxc-1: dial tcp 10.79.80.43:33062: connect: connection refused"} 2026-04-29T17:08:51.312Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "e69dac17-53c2-4191-b6d6-56fcd38c2867", "err": "failed to connect to pod pitr-pxc-0: dial tcp: lookup pitr-pxc-0.pitr-pxc.pitr-pxb-25367 on 34.118.224.10:53: no such host"} 2026-04-29T17:09:03.723Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "4123e149-98e8-4242-a0e9-852d41f1a1cf", "err": "failed to connect to pod pitr-pxc-0: dial tcp 10.79.81.53:33062: connect: connection refused"} 2026-04-29T17:09:09.526Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "f31e117d-d226-4d5f-afb5-746d663e7b9c", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-25367.svc.cluster.local"} 2026-04-29T17:09:15.226Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "7c4e920f-7734-4044-82ed-8bf4dae78b19", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-25367.svc.cluster.local"} 2026-04-29T17:09:20.939Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "62efa9bd-34f7-4287-a9b2-20a502842bbd", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-25367.svc.cluster.local"} 2026-04-29T17:09:26.654Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "128c6092-ca9b-4804-9993-a4fa3b4bde29", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-25367.svc.cluster.local"} 2026-04-29T17:09:32.424Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "39ecdafe-f05b-4efc-b2ee-a93f73c3dde7", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-25367.svc.cluster.local"} 2026-04-29T17:09:38.149Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "cb6b4167-5b21-4fba-875b-ab57bcd85c1b", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-25367.svc.cluster.local"} 2026-04-29T17:09:43.852Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "49ebaf3e-7aab-449d-9b9a-41d16906d42e", "primary name": "pitr-pxc-0.pitr-pxc.pitr-pxb-25367.svc.cluster.local"} 2026-04-29T17:09:53.303Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "5d00663d-9de4-41b3-b375-a8fb94ed72c2"} 2026-04-29T17:09:58.738Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "ede7ecb2-54c5-4f15-8bdc-944a15293253"} 2026-04-29T17:10:04.542Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "526e0aca-4f6a-4310-98ef-3d978ec945c7"} 2026-04-29T17:10:10.456Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "246cccb7-e5ca-4fcd-910b-feaeab870b0e"} 2026-04-29T17:10:16.244Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "38888d86-1f6b-4d18-8edb-4938ffaae32d"} 2026-04-29T17:10:22.037Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "ba7a8b4e-ea3d-467c-8dd0-2d3d482659bd"} 2026-04-29T17:10:28.152Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "191ca002-ffc7-4aef-b44f-d30751d0fdcc"} 2026-04-29T17:10:34.240Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "5d65c933-a099-4383-b66e-4ac63c475b08"} 2026-04-29T17:10:40.061Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "d19cae01-f196-4fcd-8446-320f974251bd"} 2026-04-29T17:10:46.048Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "903371de-cfec-4727-b883-0fc328328aca"} 2026-04-29T17:10:51.872Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "bb68f11b-05a4-498c-b175-a7fd7c06271b"} 2026-04-29T17:10:57.927Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "c93b5b0e-104f-441d-aa13-677100010d41"} 2026-04-29T17:11:03.516Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "9737a07a-2a14-45b2-ba34-508e263a11e9"} 2026-04-29T17:11:09.843Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "b1c92e56-97a5-4308-ab21-9ca5b80db8ce"} 2026-04-29T17:11:15.724Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "a9e8af05-dac0-48bb-ac34-dc8773cb5e86"} 2026-04-29T17:11:21.553Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "ae0c1f63-2d3b-4fa0-8118-1b4e0b281cf5"} 2026-04-29T17:11:27.362Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "a23f0ce7-fd96-4a0e-ba99-fb01155e0485"} 2026-04-29T17:11:32.353Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "f27d3662-882e-428f-9fd0-bb744716a0db", "latest": "2026-04-29 17:11:26 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T17:11:33.544Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "f27d3662-882e-428f-9fd0-bb744716a0db"} 2026-04-29T17:11:39.516Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "e2826250-0a14-454b-814b-ae1a09c23073"} 2026-04-29T17:11:45.550Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "b872e09d-24c0-497e-be21-13948bad07b2"} 2026-04-29T17:11:50.038Z INFO stopping cluster {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "9d1710c1-83d8-4f8c-bf24-f3098ca36217", "cluster": "pitr"} 2026-04-29T17:11:50.470Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "f0f688e2-5aa4-4577-b51a-fa237eaf7512", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} &v1.StatefulSet{ TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"}, ObjectMeta: v1.ObjectMeta{ ... // 2 identical fields Namespace: "pitr-pxb-25367", SelfLink: "", - UID: "37c2fe66-113b-4ba7-a0d8-55e1f62aecb7", + UID: "", - ResourceVersion: "1777482588903583023", + ResourceVersion: "", - Generation: 4, + Generation: 0, - CreationTimestamp: v1.Time{Time: s"2026-04-29 16:53:06 +0000 UTC"}, + CreationTimestamp: v1.Time{}, DeletionTimestamp: nil, DeletionGracePeriodSeconds: nil, Labels: nil, - Annotations: map[string]string{ - "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., - }, + Annotations: map[string]string{ + "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., + }, OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "20559ea9-e17b-447e-a2cb-6b4a334d5aaa", ...}}, Finalizers: nil, - ManagedFields: []v1.ManagedFieldsEntry{ - { - Manager: "percona-xtradb-cluster-operator", - Operation: "Update", - APIVersion: "apps/v1", - Time: s"2026-04-29 17:06:44 +0000 UTC", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., - }, - { - Manager: "kube-controller-manager", - Operation: "Update", - APIVersion: "apps/v1", - Time: s"2026-04-29 17:09:48 +0000 UTC", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., - Subresource: "status", - }, - }, + ManagedFields: nil, }, Spec: v1.StatefulSetSpec{ - Replicas: &3, + Replicas: &0, Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, Template: v1.PodTemplateSpec{ ObjectMeta: {Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27", "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "fe12d104ba487208f8c4b354dbf304cd", ...}}, Spec: v1.PodSpec{ Volumes: []v1.Volume{ {Name: "tmp", VolumeSource: {EmptyDir: &{}}}, { Name: "config", VolumeSource: v1.VolumeSource{ ... // 16 identical fields FC: nil, AzureFile: nil, ConfigMap: &v1.ConfigMapVolumeSource{ LocalObjectReference: {Name: "pitr-pxc"}, Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, VsphereVolume: nil, Quobyte: nil, ... // 9 identical fields }, }, { Name: "ssl-internal", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-ssl-internal", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "ssl", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-ssl", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &false, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "auto-config", VolumeSource: v1.VolumeSource{ ... // 16 identical fields FC: nil, AzureFile: nil, ConfigMap: &v1.ConfigMapVolumeSource{ LocalObjectReference: {Name: "auto-pitr-pxc"}, Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, VsphereVolume: nil, Quobyte: nil, ... // 9 identical fields }, }, { Name: "vault-keyring-secret", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "some-name-vault", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "mysql-users-secret-file", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "internal-pitr", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &false, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "mysql-init-file", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-mysql-init", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, {Name: "bin", VolumeSource: {EmptyDir: &{}}}, {Name: "backup-logs", VolumeSource: {EmptyDir: &{}}}, }, InitContainers: []v1.Container{ { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, }, Containers: []v1.Container{ { ... // 3 identical fields Args: nil, WorkingDir: "", Ports: []v1.ContainerPort{ { Name: "grpc", HostPort: 0, ContainerPort: 6450, - Protocol: "TCP", + Protocol: "", HostIP: "", }, }, EnvFrom: nil, Env: []v1.EnvVar{ { Name: "POD_NAMESPACE", Value: "", ValueFrom: &v1.EnvVarSource{ FieldRef: &v1.ObjectFieldSelector{ - APIVersion: "v1", + APIVersion: "", FieldPath: "metadata.namespace", }, ResourceFieldRef: nil, ConfigMapKeyRef: nil, ... // 2 identical fields }, }, {Name: "XTRABACKUP_USER_PASS", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "my-cluster-secrets"}, Key: "xtrabackup"}}}, {Name: "VAULT_KEYRING_PATH", Value: "/etc/mysql/vault-keyring-secret/keyring_vault.conf"}, }, Resources: {Limits: {s"cpu": {i: {...}, s: "200m", Format: "DecimalSI"}, s"memory": {i: {...}, Format: "BinarySI"}}, Requests: {s"cpu": {i: {...}, s: "200m", Format: "DecimalSI"}, s"memory": {i: {...}, Format: "BinarySI"}}}, ResizePolicy: nil, ... // 6 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, { ... // 3 identical fields Args: {"mysqld"}, WorkingDir: "", Ports: []v1.ContainerPort{ { Name: "mysql", HostPort: 0, ContainerPort: 3306, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "sst", HostPort: 0, ContainerPort: 4444, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "write-set", HostPort: 0, ContainerPort: 4567, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "ist", HostPort: 0, ContainerPort: 4568, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "mysql-admin", HostPort: 0, ContainerPort: 33062, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "mysqlx", HostPort: 0, ContainerPort: 33060, - Protocol: "TCP", + Protocol: "", HostIP: "", }, }, EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}}, Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...}, ... // 4 identical fields VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...}, VolumeDevices: nil, LivenessProbe: &v1.Probe{ ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}}, InitialDelaySeconds: 300, TimeoutSeconds: 5, - PeriodSeconds: 10, + PeriodSeconds: 0, SuccessThreshold: 1, FailureThreshold: 3, TerminationGracePeriodSeconds: nil, }, ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...}, StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, }, EphemeralContainers: nil, - RestartPolicy: "Always", + RestartPolicy: "", TerminationGracePeriodSeconds: &600, ActiveDeadlineSeconds: nil, - DNSPolicy: "ClusterFirst", + DNSPolicy: "", NodeSelector: nil, ServiceAccountName: "default", - DeprecatedServiceAccount: "default", + DeprecatedServiceAccount: "", AutomountServiceAccountToken: nil, NodeName: "", ... // 7 identical fields Subdomain: "", Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, - SchedulerName: "default-scheduler", + SchedulerName: "", Tolerations: nil, HostAliases: nil, ... // 6 identical fields PreemptionPolicy: nil, Overhead: nil, - TopologySpreadConstraints: nil, + TopologySpreadConstraints: []v1.TopologySpreadConstraint{}, SetHostnameAsFQDN: nil, OS: nil, ... // 6 identical fields }, }, VolumeClaimTemplates: []v1.PersistentVolumeClaim{ { TypeMeta: {}, ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, Spec: v1.PersistentVolumeClaimSpec{ ... // 3 identical fields VolumeName: "", StorageClassName: nil, - VolumeMode: &"Filesystem", + VolumeMode: nil, DataSource: nil, DataSourceRef: nil, VolumeAttributesClassName: nil, }, Status: v1.PersistentVolumeClaimStatus{ - Phase: "Pending", + Phase: "", AccessModes: nil, Capacity: nil, ... // 5 identical fields }, }, }, ServiceName: "pitr-pxc", - PodManagementPolicy: "OrderedReady", + PodManagementPolicy: "", UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, - RevisionHistoryLimit: &10, + RevisionHistoryLimit: nil, MinReadySeconds: 0, - PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", + PersistentVolumeClaimRetentionPolicy: nil, Ordinals: nil, }, Status: v1.StatefulSetStatus{ - ObservedGeneration: 4, + ObservedGeneration: 0, - Replicas: 3, + Replicas: 0, - ReadyReplicas: 3, + ReadyReplicas: 0, - CurrentReplicas: 3, + CurrentReplicas: 0, - UpdatedReplicas: 3, + UpdatedReplicas: 0, - CurrentRevision: "pitr-pxc-5b4fdc58c4", + CurrentRevision: "", - UpdateRevision: "pitr-pxc-5b4fdc58c4", + UpdateRevision: "", - CollisionCount: &0, + CollisionCount: nil, Conditions: nil, - AvailableReplicas: 3, + AvailableReplicas: 0, }, } 2026-04-29T17:11:50.542Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "f0f688e2-5aa4-4577-b51a-fa237eaf7512", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} &v1.StatefulSet{ TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"}, ObjectMeta: v1.ObjectMeta{ ... // 2 identical fields Namespace: "pitr-pxb-25367", SelfLink: "", - UID: "a6ffacfb-86f2-42c6-8055-973883f42b38", + UID: "", - ResourceVersion: "1777482320636991013", + ResourceVersion: "", - Generation: 1, + Generation: 0, - CreationTimestamp: v1.Time{Time: s"2026-04-29 16:53:06 +0000 UTC"}, + CreationTimestamp: v1.Time{}, DeletionTimestamp: nil, DeletionGracePeriodSeconds: nil, Labels: nil, - Annotations: map[string]string{ - "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., - }, + Annotations: map[string]string{ + "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., + }, OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "20559ea9-e17b-447e-a2cb-6b4a334d5aaa", ...}}, Finalizers: nil, - ManagedFields: []v1.ManagedFieldsEntry{ - { - Manager: "percona-xtradb-cluster-operator", - Operation: "Update", - APIVersion: "apps/v1", - Time: s"2026-04-29 16:53:06 +0000 UTC", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., - }, - { - Manager: "kube-controller-manager", - Operation: "Update", - APIVersion: "apps/v1", - Time: s"2026-04-29 17:05:20 +0000 UTC", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., - Subresource: "status", - }, - }, + ManagedFields: nil, }, Spec: v1.StatefulSetSpec{ - Replicas: &2, + Replicas: &0, Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, Template: v1.PodTemplateSpec{ ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "fe12d104ba487208f8c4b354dbf304cd", "percona.com/ssl-internal-hash": "1cc2c7503eff194d448bab70cf98617a"}}, Spec: v1.PodSpec{ Volumes: []v1.Volume{ { Name: "ssl-internal", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-ssl-internal", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "ssl", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-ssl", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &false, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, {Name: "bin", VolumeSource: {EmptyDir: &{}}}, }, InitContainers: []v1.Container{ { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, }, Containers: []v1.Container{ { ... // 3 identical fields Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...}, WorkingDir: "", Ports: []v1.ContainerPort{ { Name: "mysql", HostPort: 0, ContainerPort: 3306, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "proxyadm", HostPort: 0, ContainerPort: 6032, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "stats", HostPort: 0, ContainerPort: 6070, - Protocol: "TCP", + Protocol: "", HostIP: "", }, }, EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}}, Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...}, ... // 8 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, }, EphemeralContainers: nil, - RestartPolicy: "Always", + RestartPolicy: "", TerminationGracePeriodSeconds: &30, ActiveDeadlineSeconds: nil, - DNSPolicy: "ClusterFirst", + DNSPolicy: "", NodeSelector: nil, ServiceAccountName: "default", - DeprecatedServiceAccount: "default", + DeprecatedServiceAccount: "", AutomountServiceAccountToken: nil, NodeName: "", ... // 7 identical fields Subdomain: "", Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, - SchedulerName: "default-scheduler", + SchedulerName: "", Tolerations: nil, HostAliases: nil, ... // 6 identical fields PreemptionPolicy: nil, Overhead: nil, - TopologySpreadConstraints: nil, + TopologySpreadConstraints: []v1.TopologySpreadConstraint{}, SetHostnameAsFQDN: nil, OS: nil, ... // 6 identical fields }, }, VolumeClaimTemplates: []v1.PersistentVolumeClaim{ { TypeMeta: {}, ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, Spec: v1.PersistentVolumeClaimSpec{ ... // 3 identical fields VolumeName: "", StorageClassName: nil, - VolumeMode: &"Filesystem", + VolumeMode: nil, DataSource: nil, DataSourceRef: nil, VolumeAttributesClassName: nil, }, Status: v1.PersistentVolumeClaimStatus{ - Phase: "Pending", + Phase: "", AccessModes: nil, Capacity: nil, ... // 5 identical fields }, }, }, ServiceName: "pitr-proxysql-unready", - PodManagementPolicy: "OrderedReady", + PodManagementPolicy: "", UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, - RevisionHistoryLimit: &10, + RevisionHistoryLimit: nil, MinReadySeconds: 0, - PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", + PersistentVolumeClaimRetentionPolicy: nil, Ordinals: nil, }, Status: v1.StatefulSetStatus{ - ObservedGeneration: 1, + ObservedGeneration: 0, - Replicas: 2, + Replicas: 0, - ReadyReplicas: 2, + ReadyReplicas: 0, - CurrentReplicas: 2, + CurrentReplicas: 0, - UpdatedReplicas: 2, + UpdatedReplicas: 0, - CurrentRevision: "pitr-proxysql-9cb4d4785", + CurrentRevision: "", - UpdateRevision: "pitr-proxysql-9cb4d4785", + UpdateRevision: "", - CollisionCount: &0, + CollisionCount: nil, Conditions: nil, - AvailableReplicas: 2, + AvailableReplicas: 0, }, } 2026-04-29T17:11:50.621Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "f0f688e2-5aa4-4577-b51a-fa237eaf7512", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} &v1.StatefulSet{ TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"}, ObjectMeta: v1.ObjectMeta{ ... // 2 identical fields Namespace: "pitr-pxb-25367", SelfLink: "", - UID: "a6ffacfb-86f2-42c6-8055-973883f42b38", + UID: "", - ResourceVersion: "1777482320636991013", + ResourceVersion: "", - Generation: 1, + Generation: 0, - CreationTimestamp: v1.Time{Time: s"2026-04-29 16:53:06 +0000 UTC"}, + CreationTimestamp: v1.Time{}, DeletionTimestamp: nil, DeletionGracePeriodSeconds: nil, Labels: nil, - Annotations: map[string]string{ - "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., - }, + Annotations: map[string]string{ + "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., + }, OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "20559ea9-e17b-447e-a2cb-6b4a334d5aaa", ...}}, Finalizers: nil, - ManagedFields: []v1.ManagedFieldsEntry{ - { - Manager: "percona-xtradb-cluster-operator", - Operation: "Update", - APIVersion: "apps/v1", - Time: s"2026-04-29 16:53:06 +0000 UTC", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., - }, - { - Manager: "kube-controller-manager", - Operation: "Update", - APIVersion: "apps/v1", - Time: s"2026-04-29 17:05:20 +0000 UTC", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., - Subresource: "status", - }, - }, + ManagedFields: nil, }, Spec: v1.StatefulSetSpec{ - Replicas: &2, + Replicas: &0, Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, Template: v1.PodTemplateSpec{ ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "fe12d104ba487208f8c4b354dbf304cd", "percona.com/ssl-internal-hash": "1cc2c7503eff194d448bab70cf98617a"}}, Spec: v1.PodSpec{ Volumes: []v1.Volume{ { Name: "ssl-internal", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-ssl-internal", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "ssl", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-ssl", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &false, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, {Name: "bin", VolumeSource: {EmptyDir: &{}}}, }, InitContainers: []v1.Container{ { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, }, Containers: []v1.Container{ { ... // 3 identical fields Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...}, WorkingDir: "", Ports: []v1.ContainerPort{ { Name: "mysql", HostPort: 0, ContainerPort: 3306, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "proxyadm", HostPort: 0, ContainerPort: 6032, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "stats", HostPort: 0, ContainerPort: 6070, - Protocol: "TCP", + Protocol: "", HostIP: "", }, }, EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}}, Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...}, ... // 8 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, }, EphemeralContainers: nil, - RestartPolicy: "Always", + RestartPolicy: "", TerminationGracePeriodSeconds: &30, ActiveDeadlineSeconds: nil, - DNSPolicy: "ClusterFirst", + DNSPolicy: "", NodeSelector: nil, ServiceAccountName: "default", - DeprecatedServiceAccount: "default", + DeprecatedServiceAccount: "", AutomountServiceAccountToken: nil, NodeName: "", ... // 7 identical fields Subdomain: "", Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, - SchedulerName: "default-scheduler", + SchedulerName: "", Tolerations: nil, HostAliases: nil, ... // 6 identical fields PreemptionPolicy: nil, Overhead: nil, - TopologySpreadConstraints: nil, + TopologySpreadConstraints: []v1.TopologySpreadConstraint{}, SetHostnameAsFQDN: nil, OS: nil, ... // 6 identical fields }, }, VolumeClaimTemplates: []v1.PersistentVolumeClaim{ { TypeMeta: {}, ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, Spec: v1.PersistentVolumeClaimSpec{ ... // 3 identical fields VolumeName: "", StorageClassName: nil, - VolumeMode: &"Filesystem", + VolumeMode: nil, DataSource: nil, DataSourceRef: nil, VolumeAttributesClassName: nil, }, Status: v1.PersistentVolumeClaimStatus{ - Phase: "Pending", + Phase: "", AccessModes: nil, Capacity: nil, ... // 5 identical fields }, }, }, ServiceName: "pitr-proxysql-unready", - PodManagementPolicy: "OrderedReady", + PodManagementPolicy: "", UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, - RevisionHistoryLimit: &10, + RevisionHistoryLimit: nil, MinReadySeconds: 0, - PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", + PersistentVolumeClaimRetentionPolicy: nil, Ordinals: nil, }, Status: v1.StatefulSetStatus{ - ObservedGeneration: 1, + ObservedGeneration: 0, - Replicas: 2, + Replicas: 0, - ReadyReplicas: 2, + ReadyReplicas: 0, - CurrentReplicas: 2, + CurrentReplicas: 0, - UpdatedReplicas: 2, + UpdatedReplicas: 0, - CurrentRevision: "pitr-proxysql-9cb4d4785", + CurrentRevision: "", - UpdateRevision: "pitr-proxysql-9cb4d4785", + UpdateRevision: "", - CollisionCount: &0, + CollisionCount: nil, Conditions: nil, - AvailableReplicas: 2, + AvailableReplicas: 0, }, } 2026-04-29T17:11:51.932Z ERROR sync users {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "8c028556-a585-4df1-ba28-b6ac46849e2d", "error": "exec syncusers: failed to execute command in pod: unable to upgrade connection: container not found (\"proxysql\") / / ", "errorVerbose": "exec syncusers: failed to execute command in pod: unable to upgrade connection: container not found (\"proxysql\") / / \ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:993\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:832\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1 /go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:834 2026-04-29T17:12:32.224Z INFO starting restore {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "54239d22-dbf9-4811-8de6-c10a2562160e", "cluster": "pitr", "backup": "on-pitr-minio"} 2026-04-29T17:12:32.325Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "4086af81-5421-4501-a278-7e8c8f6c811c", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:12:37.373Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "0010c8fe-5733-492c-994e-5071e6153510", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:12:42.424Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "f328ee57-97a8-4f1a-a5c5-6fcf484edccd", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:12:47.440Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "4f50b51f-c141-4be9-825b-9fd28935e800", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:12:52.511Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "37d5108f-3ace-48be-bb35-de577ad62347", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:12:57.530Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "729a61c5-5c48-4ef8-a1a9-ecc3ed0d423e", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:13:02.547Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "f6c2c04d-95e2-42a2-994d-87982feb499b", "job": "restore-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:13:07.676Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "079d83f8-ed8e-4c3d-99fa-4b032f6296ed", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:13:07.676Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "079d83f8-ed8e-4c3d-99fa-4b032f6296ed", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:07.677Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "079d83f8-ed8e-4c3d-99fa-4b032f6296ed", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:07.901Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "50f630f4-2133-4153-b53b-f36db71b33d7", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} &v1.StatefulSet{ TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"}, ObjectMeta: v1.ObjectMeta{ ... // 2 identical fields Namespace: "pitr-pxb-25367", SelfLink: "", - UID: "37c2fe66-113b-4ba7-a0d8-55e1f62aecb7", + UID: "", - ResourceVersion: "1777482750227727023", + ResourceVersion: "", - Generation: 5, + Generation: 0, - CreationTimestamp: v1.Time{Time: s"2026-04-29 16:53:06 +0000 UTC"}, + CreationTimestamp: v1.Time{}, DeletionTimestamp: nil, DeletionGracePeriodSeconds: nil, Labels: nil, - Annotations: map[string]string{ - "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., - }, + Annotations: map[string]string{ + "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MSwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., + }, OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "20559ea9-e17b-447e-a2cb-6b4a334d5aaa", ...}}, Finalizers: nil, - ManagedFields: []v1.ManagedFieldsEntry{ - { - Manager: "percona-xtradb-cluster-operator", - Operation: "Update", - APIVersion: "apps/v1", - Time: s"2026-04-29 17:11:50 +0000 UTC", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., - }, - { - Manager: "kube-controller-manager", - Operation: "Update", - APIVersion: "apps/v1", - Time: s"2026-04-29 17:12:30 +0000 UTC", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., - Subresource: "status", - }, - }, + ManagedFields: nil, }, Spec: v1.StatefulSetSpec{ - Replicas: &0, + Replicas: &1, Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, Template: v1.PodTemplateSpec{ ObjectMeta: {Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27", "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "fe12d104ba487208f8c4b354dbf304cd", ...}}, Spec: v1.PodSpec{ Volumes: []v1.Volume{ {Name: "tmp", VolumeSource: {EmptyDir: &{}}}, { Name: "config", VolumeSource: v1.VolumeSource{ ... // 16 identical fields FC: nil, AzureFile: nil, ConfigMap: &v1.ConfigMapVolumeSource{ LocalObjectReference: {Name: "pitr-pxc"}, Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, VsphereVolume: nil, Quobyte: nil, ... // 9 identical fields }, }, { Name: "ssl-internal", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-ssl-internal", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "ssl", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-ssl", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &false, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "auto-config", VolumeSource: v1.VolumeSource{ ... // 16 identical fields FC: nil, AzureFile: nil, ConfigMap: &v1.ConfigMapVolumeSource{ LocalObjectReference: {Name: "auto-pitr-pxc"}, Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, VsphereVolume: nil, Quobyte: nil, ... // 9 identical fields }, }, { Name: "vault-keyring-secret", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "some-name-vault", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "mysql-users-secret-file", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "internal-pitr", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &false, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "mysql-init-file", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-mysql-init", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, {Name: "bin", VolumeSource: {EmptyDir: &{}}}, {Name: "backup-logs", VolumeSource: {EmptyDir: &{}}}, }, InitContainers: []v1.Container{ { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, }, Containers: []v1.Container{ { ... // 3 identical fields Args: nil, WorkingDir: "", Ports: []v1.ContainerPort{ { Name: "grpc", HostPort: 0, ContainerPort: 6450, - Protocol: "TCP", + Protocol: "", HostIP: "", }, }, EnvFrom: nil, Env: []v1.EnvVar{ { Name: "POD_NAMESPACE", Value: "", ValueFrom: &v1.EnvVarSource{ FieldRef: &v1.ObjectFieldSelector{ - APIVersion: "v1", + APIVersion: "", FieldPath: "metadata.namespace", }, ResourceFieldRef: nil, ConfigMapKeyRef: nil, ... // 2 identical fields }, }, {Name: "XTRABACKUP_USER_PASS", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "my-cluster-secrets"}, Key: "xtrabackup"}}}, {Name: "VAULT_KEYRING_PATH", Value: "/etc/mysql/vault-keyring-secret/keyring_vault.conf"}, }, Resources: {Limits: {s"cpu": {i: {...}, s: "200m", Format: "DecimalSI"}, s"memory": {i: {...}, Format: "BinarySI"}}, Requests: {s"cpu": {i: {...}, s: "200m", Format: "DecimalSI"}, s"memory": {i: {...}, Format: "BinarySI"}}}, ResizePolicy: nil, ... // 6 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, { ... // 3 identical fields Args: {"mysqld"}, WorkingDir: "", Ports: []v1.ContainerPort{ { Name: "mysql", HostPort: 0, ContainerPort: 3306, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "sst", HostPort: 0, ContainerPort: 4444, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "write-set", HostPort: 0, ContainerPort: 4567, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "ist", HostPort: 0, ContainerPort: 4568, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "mysql-admin", HostPort: 0, ContainerPort: 33062, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "mysqlx", HostPort: 0, ContainerPort: 33060, - Protocol: "TCP", + Protocol: "", HostIP: "", }, }, EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}}, Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...}, ... // 4 identical fields VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...}, VolumeDevices: nil, LivenessProbe: &v1.Probe{ ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}}, InitialDelaySeconds: 300, TimeoutSeconds: 5, - PeriodSeconds: 10, + PeriodSeconds: 0, SuccessThreshold: 1, FailureThreshold: 3, TerminationGracePeriodSeconds: nil, }, ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...}, StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, }, EphemeralContainers: nil, - RestartPolicy: "Always", + RestartPolicy: "", TerminationGracePeriodSeconds: &600, ActiveDeadlineSeconds: nil, - DNSPolicy: "ClusterFirst", + DNSPolicy: "", NodeSelector: nil, ServiceAccountName: "default", - DeprecatedServiceAccount: "default", + DeprecatedServiceAccount: "", AutomountServiceAccountToken: nil, NodeName: "", ... // 7 identical fields Subdomain: "", Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, - SchedulerName: "default-scheduler", + SchedulerName: "", Tolerations: nil, HostAliases: nil, ... // 6 identical fields PreemptionPolicy: nil, Overhead: nil, - TopologySpreadConstraints: nil, + TopologySpreadConstraints: []v1.TopologySpreadConstraint{}, SetHostnameAsFQDN: nil, OS: nil, ... // 6 identical fields }, }, VolumeClaimTemplates: []v1.PersistentVolumeClaim{ { TypeMeta: {}, ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, Spec: v1.PersistentVolumeClaimSpec{ ... // 3 identical fields VolumeName: "", StorageClassName: nil, - VolumeMode: &"Filesystem", + VolumeMode: nil, DataSource: nil, DataSourceRef: nil, VolumeAttributesClassName: nil, }, Status: v1.PersistentVolumeClaimStatus{ - Phase: "Pending", + Phase: "", AccessModes: nil, Capacity: nil, ... // 5 identical fields }, }, }, ServiceName: "pitr-pxc", - PodManagementPolicy: "OrderedReady", + PodManagementPolicy: "", UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, - RevisionHistoryLimit: &10, + RevisionHistoryLimit: nil, MinReadySeconds: 0, - PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", + PersistentVolumeClaimRetentionPolicy: nil, Ordinals: nil, }, Status: v1.StatefulSetStatus{ - ObservedGeneration: 5, + ObservedGeneration: 0, Replicas: 0, ReadyReplicas: 0, CurrentReplicas: 0, UpdatedReplicas: 0, - CurrentRevision: "pitr-pxc-5b4fdc58c4", + CurrentRevision: "", - UpdateRevision: "pitr-pxc-5b4fdc58c4", + UpdateRevision: "", - CollisionCount: &0, + CollisionCount: nil, Conditions: nil, AvailableReplicas: 0, }, } 2026-04-29T17:13:08.028Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "50f630f4-2133-4153-b53b-f36db71b33d7", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} &v1.StatefulSet{ TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"}, ObjectMeta: v1.ObjectMeta{ ... // 2 identical fields Namespace: "pitr-pxb-25367", SelfLink: "", - UID: "a6ffacfb-86f2-42c6-8055-973883f42b38", + UID: "", - ResourceVersion: "1777482713347839013", + ResourceVersion: "", - Generation: 2, + Generation: 0, - CreationTimestamp: v1.Time{Time: s"2026-04-29 16:53:06 +0000 UTC"}, + CreationTimestamp: v1.Time{}, DeletionTimestamp: nil, DeletionGracePeriodSeconds: nil, Labels: nil, - Annotations: map[string]string{ - "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJmZTEyZDEwNGJhNDg3MjA4ZjhjNGIzNTRkYmYzMDRjZCIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiMWNjMmM3NTAzZWZmMTk0ZDQ0OGJhYjcwY2Y5ODYxN2EifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjAifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3NzbCJ9LHsibmFtZSI6"..., - }, + Annotations: map[string]string{ + "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJmZTEyZDEwNGJhNDg3MjA4ZjhjNGIzNTRkYmYzMDRjZCIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiMWNjMmM3NTAzZWZmMTk0ZDQ0OGJhYjcwY2Y5ODYxN2EifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMjAwMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIxIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjEifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIzIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIxMDAwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIyMDAwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIxMDAwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3Nz"..., + }, OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "20559ea9-e17b-447e-a2cb-6b4a334d5aaa", ...}}, Finalizers: nil, - ManagedFields: []v1.ManagedFieldsEntry{ - { - Manager: "percona-xtradb-cluster-operator", - Operation: "Update", - APIVersion: "apps/v1", - Time: s"2026-04-29 17:11:50 +0000 UTC", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., - }, - { - Manager: "kube-controller-manager", - Operation: "Update", - APIVersion: "apps/v1", - Time: s"2026-04-29 17:11:53 +0000 UTC", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., - Subresource: "status", - }, - }, + ManagedFields: nil, }, Spec: v1.StatefulSetSpec{ Replicas: &0, Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, Template: v1.PodTemplateSpec{ ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "fe12d104ba487208f8c4b354dbf304cd", "percona.com/ssl-internal-hash": "1cc2c7503eff194d448bab70cf98617a"}}, Spec: v1.PodSpec{ Volumes: []v1.Volume{ { Name: "ssl-internal", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-ssl-internal", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "ssl", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-ssl", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &false, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, {Name: "bin", VolumeSource: {EmptyDir: &{}}}, }, InitContainers: []v1.Container{ { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, }, Containers: []v1.Container{ { ... // 3 identical fields Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...}, WorkingDir: "", Ports: []v1.ContainerPort{ { Name: "mysql", HostPort: 0, ContainerPort: 3306, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "proxyadm", HostPort: 0, ContainerPort: 6032, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "stats", HostPort: 0, ContainerPort: 6070, - Protocol: "TCP", + Protocol: "", HostIP: "", }, }, EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}}, Env: []v1.EnvVar{ ... // 3 identical elements {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, {Name: "MONITOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "monitor"}}}, { Name: "SCHEDULER_CHECKTIMEOUT", - Value: "0", + Value: "2000", ValueFrom: nil, }, { Name: "SCHEDULER_WRITERALSOREADER", - Value: "0", + Value: "1", ValueFrom: nil, }, { Name: "SCHEDULER_RETRYUP", - Value: "0", + Value: "1", ValueFrom: nil, }, { Name: "SCHEDULER_RETRYDOWN", - Value: "0", + Value: "3", ValueFrom: nil, }, { Name: "SCHEDULER_PINGTIMEOUT", - Value: "0", + Value: "1000", ValueFrom: nil, }, { Name: "SCHEDULER_NODECHECKINTERVAL", - Value: "0", + Value: "2000", ValueFrom: nil, }, { Name: "SCHEDULER_MAXCONNECTIONS", - Value: "0", + Value: "1000", ValueFrom: nil, }, {Name: "PERCONA_SCHEDULER_CFG", Value: "/tmp/scheduler-config.toml"}, {Name: "PXC_READ_ONLY", Value: "false"}, }, Resources: {Limits: {s"cpu": {i: {...}, s: "700m", Format: "DecimalSI"}, s"memory": {i: {...}, s: "1G", Format: "DecimalSI"}}, Requests: {s"cpu": {i: {...}, s: "100m", Format: "DecimalSI"}, s"memory": {i: {...}, s: "100M", Format: "DecimalSI"}}}, ResizePolicy: nil, ... // 6 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, { ... // 5 identical fields Ports: nil, EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}}, Env: []v1.EnvVar{ ... // 3 identical elements {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, {Name: "MONITOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "monitor"}}}, { Name: "SCHEDULER_CHECKTIMEOUT", - Value: "0", + Value: "2000", ValueFrom: nil, }, { Name: "SCHEDULER_WRITERALSOREADER", - Value: "0", + Value: "1", ValueFrom: nil, }, { Name: "SCHEDULER_RETRYUP", - Value: "0", + Value: "1", ValueFrom: nil, }, { Name: "SCHEDULER_RETRYDOWN", - Value: "0", + Value: "3", ValueFrom: nil, }, { Name: "SCHEDULER_PINGTIMEOUT", - Value: "0", + Value: "1000", ValueFrom: nil, }, { Name: "SCHEDULER_NODECHECKINTERVAL", - Value: "0", + Value: "2000", ValueFrom: nil, }, { Name: "SCHEDULER_MAXCONNECTIONS", - Value: "0", + Value: "1000", ValueFrom: nil, }, {Name: "PERCONA_SCHEDULER_CFG", Value: "/tmp/scheduler-config.toml"}, {Name: "PXC_READ_ONLY", Value: "false"}, }, Resources: {}, ResizePolicy: nil, ... // 6 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, }, EphemeralContainers: nil, - RestartPolicy: "Always", + RestartPolicy: "", TerminationGracePeriodSeconds: &30, ActiveDeadlineSeconds: nil, - DNSPolicy: "ClusterFirst", + DNSPolicy: "", NodeSelector: nil, ServiceAccountName: "default", - DeprecatedServiceAccount: "default", + DeprecatedServiceAccount: "", AutomountServiceAccountToken: nil, NodeName: "", ... // 7 identical fields Subdomain: "", Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, - SchedulerName: "default-scheduler", + SchedulerName: "", Tolerations: nil, HostAliases: nil, ... // 6 identical fields PreemptionPolicy: nil, Overhead: nil, - TopologySpreadConstraints: nil, + TopologySpreadConstraints: []v1.TopologySpreadConstraint{}, SetHostnameAsFQDN: nil, OS: nil, ... // 6 identical fields }, }, VolumeClaimTemplates: []v1.PersistentVolumeClaim{ { TypeMeta: {}, ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, Spec: v1.PersistentVolumeClaimSpec{ ... // 3 identical fields VolumeName: "", StorageClassName: nil, - VolumeMode: &"Filesystem", + VolumeMode: nil, DataSource: nil, DataSourceRef: nil, VolumeAttributesClassName: nil, }, Status: v1.PersistentVolumeClaimStatus{ - Phase: "Pending", + Phase: "", AccessModes: nil, Capacity: nil, ... // 5 identical fields }, }, }, ServiceName: "pitr-proxysql-unready", - PodManagementPolicy: "OrderedReady", + PodManagementPolicy: "", UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, - RevisionHistoryLimit: &10, + RevisionHistoryLimit: nil, MinReadySeconds: 0, - PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", + PersistentVolumeClaimRetentionPolicy: nil, Ordinals: nil, }, Status: v1.StatefulSetStatus{ - ObservedGeneration: 2, + ObservedGeneration: 0, Replicas: 0, ReadyReplicas: 0, CurrentReplicas: 0, UpdatedReplicas: 0, - CurrentRevision: "pitr-proxysql-9cb4d4785", + CurrentRevision: "", - UpdateRevision: "pitr-proxysql-9cb4d4785", + UpdateRevision: "", - CollisionCount: &0, + CollisionCount: nil, Conditions: nil, AvailableReplicas: 0, }, } 2026-04-29T17:13:08.199Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "50f630f4-2133-4153-b53b-f36db71b33d7", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} &v1.StatefulSet{ TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"}, ObjectMeta: v1.ObjectMeta{ ... // 2 identical fields Namespace: "pitr-pxb-25367", SelfLink: "", - UID: "a6ffacfb-86f2-42c6-8055-973883f42b38", + UID: "", - ResourceVersion: "1777482713347839013", + ResourceVersion: "", - Generation: 2, + Generation: 0, - CreationTimestamp: v1.Time{Time: s"2026-04-29 16:53:06 +0000 UTC"}, + CreationTimestamp: v1.Time{}, DeletionTimestamp: nil, DeletionGracePeriodSeconds: nil, Labels: nil, - Annotations: map[string]string{ - "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJmZTEyZDEwNGJhNDg3MjA4ZjhjNGIzNTRkYmYzMDRjZCIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiMWNjMmM3NTAzZWZmMTk0ZDQ0OGJhYjcwY2Y5ODYxN2EifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjAifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3NzbCJ9LHsibmFtZSI6"..., - }, + Annotations: map[string]string{ + "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJlcm5ldGVzLmlvL25hbWUiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyIiwiYXBwLmt1YmVybmV0ZXMuaW8vcGFydC1vZiI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIifX0sInRlbXBsYXRlIjp7Im1ldGFkYXRhIjp7ImxhYmVscyI6eyJhcHAua3ViZXJuZXRlcy5pby9jb21wb25lbnQiOiJwcm94eXNxbCIsImFwcC5rdWJlcm5ldGVzLmlvL2luc3RhbmNlIjoicGl0ciIsImFwcC5rdWJlcm5ldGVzLmlvL21hbmFnZWQtYnkiOiJwZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yIiwiYXBwLmt1YmVybmV0ZXMuaW8vbmFtZSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXIiLCJhcHAua3ViZXJuZXRlcy5pby9wYXJ0LW9mIjoicGVyY29uYS14dHJhZGItY2x1c3RlciJ9LCJhbm5vdGF0aW9ucyI6eyJrdWJlY3RsLmt1YmVybmV0ZXMuaW8vZGVmYXVsdC1jb250YWluZXIiOiJwcm94eXNxbCIsInBlcmNvbmEuY29tL2NvbmZpZ3VyYXRpb24taGFzaCI6ImQ0MWQ4Y2Q5OGYwMGIyMDRlOTgwMDk5OGVjZjg0MjdlIiwicGVyY29uYS5jb20vc3NsLWhhc2giOiJmZTEyZDEwNGJhNDg3MjA4ZjhjNGIzNTRkYmYzMDRjZCIsInBlcmNvbmEuY29tL3NzbC1pbnRlcm5hbC1oYXNoIjoiMWNjMmM3NTAzZWZmMTk0ZDQ0OGJhYjcwY2Y5ODYxN2EifX0sInNwZWMiOnsidm9sdW1lcyI6W3sibmFtZSI6InNzbC1pbnRlcm5hbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wtaW50ZXJuYWwiLCJvcHRpb25hbCI6dHJ1ZX19LHsibmFtZSI6InNzbCIsInNlY3JldCI6eyJzZWNyZXROYW1lIjoicGl0ci1zc2wiLCJvcHRpb25hbCI6ZmFsc2V9fSx7Im5hbWUiOiJiaW4iLCJlbXB0eURpciI6e319XSwiaW5pdENvbnRhaW5lcnMiOlt7Im5hbWUiOiJweGMtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9weGMtaW5pdC1lbnRyeXBvaW50LnNoIl0sInJlc291cmNlcyI6eyJsaW1pdHMiOnsiY3B1IjoiNTBtIiwibWVtb3J5IjoiNTBNIn19LCJ2b2x1bWVNb3VudHMiOlt7Im5hbWUiOiJiaW4iLCJtb3VudFBhdGgiOiIvdmFyL2xpYi9teXNxbCJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn0seyJuYW1lIjoicHJveHlzcWwtaW5pdCIsImltYWdlIjoicGVyY29uYWxhYi9wZXJjb25hLXh0cmFkYi1jbHVzdGVyLW9wZXJhdG9yOlBSLTI0MzMtOWJlYzU1MGUiLCJjb21tYW5kIjpbIi9wcm94eXNxbC1pbml0LWVudHJ5cG9pbnQuc2giXSwicmVzb3VyY2VzIjp7ImxpbWl0cyI6eyJjcHUiOiI1MG0iLCJtZW1vcnkiOiI1ME0ifX0sInZvbHVtZU1vdW50cyI6W3sibmFtZSI6ImJpbiIsIm1vdW50UGF0aCI6Ii9vcHQvcGVyY29uYSJ9XSwiaW1hZ2VQdWxsUG9saWN5IjoiQWx3YXlzIn1dLCJjb250YWluZXJzIjpbeyJuYW1lIjoicHJveHlzcWwiLCJpbWFnZSI6InBlcmNvbmFsYWIvcGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvcjptYWluLXByb3h5c3FsIiwiY29tbWFuZCI6WyIvb3B0L3BlcmNvbmEvcHJveHlzcWwtZW50cnlwb2ludC5zaCJdLCJhcmdzIjpbInByb3h5c3FsIiwiLWYiLCItYyIsIi9ldGMvcHJveHlzcWwvcHJveHlzcWwuY25mIiwiLS1yZWxvYWQiXSwicG9ydHMiOlt7Im5hbWUiOiJteXNxbCIsImNvbnRhaW5lclBvcnQiOjMzMDZ9LHsibmFtZSI6InByb3h5YWRtIiwiY29udGFpbmVyUG9ydCI6NjAzMn0seyJuYW1lIjoic3RhdHMiLCJjb250YWluZXJQb3J0Ijo2MDcwfV0sImVudkZyb20iOlt7InNlY3JldFJlZiI6eyJuYW1lIjoicGl0ci1lbnYtdmFycy1wcm94eXNxbCIsIm9wdGlvbmFsIjp0cnVlfX1dLCJlbnYiOlt7Im5hbWUiOiJQWENfU0VSVklDRSIsInZhbHVlIjoicGl0ci1weGMifSx7Im5hbWUiOiJPUEVSQVRPUl9QQVNTV09SRCIsInZhbHVlRnJvbSI6eyJzZWNyZXRLZXlSZWYiOnsibmFtZSI6ImludGVybmFsLXBpdHIiLCJrZXkiOiJvcGVyYXRvciJ9fX0seyJuYW1lIjoiUFJPWFlfQURNSU5fVVNFUiIsInZhbHVlIjoicHJveHlhZG1pbiJ9LHsibmFtZSI6IlBST1hZX0FETUlOX1BBU1NXT1JEIiwidmFsdWVGcm9tIjp7InNlY3JldEtleVJlZiI6eyJuYW1lIjoiaW50ZXJuYWwtcGl0ciIsImtleSI6InByb3h5YWRtaW4ifX19LHsibmFtZSI6Ik1PTklUT1JfUEFTU1dPUkQiLCJ2YWx1ZUZyb20iOnsic2VjcmV0S2V5UmVmIjp7Im5hbWUiOiJpbnRlcm5hbC1waXRyIiwia2V5IjoibW9uaXRvciJ9fX0seyJuYW1lIjoiU0NIRURVTEVSX0NIRUNLVElNRU9VVCIsInZhbHVlIjoiMjAwMCJ9LHsibmFtZSI6IlNDSEVEVUxFUl9XUklURVJBTFNPUkVBREVSIiwidmFsdWUiOiIxIn0seyJuYW1lIjoiU0NIRURVTEVSX1JFVFJZVVAiLCJ2YWx1ZSI6IjEifSx7Im5hbWUiOiJTQ0hFRFVMRVJfUkVUUllET1dOIiwidmFsdWUiOiIzIn0seyJuYW1lIjoiU0NIRURVTEVSX1BJTkdUSU1FT1VUIiwidmFsdWUiOiIxMDAwIn0seyJuYW1lIjoiU0NIRURVTEVSX05PREVDSEVDS0lOVEVSVkFMIiwidmFsdWUiOiIyMDAwIn0seyJuYW1lIjoiU0NIRURVTEVSX01BWENPTk5FQ1RJT05TIiwidmFsdWUiOiIxMDAwIn0seyJuYW1lIjoiUEVSQ09OQV9TQ0hFRFVMRVJfQ0ZHIiwidmFsdWUiOiIvdG1wL3NjaGVkdWxlci1jb25maWcudG9tbCJ9LHsibmFtZSI6IlBYQ19SRUFEX09OTFkiLCJ2YWx1ZSI6ImZhbHNlIn1dLCJyZXNvdXJjZXMiOnsibGltaXRzIjp7ImNwdSI6IjcwMG0iLCJtZW1vcnkiOiIxRyJ9LCJyZXF1ZXN0cyI6eyJjcHUiOiIxMDBtIiwibWVtb3J5IjoiMTAwTSJ9fSwidm9sdW1lTW91bnRzIjpbeyJuYW1lIjoicHJveHlkYXRhIiwibW91bnRQYXRoIjoiL3Zhci9saWIvcHJveHlzcWwifSx7Im5hbWUiOiJzc2wiLCJtb3VudFBhdGgiOiIvZXRjL3Byb3h5c3FsL3Nz"..., + }, OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "20559ea9-e17b-447e-a2cb-6b4a334d5aaa", ...}}, Finalizers: nil, - ManagedFields: []v1.ManagedFieldsEntry{ - { - Manager: "percona-xtradb-cluster-operator", - Operation: "Update", - APIVersion: "apps/v1", - Time: s"2026-04-29 17:11:50 +0000 UTC", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., - }, - { - Manager: "kube-controller-manager", - Operation: "Update", - APIVersion: "apps/v1", - Time: s"2026-04-29 17:11:53 +0000 UTC", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., - Subresource: "status", - }, - }, + ManagedFields: nil, }, Spec: v1.StatefulSetSpec{ Replicas: &0, Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, Template: v1.PodTemplateSpec{ ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "fe12d104ba487208f8c4b354dbf304cd", "percona.com/ssl-internal-hash": "1cc2c7503eff194d448bab70cf98617a"}}, Spec: v1.PodSpec{ Volumes: []v1.Volume{ { Name: "ssl-internal", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-ssl-internal", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "ssl", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-ssl", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &false, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, {Name: "bin", VolumeSource: {EmptyDir: &{}}}, }, InitContainers: []v1.Container{ { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, }, Containers: []v1.Container{ { ... // 3 identical fields Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...}, WorkingDir: "", Ports: []v1.ContainerPort{ { Name: "mysql", HostPort: 0, ContainerPort: 3306, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "proxyadm", HostPort: 0, ContainerPort: 6032, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "stats", HostPort: 0, ContainerPort: 6070, - Protocol: "TCP", + Protocol: "", HostIP: "", }, }, EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}}, Env: []v1.EnvVar{ ... // 3 identical elements {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, {Name: "MONITOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "monitor"}}}, { Name: "SCHEDULER_CHECKTIMEOUT", - Value: "0", + Value: "2000", ValueFrom: nil, }, { Name: "SCHEDULER_WRITERALSOREADER", - Value: "0", + Value: "1", ValueFrom: nil, }, { Name: "SCHEDULER_RETRYUP", - Value: "0", + Value: "1", ValueFrom: nil, }, { Name: "SCHEDULER_RETRYDOWN", - Value: "0", + Value: "3", ValueFrom: nil, }, { Name: "SCHEDULER_PINGTIMEOUT", - Value: "0", + Value: "1000", ValueFrom: nil, }, { Name: "SCHEDULER_NODECHECKINTERVAL", - Value: "0", + Value: "2000", ValueFrom: nil, }, { Name: "SCHEDULER_MAXCONNECTIONS", - Value: "0", + Value: "1000", ValueFrom: nil, }, {Name: "PERCONA_SCHEDULER_CFG", Value: "/tmp/scheduler-config.toml"}, {Name: "PXC_READ_ONLY", Value: "false"}, }, Resources: {Limits: {s"cpu": {i: {...}, s: "700m", Format: "DecimalSI"}, s"memory": {i: {...}, s: "1G", Format: "DecimalSI"}}, Requests: {s"cpu": {i: {...}, s: "100m", Format: "DecimalSI"}, s"memory": {i: {...}, s: "100M", Format: "DecimalSI"}}}, ResizePolicy: nil, ... // 6 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, { ... // 5 identical fields Ports: nil, EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}}, Env: []v1.EnvVar{ ... // 3 identical elements {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, {Name: "MONITOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "monitor"}}}, { Name: "SCHEDULER_CHECKTIMEOUT", - Value: "0", + Value: "2000", ValueFrom: nil, }, { Name: "SCHEDULER_WRITERALSOREADER", - Value: "0", + Value: "1", ValueFrom: nil, }, { Name: "SCHEDULER_RETRYUP", - Value: "0", + Value: "1", ValueFrom: nil, }, { Name: "SCHEDULER_RETRYDOWN", - Value: "0", + Value: "3", ValueFrom: nil, }, { Name: "SCHEDULER_PINGTIMEOUT", - Value: "0", + Value: "1000", ValueFrom: nil, }, { Name: "SCHEDULER_NODECHECKINTERVAL", - Value: "0", + Value: "2000", ValueFrom: nil, }, { Name: "SCHEDULER_MAXCONNECTIONS", - Value: "0", + Value: "1000", ValueFrom: nil, }, {Name: "PERCONA_SCHEDULER_CFG", Value: "/tmp/scheduler-config.toml"}, {Name: "PXC_READ_ONLY", Value: "false"}, }, Resources: {}, ResizePolicy: nil, ... // 6 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, }, EphemeralContainers: nil, - RestartPolicy: "Always", + RestartPolicy: "", TerminationGracePeriodSeconds: &30, ActiveDeadlineSeconds: nil, - DNSPolicy: "ClusterFirst", + DNSPolicy: "", NodeSelector: nil, ServiceAccountName: "default", - DeprecatedServiceAccount: "default", + DeprecatedServiceAccount: "", AutomountServiceAccountToken: nil, NodeName: "", ... // 7 identical fields Subdomain: "", Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, - SchedulerName: "default-scheduler", + SchedulerName: "", Tolerations: nil, HostAliases: nil, ... // 6 identical fields PreemptionPolicy: nil, Overhead: nil, - TopologySpreadConstraints: nil, + TopologySpreadConstraints: []v1.TopologySpreadConstraint{}, SetHostnameAsFQDN: nil, OS: nil, ... // 6 identical fields }, }, VolumeClaimTemplates: []v1.PersistentVolumeClaim{ { TypeMeta: {}, ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, Spec: v1.PersistentVolumeClaimSpec{ ... // 3 identical fields VolumeName: "", StorageClassName: nil, - VolumeMode: &"Filesystem", + VolumeMode: nil, DataSource: nil, DataSourceRef: nil, VolumeAttributesClassName: nil, }, Status: v1.PersistentVolumeClaimStatus{ - Phase: "Pending", + Phase: "", AccessModes: nil, Capacity: nil, ... // 5 identical fields }, }, }, ServiceName: "pitr-proxysql-unready", - PodManagementPolicy: "OrderedReady", + PodManagementPolicy: "", UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, - RevisionHistoryLimit: &10, + RevisionHistoryLimit: nil, MinReadySeconds: 0, - PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", + PersistentVolumeClaimRetentionPolicy: nil, Ordinals: nil, }, Status: v1.StatefulSetStatus{ - ObservedGeneration: 2, + ObservedGeneration: 0, Replicas: 0, ReadyReplicas: 0, CurrentReplicas: 0, UpdatedReplicas: 0, - CurrentRevision: "pitr-proxysql-9cb4d4785", + CurrentRevision: "", - UpdateRevision: "pitr-proxysql-9cb4d4785", + UpdateRevision: "", - CollisionCount: &0, + CollisionCount: nil, Conditions: nil, AvailableReplicas: 0, }, } 2026-04-29T17:13:12.769Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "b03181e3-ae49-45f3-abd9-9d111471de1d", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:13:12.769Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "b03181e3-ae49-45f3-abd9-9d111471de1d", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:12.770Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "b03181e3-ae49-45f3-abd9-9d111471de1d", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:12.770Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "b03181e3-ae49-45f3-abd9-9d111471de1d", "cluster": "pitr"} 2026-04-29T17:13:17.789Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "e65d488c-b624-4a91-b963-34433113c1d3", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:13:17.789Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "e65d488c-b624-4a91-b963-34433113c1d3", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:17.791Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "e65d488c-b624-4a91-b963-34433113c1d3", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:17.791Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "e65d488c-b624-4a91-b963-34433113c1d3", "cluster": "pitr"} 2026-04-29T17:13:22.813Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "91cbb096-2663-4d3a-9c43-a45064757988", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:13:22.813Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "91cbb096-2663-4d3a-9c43-a45064757988", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:22.814Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "91cbb096-2663-4d3a-9c43-a45064757988", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:22.814Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "91cbb096-2663-4d3a-9c43-a45064757988", "cluster": "pitr"} 2026-04-29T17:13:27.832Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "5713542f-fe9b-4055-9606-d388077f2017", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:13:27.832Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "5713542f-fe9b-4055-9606-d388077f2017", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:27.833Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "5713542f-fe9b-4055-9606-d388077f2017", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:27.833Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "5713542f-fe9b-4055-9606-d388077f2017", "cluster": "pitr"} 2026-04-29T17:13:32.854Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "c8e5fefc-bdf0-417b-a8b1-089b027ed912", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:13:32.854Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "c8e5fefc-bdf0-417b-a8b1-089b027ed912", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:32.855Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "c8e5fefc-bdf0-417b-a8b1-089b027ed912", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:32.855Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "c8e5fefc-bdf0-417b-a8b1-089b027ed912", "cluster": "pitr"} 2026-04-29T17:13:37.884Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "0c220f3f-a08e-446c-ab91-11ecc6870e82", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:13:37.884Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "0c220f3f-a08e-446c-ab91-11ecc6870e82", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:37.885Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "0c220f3f-a08e-446c-ab91-11ecc6870e82", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:37.885Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "0c220f3f-a08e-446c-ab91-11ecc6870e82", "cluster": "pitr"} 2026-04-29T17:13:42.902Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "a1bcd1e7-0527-4413-9521-0f1549666962", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:13:42.902Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "a1bcd1e7-0527-4413-9521-0f1549666962", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:42.903Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "a1bcd1e7-0527-4413-9521-0f1549666962", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:42.903Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "a1bcd1e7-0527-4413-9521-0f1549666962", "cluster": "pitr"} 2026-04-29T17:13:47.920Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "5d0efcd8-068c-4020-93c3-5f5a6f46597f", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:13:47.920Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "5d0efcd8-068c-4020-93c3-5f5a6f46597f", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:47.921Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "5d0efcd8-068c-4020-93c3-5f5a6f46597f", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:47.921Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "5d0efcd8-068c-4020-93c3-5f5a6f46597f", "cluster": "pitr"} 2026-04-29T17:13:52.947Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "9fe5869c-10b4-4f67-936f-4b3d89d51d28", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:13:52.947Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "9fe5869c-10b4-4f67-936f-4b3d89d51d28", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:52.947Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "9fe5869c-10b4-4f67-936f-4b3d89d51d28", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:52.947Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "9fe5869c-10b4-4f67-936f-4b3d89d51d28", "cluster": "pitr"} 2026-04-29T17:13:55.484Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "7a760d92-2690-48b0-bc58-143f1a803c14"} 2026-04-29T17:13:55.550Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "7a760d92-2690-48b0-bc58-143f1a803c14", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-pxb-25367 on 34.118.224.10:53: no such host"} 2026-04-29T17:13:57.973Z INFO invalidating binlog collector cache {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "1434d21c-ffda-42cf-9836-9430fc705e2e", "storage": "minio-binlogs", "file": "gtid-binlog-cache.json"} 2026-04-29T17:13:57.973Z DEBUG deleting object {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "1434d21c-ffda-42cf-9836-9430fc705e2e", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:57.974Z DEBUG object deleted {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "1434d21c-ffda-42cf-9836-9430fc705e2e", "bucket": "operator-testing", "prefix": "binlogs/", "object": "binlogs/gtid-binlog-cache.json"} 2026-04-29T17:13:57.974Z INFO point-in-time recovering {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "1434d21c-ffda-42cf-9836-9430fc705e2e", "cluster": "pitr"} 2026-04-29T17:13:58.195Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "0282af68-2b26-42d5-8505-b29adf5e2731", "job": "pitr-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:14:00.773Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "f3cf47c7-0ebe-41c4-bd14-395fc55a5e22"} 2026-04-29T17:14:00.833Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "f3cf47c7-0ebe-41c4-bd14-395fc55a5e22", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-pxb-25367 on 34.118.224.10:53: no such host"} 2026-04-29T17:14:03.215Z INFO Waiting for restore job to finish {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "2ad4f922-387b-42ff-943d-9fcd548aa7ad", "job": "pitr-job-restore-on-pitr-minio-gtid-pitr"} 2026-04-29T17:14:06.075Z INFO Old password discarded {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "03c8dc3a-55e9-47cc-bf1d-ec7ec0ae6dcf", "user": "xtrabackup"} 2026-04-29T17:14:06.078Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "03c8dc3a-55e9-47cc-bf1d-ec7ec0ae6dcf"} 2026-04-29T17:14:06.154Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "03c8dc3a-55e9-47cc-bf1d-ec7ec0ae6dcf", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-pxb-25367 on 34.118.224.10:53: no such host"} 2026-04-29T17:14:08.264Z INFO starting cluster {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "43ac1e5e-fcf9-4979-bfb7-c8fce8b4ac4b", "cluster": "pitr"} 2026-04-29T17:14:08.550Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "d5e0d7f2-21fe-4093-97e2-4dc10c2274d1", "object": "pitr-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} &v1.StatefulSet{ TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"}, ObjectMeta: v1.ObjectMeta{ ... // 2 identical fields Namespace: "pitr-pxb-25367", SelfLink: "", - UID: "37c2fe66-113b-4ba7-a0d8-55e1f62aecb7", + UID: "", - ResourceVersion: "1777482834596623023", + ResourceVersion: "", - Generation: 6, + Generation: 0, - CreationTimestamp: v1.Time{Time: s"2026-04-29 16:53:06 +0000 UTC"}, + CreationTimestamp: v1.Time{}, DeletionTimestamp: nil, DeletionGracePeriodSeconds: nil, Labels: nil, - Annotations: map[string]string{ - "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MSwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., - }, + Annotations: map[string]string{ + "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Mywic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHhjIiwiYXBwLmt1YmVybmV0ZXMuaW8vaW5zdGFuY2UiOiJwaXRyIiwiYXBwLmt1YmVybmV0ZXMuaW8vbWFuYWdlZC1ieSI6InBlcmNvbmEteHRyYWRiLWNsdXN0ZXItb3BlcmF0b3IiLCJhcHAua3ViZXJuZXRl"..., + }, OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "20559ea9-e17b-447e-a2cb-6b4a334d5aaa", ...}}, Finalizers: nil, - ManagedFields: []v1.ManagedFieldsEntry{ - { - Manager: "percona-xtradb-cluster-operator", - Operation: "Update", - APIVersion: "apps/v1", - Time: s"2026-04-29 17:13:07 +0000 UTC", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., - }, - { - Manager: "kube-controller-manager", - Operation: "Update", - APIVersion: "apps/v1", - Time: s"2026-04-29 17:13:54 +0000 UTC", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., - Subresource: "status", - }, - }, + ManagedFields: nil, }, Spec: v1.StatefulSetSpec{ - Replicas: &1, + Replicas: &3, Selector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, Template: v1.PodTemplateSpec{ ObjectMeta: {Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "pxc", "last-applied-secret": "df8cbb9bd67352c10228f4870690fe8020c1933c6911f7b946c47e2b76080b27", "percona.com/configuration-hash": "ccc80fc959e03c87401faeb7a6adeb79", "percona.com/ssl-hash": "fe12d104ba487208f8c4b354dbf304cd", ...}}, Spec: v1.PodSpec{ Volumes: []v1.Volume{ {Name: "tmp", VolumeSource: {EmptyDir: &{}}}, { Name: "config", VolumeSource: v1.VolumeSource{ ... // 16 identical fields FC: nil, AzureFile: nil, ConfigMap: &v1.ConfigMapVolumeSource{ LocalObjectReference: {Name: "pitr-pxc"}, Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, VsphereVolume: nil, Quobyte: nil, ... // 9 identical fields }, }, { Name: "ssl-internal", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-ssl-internal", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "ssl", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-ssl", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &false, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "auto-config", VolumeSource: v1.VolumeSource{ ... // 16 identical fields FC: nil, AzureFile: nil, ConfigMap: &v1.ConfigMapVolumeSource{ LocalObjectReference: {Name: "auto-pitr-pxc"}, Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, VsphereVolume: nil, Quobyte: nil, ... // 9 identical fields }, }, { Name: "vault-keyring-secret", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "some-name-vault", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "mysql-users-secret-file", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "internal-pitr", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &false, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "mysql-init-file", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-mysql-init", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, {Name: "bin", VolumeSource: {EmptyDir: &{}}}, {Name: "backup-logs", VolumeSource: {EmptyDir: &{}}}, }, InitContainers: []v1.Container{ { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, }, Containers: []v1.Container{ { ... // 3 identical fields Args: nil, WorkingDir: "", Ports: []v1.ContainerPort{ { Name: "grpc", HostPort: 0, ContainerPort: 6450, - Protocol: "TCP", + Protocol: "", HostIP: "", }, }, EnvFrom: nil, Env: []v1.EnvVar{ { Name: "POD_NAMESPACE", Value: "", ValueFrom: &v1.EnvVarSource{ FieldRef: &v1.ObjectFieldSelector{ - APIVersion: "v1", + APIVersion: "", FieldPath: "metadata.namespace", }, ResourceFieldRef: nil, ConfigMapKeyRef: nil, ... // 2 identical fields }, }, {Name: "XTRABACKUP_USER_PASS", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "my-cluster-secrets"}, Key: "xtrabackup"}}}, {Name: "VAULT_KEYRING_PATH", Value: "/etc/mysql/vault-keyring-secret/keyring_vault.conf"}, }, Resources: {Limits: {s"cpu": {i: {...}, s: "200m", Format: "DecimalSI"}, s"memory": {i: {...}, Format: "BinarySI"}}, Requests: {s"cpu": {i: {...}, s: "200m", Format: "DecimalSI"}, s"memory": {i: {...}, Format: "BinarySI"}}}, ResizePolicy: nil, ... // 6 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, { ... // 3 identical fields Args: {"mysqld"}, WorkingDir: "", Ports: []v1.ContainerPort{ { Name: "mysql", HostPort: 0, ContainerPort: 3306, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "sst", HostPort: 0, ContainerPort: 4444, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "write-set", HostPort: 0, ContainerPort: 4567, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "ist", HostPort: 0, ContainerPort: 4568, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "mysql-admin", HostPort: 0, ContainerPort: 33062, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "mysqlx", HostPort: 0, ContainerPort: 33060, - Protocol: "TCP", + Protocol: "", HostIP: "", }, }, EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-pxc"}, Optional: &true}}}, Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc-unready"}, {Name: "MONITOR_HOST", Value: "%"}, {Name: "MYSQL_ROOT_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "root"}}}, {Name: "XTRABACKUP_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "xtrabackup"}}}, ...}, ... // 4 identical fields VolumeMounts: {{Name: "datadir", MountPath: "/var/lib/mysql"}, {Name: "config", MountPath: "/etc/percona-xtradb-cluster.conf.d"}, {Name: "tmp", MountPath: "/tmp"}, {Name: "ssl", MountPath: "/etc/mysql/ssl"}, ...}, VolumeDevices: nil, LivenessProbe: &v1.Probe{ ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/liveness-check.sh"}}}, InitialDelaySeconds: 300, TimeoutSeconds: 5, - PeriodSeconds: 10, + PeriodSeconds: 0, SuccessThreshold: 1, FailureThreshold: 3, TerminationGracePeriodSeconds: nil, }, ReadinessProbe: &{ProbeHandler: {Exec: &{Command: {"/var/lib/mysql/readiness-check.sh"}}}, InitialDelaySeconds: 15, TimeoutSeconds: 15, PeriodSeconds: 30, ...}, StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, }, EphemeralContainers: nil, - RestartPolicy: "Always", + RestartPolicy: "", TerminationGracePeriodSeconds: &600, ActiveDeadlineSeconds: nil, - DNSPolicy: "ClusterFirst", + DNSPolicy: "", NodeSelector: nil, ServiceAccountName: "default", - DeprecatedServiceAccount: "default", + DeprecatedServiceAccount: "", AutomountServiceAccountToken: nil, NodeName: "", ... // 7 identical fields Subdomain: "", Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, - SchedulerName: "default-scheduler", + SchedulerName: "", Tolerations: nil, HostAliases: nil, ... // 6 identical fields PreemptionPolicy: nil, Overhead: nil, - TopologySpreadConstraints: nil, + TopologySpreadConstraints: []v1.TopologySpreadConstraint{}, SetHostnameAsFQDN: nil, OS: nil, ... // 6 identical fields }, }, VolumeClaimTemplates: []v1.PersistentVolumeClaim{ { TypeMeta: {}, ObjectMeta: {Name: "datadir", Labels: {"app.kubernetes.io/component": "pxc", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, Spec: v1.PersistentVolumeClaimSpec{ ... // 3 identical fields VolumeName: "", StorageClassName: nil, - VolumeMode: &"Filesystem", + VolumeMode: nil, DataSource: nil, DataSourceRef: nil, VolumeAttributesClassName: nil, }, Status: v1.PersistentVolumeClaimStatus{ - Phase: "Pending", + Phase: "", AccessModes: nil, Capacity: nil, ... // 5 identical fields }, }, }, ServiceName: "pitr-pxc", - PodManagementPolicy: "OrderedReady", + PodManagementPolicy: "", UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, - RevisionHistoryLimit: &10, + RevisionHistoryLimit: nil, MinReadySeconds: 0, - PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", + PersistentVolumeClaimRetentionPolicy: nil, Ordinals: nil, }, Status: v1.StatefulSetStatus{ - ObservedGeneration: 6, + ObservedGeneration: 0, - Replicas: 1, + Replicas: 0, - ReadyReplicas: 1, + ReadyReplicas: 0, - CurrentReplicas: 1, + CurrentReplicas: 0, - UpdatedReplicas: 1, + UpdatedReplicas: 0, - CurrentRevision: "pitr-pxc-5b4fdc58c4", + CurrentRevision: "", - UpdateRevision: "pitr-pxc-5b4fdc58c4", + UpdateRevision: "", - CollisionCount: &0, + CollisionCount: nil, Conditions: nil, - AvailableReplicas: 1, + AvailableReplicas: 0, }, } 2026-04-29T17:14:08.617Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "d5e0d7f2-21fe-4093-97e2-4dc10c2274d1", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} &v1.StatefulSet{ TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"}, ObjectMeta: v1.ObjectMeta{ ... // 2 identical fields Namespace: "pitr-pxb-25367", SelfLink: "", - UID: "a6ffacfb-86f2-42c6-8055-973883f42b38", + UID: "", - ResourceVersion: "1777482788360287013", + ResourceVersion: "", - Generation: 3, + Generation: 0, - CreationTimestamp: v1.Time{Time: s"2026-04-29 16:53:06 +0000 UTC"}, + CreationTimestamp: v1.Time{}, DeletionTimestamp: nil, DeletionGracePeriodSeconds: nil, Labels: nil, - Annotations: map[string]string{ - "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., - }, + Annotations: map[string]string{ + "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., + }, OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "20559ea9-e17b-447e-a2cb-6b4a334d5aaa", ...}}, Finalizers: nil, - ManagedFields: []v1.ManagedFieldsEntry{ - { - Manager: "kube-controller-manager", - Operation: "Update", - APIVersion: "apps/v1", - Time: s"2026-04-29 17:13:08 +0000 UTC", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., - Subresource: "status", - }, - { - Manager: "percona-xtradb-cluster-operator", - Operation: "Update", - APIVersion: "apps/v1", - Time: s"2026-04-29 17:13:08 +0000 UTC", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., - }, - }, + ManagedFields: nil, }, Spec: v1.StatefulSetSpec{ - Replicas: &0, + Replicas: &2, Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, Template: v1.PodTemplateSpec{ ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "fe12d104ba487208f8c4b354dbf304cd", "percona.com/ssl-internal-hash": "1cc2c7503eff194d448bab70cf98617a"}}, Spec: v1.PodSpec{ Volumes: []v1.Volume{ { Name: "ssl-internal", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-ssl-internal", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "ssl", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-ssl", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &false, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, {Name: "bin", VolumeSource: {EmptyDir: &{}}}, }, InitContainers: []v1.Container{ { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, }, Containers: []v1.Container{ { ... // 3 identical fields Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...}, WorkingDir: "", Ports: []v1.ContainerPort{ { Name: "mysql", HostPort: 0, ContainerPort: 3306, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "proxyadm", HostPort: 0, ContainerPort: 6032, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "stats", HostPort: 0, ContainerPort: 6070, - Protocol: "TCP", + Protocol: "", HostIP: "", }, }, EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}}, Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...}, ... // 8 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, }, EphemeralContainers: nil, - RestartPolicy: "Always", + RestartPolicy: "", TerminationGracePeriodSeconds: &30, ActiveDeadlineSeconds: nil, - DNSPolicy: "ClusterFirst", + DNSPolicy: "", NodeSelector: nil, ServiceAccountName: "default", - DeprecatedServiceAccount: "default", + DeprecatedServiceAccount: "", AutomountServiceAccountToken: nil, NodeName: "", ... // 7 identical fields Subdomain: "", Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, - SchedulerName: "default-scheduler", + SchedulerName: "", Tolerations: nil, HostAliases: nil, ... // 6 identical fields PreemptionPolicy: nil, Overhead: nil, - TopologySpreadConstraints: nil, + TopologySpreadConstraints: []v1.TopologySpreadConstraint{}, SetHostnameAsFQDN: nil, OS: nil, ... // 6 identical fields }, }, VolumeClaimTemplates: []v1.PersistentVolumeClaim{ { TypeMeta: {}, ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, Spec: v1.PersistentVolumeClaimSpec{ ... // 3 identical fields VolumeName: "", StorageClassName: nil, - VolumeMode: &"Filesystem", + VolumeMode: nil, DataSource: nil, DataSourceRef: nil, VolumeAttributesClassName: nil, }, Status: v1.PersistentVolumeClaimStatus{ - Phase: "Pending", + Phase: "", AccessModes: nil, Capacity: nil, ... // 5 identical fields }, }, }, ServiceName: "pitr-proxysql-unready", - PodManagementPolicy: "OrderedReady", + PodManagementPolicy: "", UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, - RevisionHistoryLimit: &10, + RevisionHistoryLimit: nil, MinReadySeconds: 0, - PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", + PersistentVolumeClaimRetentionPolicy: nil, Ordinals: nil, }, Status: v1.StatefulSetStatus{ - ObservedGeneration: 3, + ObservedGeneration: 0, Replicas: 0, ReadyReplicas: 0, CurrentReplicas: 0, UpdatedReplicas: 0, - CurrentRevision: "pitr-proxysql-67d5d5475f", + CurrentRevision: "", - UpdateRevision: "pitr-proxysql-67d5d5475f", + UpdateRevision: "", - CollisionCount: &0, + CollisionCount: nil, Conditions: nil, AvailableReplicas: 0, }, } 2026-04-29T17:14:08.710Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "d5e0d7f2-21fe-4093-97e2-4dc10c2274d1", "object": "pitr-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} &v1.StatefulSet{ TypeMeta: {Kind: "StatefulSet", APIVersion: "apps/v1"}, ObjectMeta: v1.ObjectMeta{ ... // 2 identical fields Namespace: "pitr-pxb-25367", SelfLink: "", - UID: "a6ffacfb-86f2-42c6-8055-973883f42b38", + UID: "", - ResourceVersion: "1777482788360287013", + ResourceVersion: "", - Generation: 3, + Generation: 0, - CreationTimestamp: v1.Time{Time: s"2026-04-29 16:53:06 +0000 UTC"}, + CreationTimestamp: v1.Time{}, DeletionTimestamp: nil, DeletionGracePeriodSeconds: nil, Labels: nil, - Annotations: map[string]string{ - "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6MCwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., - }, + Annotations: map[string]string{ + "percona.com/last-config-hash": "eyJyZXBsaWNhcyI6Miwic2VsZWN0b3IiOnsibWF0Y2hMYWJlbHMiOnsiYXBwLmt1YmVybmV0ZXMuaW8vY29tcG9uZW50IjoicHJveHlzcWwiLCJhcHAua3ViZXJuZXRlcy5pby9pbnN0YW5jZSI6InBpdHIiLCJhcHAua3ViZXJuZXRlcy5pby9tYW5hZ2VkLWJ5IjoicGVyY29uYS14dHJhZGItY2x1c3Rlci1vcGVyYXRvciIsImFwcC5rdWJl"..., + }, OwnerReferences: {{APIVersion: "pxc.percona.com/v1", Kind: "PerconaXtraDBCluster", Name: "pitr", UID: "20559ea9-e17b-447e-a2cb-6b4a334d5aaa", ...}}, Finalizers: nil, - ManagedFields: []v1.ManagedFieldsEntry{ - { - Manager: "kube-controller-manager", - Operation: "Update", - APIVersion: "apps/v1", - Time: s"2026-04-29 17:13:08 +0000 UTC", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:status":{"f:availableReplicas":{},"f:collisionCount":{},"f:c`..., - Subresource: "status", - }, - { - Manager: "percona-xtradb-cluster-operator", - Operation: "Update", - APIVersion: "apps/v1", - Time: s"2026-04-29 17:13:08 +0000 UTC", - FieldsType: "FieldsV1", - FieldsV1: s`{"f:metadata":{"f:annotations":{".":{},"f:percona.com/last-confi`..., - }, - }, + ManagedFields: nil, }, Spec: v1.StatefulSetSpec{ - Replicas: &0, + Replicas: &2, Selector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, Template: v1.PodTemplateSpec{ ObjectMeta: {Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}, Annotations: {"kubectl.kubernetes.io/default-container": "proxysql", "percona.com/configuration-hash": "d41d8cd98f00b204e9800998ecf8427e", "percona.com/ssl-hash": "fe12d104ba487208f8c4b354dbf304cd", "percona.com/ssl-internal-hash": "1cc2c7503eff194d448bab70cf98617a"}}, Spec: v1.PodSpec{ Volumes: []v1.Volume{ { Name: "ssl-internal", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-ssl-internal", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &true, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, { Name: "ssl", VolumeSource: v1.VolumeSource{ ... // 3 identical fields AWSElasticBlockStore: nil, GitRepo: nil, Secret: &v1.SecretVolumeSource{ SecretName: "pitr-ssl", Items: nil, - DefaultMode: &420, + DefaultMode: nil, Optional: &false, }, NFS: nil, ISCSI: nil, ... // 22 identical fields }, }, {Name: "bin", VolumeSource: {EmptyDir: &{}}}, }, InitContainers: []v1.Container{ { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, }, Containers: []v1.Container{ { ... // 3 identical fields Args: {"proxysql", "-f", "-c", "/etc/proxysql/proxysql.cnf", ...}, WorkingDir: "", Ports: []v1.ContainerPort{ { Name: "mysql", HostPort: 0, ContainerPort: 3306, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "proxyadm", HostPort: 0, ContainerPort: 6032, - Protocol: "TCP", + Protocol: "", HostIP: "", }, { Name: "stats", HostPort: 0, ContainerPort: 6070, - Protocol: "TCP", + Protocol: "", HostIP: "", }, }, EnvFrom: {{SecretRef: &{LocalObjectReference: {Name: "pitr-env-vars-proxysql"}, Optional: &true}}}, Env: {{Name: "PXC_SERVICE", Value: "pitr-pxc"}, {Name: "OPERATOR_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "operator"}}}, {Name: "PROXY_ADMIN_USER", Value: "proxyadmin"}, {Name: "PROXY_ADMIN_PASSWORD", ValueFrom: &{SecretKeyRef: &{LocalObjectReference: {Name: "internal-pitr"}, Key: "proxyadmin"}}}, ...}, ... // 8 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, { ... // 16 identical fields StartupProbe: nil, Lifecycle: nil, - TerminationMessagePath: "/dev/termination-log", + TerminationMessagePath: "", - TerminationMessagePolicy: "File", + TerminationMessagePolicy: "", ImagePullPolicy: "Always", SecurityContext: nil, ... // 3 identical fields }, }, EphemeralContainers: nil, - RestartPolicy: "Always", + RestartPolicy: "", TerminationGracePeriodSeconds: &30, ActiveDeadlineSeconds: nil, - DNSPolicy: "ClusterFirst", + DNSPolicy: "", NodeSelector: nil, ServiceAccountName: "default", - DeprecatedServiceAccount: "default", + DeprecatedServiceAccount: "", AutomountServiceAccountToken: nil, NodeName: "", ... // 7 identical fields Subdomain: "", Affinity: &{PodAntiAffinity: &{RequiredDuringSchedulingIgnoredDuringExecution: {{LabelSelector: &{MatchLabels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, TopologyKey: "kubernetes.io/hostname"}}}}, - SchedulerName: "default-scheduler", + SchedulerName: "", Tolerations: nil, HostAliases: nil, ... // 6 identical fields PreemptionPolicy: nil, Overhead: nil, - TopologySpreadConstraints: nil, + TopologySpreadConstraints: []v1.TopologySpreadConstraint{}, SetHostnameAsFQDN: nil, OS: nil, ... // 6 identical fields }, }, VolumeClaimTemplates: []v1.PersistentVolumeClaim{ { TypeMeta: {}, ObjectMeta: {Name: "proxydata", Labels: {"app.kubernetes.io/component": "proxysql", "app.kubernetes.io/instance": "pitr", "app.kubernetes.io/managed-by": "percona-xtradb-cluster-operator", "app.kubernetes.io/name": "percona-xtradb-cluster", ...}}, Spec: v1.PersistentVolumeClaimSpec{ ... // 3 identical fields VolumeName: "", StorageClassName: nil, - VolumeMode: &"Filesystem", + VolumeMode: nil, DataSource: nil, DataSourceRef: nil, VolumeAttributesClassName: nil, }, Status: v1.PersistentVolumeClaimStatus{ - Phase: "Pending", + Phase: "", AccessModes: nil, Capacity: nil, ... // 5 identical fields }, }, }, ServiceName: "pitr-proxysql-unready", - PodManagementPolicy: "OrderedReady", + PodManagementPolicy: "", UpdateStrategy: {Type: "RollingUpdate", RollingUpdate: &{Partition: &0}}, - RevisionHistoryLimit: &10, + RevisionHistoryLimit: nil, MinReadySeconds: 0, - PersistentVolumeClaimRetentionPolicy: s"&StatefulSetPersistentVolumeClaimRetentionPolicy{WhenDeleted:Retain,WhenScaled:Retain,}", + PersistentVolumeClaimRetentionPolicy: nil, Ordinals: nil, }, Status: v1.StatefulSetStatus{ - ObservedGeneration: 3, + ObservedGeneration: 0, Replicas: 0, ReadyReplicas: 0, CurrentReplicas: 0, UpdatedReplicas: 0, - CurrentRevision: "pitr-proxysql-67d5d5475f", + CurrentRevision: "", - UpdateRevision: "pitr-proxysql-67d5d5475f", + UpdateRevision: "", - CollisionCount: &0, + CollisionCount: nil, Conditions: nil, AvailableReplicas: 0, }, } 2026-04-29T17:14:11.354Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "d5e0d7f2-21fe-4093-97e2-4dc10c2274d1", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-pxb-25367 on 34.118.224.10:53: no such host"} 2026-04-29T17:14:13.407Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "3d0c9be4-d14f-4646-b924-2a500ff9453e", "cluster": "pitr"} 2026-04-29T17:14:14.227Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "48637f92-8285-445f-8655-e58723019f52", "err": "get primary pxc pod: failed to get proxy connection: dial tcp: lookup pitr-proxysql-unready.pitr-pxb-25367 on 34.118.224.10:53: no such host"} 2026-04-29T17:14:18.433Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "1178b496-3c79-4385-af17-930fa694f8f1", "cluster": "pitr"} 2026-04-29T17:14:21.804Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "360af132-4923-4138-a192-0982e052e4b0", "err": "get primary pxc pod: failed to get proxy connection: dial tcp 10.79.82.65:6032: connect: connection refused"} 2026-04-29T17:14:23.453Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "9e611d44-fb6e-4ea4-a29b-2fccc779e932", "cluster": "pitr"} 2026-04-29T17:14:28.496Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "087af2a8-b46b-4572-b5c2-05f6c70c598f", "cluster": "pitr"} 2026-04-29T17:14:33.516Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "1daabd36-c12f-473a-a9b6-5faf76a054b6", "cluster": "pitr"} 2026-04-29T17:14:38.539Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "b805ec7c-a8ae-425d-b955-3c2cf5530fb4", "cluster": "pitr"} 2026-04-29T17:14:43.592Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "7e72a143-2c0b-4626-9d1c-298de41ba3e8", "cluster": "pitr"} 2026-04-29T17:14:48.628Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "3310f5b1-348d-44c7-ac1b-f32a8b89f7f8", "cluster": "pitr"} 2026-04-29T17:14:53.646Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "63af7094-2df0-4650-bec0-31862491909f", "cluster": "pitr"} 2026-04-29T17:14:58.664Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "a8e84d5c-48e3-4cec-83c9-7925fd90bc47", "cluster": "pitr"} 2026-04-29T17:15:03.682Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "7b685c02-84d3-4e68-a64f-902b79ceb0cb", "cluster": "pitr"} 2026-04-29T17:15:08.701Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "30554df6-5bce-431a-ad31-ed0a8a195d7b", "cluster": "pitr"} 2026-04-29T17:15:13.722Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "46204c28-1ce3-46b5-af8b-8cfe900edeb0", "cluster": "pitr"} 2026-04-29T17:15:18.749Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "08553f99-8743-4113-88f1-f6f8ac09dc7a", "cluster": "pitr"} 2026-04-29T17:15:23.767Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "7ac5954e-22ba-4b89-81a7-7c7a3e7da101", "cluster": "pitr"} 2026-04-29T17:15:28.793Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "0da72cf9-22b7-4c43-bd13-f20221118343", "cluster": "pitr"} 2026-04-29T17:15:33.814Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "a033c00d-7475-453f-8731-8602156920e5", "cluster": "pitr"} 2026-04-29T17:15:38.834Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "6a6f4a5d-97bb-45d2-a8b3-5b7d8ab66228", "cluster": "pitr"} 2026-04-29T17:15:43.868Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "3806082f-d771-4d24-8875-57532e36cf16", "cluster": "pitr"} 2026-04-29T17:15:48.884Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "de196b14-4f3f-4afc-bd1f-e6886832b16a", "cluster": "pitr"} 2026-04-29T17:15:53.902Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "95d68834-cbbd-48b7-9ab2-7392fbf8589c", "cluster": "pitr"} 2026-04-29T17:15:58.922Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "aabd2e20-b57a-4abe-88e8-5149d7c38d89", "cluster": "pitr"} 2026-04-29T17:16:03.943Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "43498f9a-b911-4b14-9309-4dc467f0a6da", "cluster": "pitr"} 2026-04-29T17:16:08.973Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "feaa6b0d-498b-4a06-a114-df8901d82c6f", "cluster": "pitr"} 2026-04-29T17:16:14.014Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "2d9a5058-8248-4ff3-8b8d-9fb00a3becbb", "cluster": "pitr"} 2026-04-29T17:16:19.036Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "0ec5e5c0-d19b-4d2f-8d97-1131370a811d", "cluster": "pitr"} 2026-04-29T17:16:24.059Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "cec83d03-56b2-4a4c-aaf1-c41b5f6095b4", "cluster": "pitr"} 2026-04-29T17:16:29.080Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "5f971239-7771-4fa6-b5d7-2d13a3b67319", "cluster": "pitr"} 2026-04-29T17:16:34.109Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "3894512c-f986-443b-86ea-99d0e7ae7779", "cluster": "pitr"} 2026-04-29T17:16:39.126Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "2284f8d6-44b8-4e18-996a-048ec9626cf5", "cluster": "pitr"} 2026-04-29T17:16:44.184Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "76717b09-9eca-4df2-9081-e63770a73457", "cluster": "pitr"} 2026-04-29T17:16:49.203Z INFO Waiting for cluster to start {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "PerconaXtraDBClusterRestore": {"name":"restore-on-pitr-minio-gtid","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "restore-on-pitr-minio-gtid", "reconcileID": "bbcec834-7b15-49a8-b591-bc504e6fac6a", "cluster": "pitr"} 2026-04-29T17:16:51.657Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "d7a81a61-a079-4a3e-bfed-fbbb549acb12"} 2026-04-29T17:16:55.162Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "a64cd75f-c38e-437e-911e-248fc268a395", "object": "pitr-pitr", "kind": "&TypeMeta{Kind:Deployment,APIVersion:apps/v1,}"} 2026-04-29T17:16:57.168Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "a64cd75f-c38e-437e-911e-248fc268a395"} 2026-04-29T17:17:01.300Z INFO Updated PITR timelines {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "9d5dc059-6a46-4a9d-a6c6-4cae24493ea5", "latest": "2026-04-29 17:14:55 +0000 UTC", "lastBackup": "on-pitr-minio"} 2026-04-29T17:17:02.479Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "9d5dc059-6a46-4a9d-a6c6-4cae24493ea5"} 2026-04-29T17:17:08.162Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "20b9cc75-8eae-4e90-b37b-332ffe38ae99"} 2026-04-29T17:17:13.856Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "0c12dc26-3c3e-43b8-9079-b0dbe481b42b"} 2026-04-29T17:17:19.988Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "5f24bd8f-0a7c-4081-83ef-05b1323067f3"} 2026-04-29T17:17:25.933Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "d9d3d34b-e3de-4759-b1de-56654a5ed08b"} 2026-04-29T17:17:31.653Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "6e0d7a7a-908e-4f72-91c0-022b3faf51d2"} 2026-04-29T17:17:37.663Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "c8924f69-bf2c-4d49-ba8c-0d8a2ec825ac"} 2026-04-29T17:17:43.464Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "e5b5bd63-ec5d-42c6-937e-f358da5ca6de"} 2026-04-29T17:17:49.255Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "5f1d6966-8e26-45f5-aea2-8e4640a925be"} 2026-04-29T17:17:55.371Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "06b265f1-9037-4c6c-9b99-4dc9436c645e"} 2026-04-29T17:18:01.170Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "6881e7b9-f769-4b47-bc32-41b9aebf1d07"} 2026-04-29T17:18:07.055Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "86fdf87c-7821-430b-b269-a1cc90f9d406"} 2026-04-29T17:18:13.353Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "f9930e02-bb4f-47f5-ac64-057da0d59710"} 2026-04-29T17:18:19.070Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "17e154c2-20bb-4fa9-9b46-9bc4338c3499"} 2026-04-29T17:18:25.180Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "30ea9966-2269-4ea7-b4fd-9c2456ed1f8d"} 2026-04-29T17:18:31.132Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "9677c45c-3da2-4afd-8220-e13de1954b98"} 2026-04-29T17:18:36.980Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "716b34d1-f32f-4cf8-9396-651d5b3e3e50"} 2026-04-29T17:18:42.570Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "0af61d04-6c6a-44a6-9f71-ccc7332ce825"} 2026-04-29T17:18:48.784Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "8f570b09-c264-4ad9-96a1-1ed2157c76e7"} 2026-04-29T17:18:54.949Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "c7b3003c-2544-42e9-a5f7-b4d52f23fb1a"} 2026-04-29T17:19:00.861Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "34da751a-42ff-4866-b4f6-cd053838cf1c"} 2026-04-29T17:19:06.462Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "bd3869a1-05bf-4ffe-8451-ac01802b303f"} 2026-04-29T17:19:12.829Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "b5939f31-1bbc-46bb-93bf-b3a5695664b4"} 2026-04-29T17:19:19.035Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "2139ae7b-a9e4-42b2-8b61-34b1af7981ec"} 2026-04-29T17:19:25.067Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "a32607cd-8117-4f7f-952f-f7b21fc9b965"} 2026-04-29T17:19:30.471Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "923f60cb-e564-4651-8867-ed8b2fc2b68e"} 2026-04-29T17:19:36.568Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "9298acac-4773-4ab5-987c-42a38e9d827c"} 2026-04-29T17:19:42.356Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "f714421b-45fd-44a4-bfc4-22d7e84386b3"} 2026-04-29T17:19:48.483Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "81c09d01-8388-4cd0-9b20-d9c440dd9a30"} 2026-04-29T17:19:53.751Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "5c0ff9d1-9309-4bbe-9cab-86fd2111d16f"} 2026-04-29T17:19:59.635Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "a334fc70-39df-4dd8-90cb-8d05e6b277e8"} 2026-04-29T17:20:05.240Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "49dc1832-8fac-4d95-b686-30f38ed34680"} 2026-04-29T17:20:11.571Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "3ffae360-8a55-4899-896d-854fc134d53f"} 2026-04-29T17:20:17.551Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "ed2711ed-9b72-4bb6-8015-fe4757d12f08"} 2026-04-29T17:20:23.245Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "9a446599-c441-45a3-a3af-47952f96cfd5"} 2026-04-29T17:20:29.180Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"pitr","namespace":"pitr-pxb-25367"}, "namespace": "pitr-pxb-25367", "name": "pitr", "reconcileID": "32ff1bcf-9445-4431-9fac-babbff01fb0f"} max retry count 120 reached. something went wrong with operator or kubernetes cluster