Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/logs/demand-backup-8-0.log Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 + main + create_infra demand-backup-28441 + local ns=demand-backup-28441 + '[' -n pxc-operator ']' + kubectl get pxc --all-namespaces -o wide + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + grep -v NAMESPACE + kubectl patch pxc -n demand-backup-29023 demand-backup --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/demand-backup patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.8JNVjdXi4E ++ mktemp + local LAST_ERR=/tmp/tmp.IgYKkZBUvD + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.8JNVjdXi4E perconaxtradbcluster.pxc.percona.com "demand-backup" deleted from demand-backup-29023 namespace + cat /tmp/tmp.IgYKkZBUvD + rm /tmp/tmp.8JNVjdXi4E /tmp/tmp.IgYKkZBUvD + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.r1e6uaXlNW ++ mktemp + local LAST_ERR=/tmp/tmp.RdaMKE7Xda + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.r1e6uaXlNW perconaxtradbclusterbackup.pxc.percona.com "on-demand-backup-minio" deleted from demand-backup-29023 namespace perconaxtradbclusterbackup.pxc.percona.com "on-demand-backup-pvc-bsource" deleted from demand-backup-29023 namespace perconaxtradbclusterbackup.pxc.percona.com "on-demand-backup-wrong-creds" deleted from demand-backup-29023 namespace + cat /tmp/tmp.RdaMKE7Xda + rm /tmp/tmp.r1e6uaXlNW /tmp/tmp.RdaMKE7Xda + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.jDyJfYAM29 ++ mktemp + local LAST_ERR=/tmp/tmp.XWyzK0MSbV + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.jDyJfYAM29 perconaxtradbclusterrestore.pxc.percona.com "on-demand-backup-minio" deleted from demand-backup-29023 namespace perconaxtradbclusterrestore.pxc.percona.com "on-demand-backup-pvc" deleted from demand-backup-29023 namespace perconaxtradbclusterrestore.pxc.percona.com "on-demand-backup-pvc-bsource" deleted from demand-backup-29023 namespace + cat /tmp/tmp.XWyzK0MSbV + rm /tmp/tmp.jDyJfYAM29 /tmp/tmp.XWyzK0MSbV + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ sed s/NAMESPACE// ++ tail -n1 ++ awk '-F ' '{print $2}' ++ helm list --all-namespaces --filter chaos-mesh + local chaos_mesh_ns= + '[' -n '' ']' ++ awk '{print $1}' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ grep validate-auth ++ kubectl get ValidatingWebhookConfiguration + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ grep chaos-mesh.org ++ awk '{print $1}' ++ kubectl get crd + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get clusterrolebinding + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get clusterrole + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + '[' -n '' ']' + desc 'cleaned up old namespaces pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pxc-operator ++ mktemp + xargs kubectl delete ns + awk '{print$1}' + kubectl_bin get ns ++ mktemp + local LAST_OUT=/tmp/tmp.qGDqehLz90 + grep -E -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' ++ mktemp + local LAST_ERR=/tmp/tmp.yBkBpBCcTg + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + local LAST_OUT=/tmp/tmp.csmFR1iQzW ++ mktemp + local LAST_ERR=/tmp/tmp.fqBcAv3G5b + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.qGDqehLz90 + cat /tmp/tmp.yBkBpBCcTg + rm /tmp/tmp.qGDqehLz90 /tmp/tmp.yBkBpBCcTg + return 0 namespace "demand-backup-29023" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.csmFR1iQzW namespace "pxc-operator" deleted + cat /tmp/tmp.fqBcAv3G5b + rm /tmp/tmp.csmFR1iQzW /tmp/tmp.fqBcAv3G5b + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + echo -n 'waiting for namespace/pxc-operator to be deleted' waiting for namespace/pxc-operator to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pxc-operator" not found + desc 'create namespace pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.F0eBltC9Aa ++ mktemp + local LAST_ERR=/tmp/tmp.WHkKpFo3Pk + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.F0eBltC9Aa namespace/pxc-operator created + cat /tmp/tmp.WHkKpFo3Pk + rm /tmp/tmp.F0eBltC9Aa /tmp/tmp.WHkKpFo3Pk + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.UeqFx80xeq +++ mktemp ++ local LAST_ERR=/tmp/tmp.LSDSMOa8fv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.UeqFx80xeq ++ cat /tmp/tmp.LSDSMOa8fv ++ rm /tmp/tmp.UeqFx80xeq /tmp/tmp.LSDSMOa8fv ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2382-13dc5c0a-12-cluster1 --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.jqlfXFVBC1 ++ mktemp + local LAST_ERR=/tmp/tmp.2ll5fbx59F + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2382-13dc5c0a-12-cluster1 --namespace=pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.jqlfXFVBC1 Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2382-13dc5c0a-12-cluster1" modified. + cat /tmp/tmp.2ll5fbx59F + rm /tmp/tmp.jqlfXFVBC1 /tmp/tmp.2ll5fbx59F + return 0 + deploy_operator + desc 'start PXC operator' + set +o xtrace ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.mfUUD2gG9c ++ mktemp + local LAST_ERR=/tmp/tmp.rmlBKyloWg + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.mfUUD2gG9c customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied + cat /tmp/tmp.rmlBKyloWg + rm /tmp/tmp.mfUUD2gG9c /tmp/tmp.rmlBKyloWg + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + kubectl_bin apply -f - + sed -e 's^namespace: .*^namespace: pxc-operator^' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/deploy/cw-rbac.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.qcJMnfg5uP ++ mktemp + local LAST_ERR=/tmp/tmp.jtmmUuNexq + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.qcJMnfg5uP clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.jtmmUuNexq + rm /tmp/tmp.qcJMnfg5uP /tmp/tmp.jtmmUuNexq + return 0 + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "LOG_LEVEL").value) = "VERBOSE"' - + sed -e 's^failureThreshold: .*^failureThreshold: 10^' + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-2382-13dc5c0a^' + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "PXCO_FEATURE_GATES").value) = ""' - + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "DISABLE_TELEMETRY").value) = "true"' - + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/deploy/cw-operator.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.qZceGGQLer ++ mktemp + local LAST_ERR=/tmp/tmp.AMIU9V1gDF + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.qZceGGQLer deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.AMIU9V1gDF + rm /tmp/tmp.qZceGGQLer /tmp/tmp.AMIU9V1gDF + return 0 + sleep 10 + kubectl_bin wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s ++ mktemp + local LAST_OUT=/tmp/tmp.d2yyjqfaYE ++ mktemp + local LAST_ERR=/tmp/tmp.pwNDXetp4d + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.d2yyjqfaYE pod/percona-xtradb-cluster-operator-54849d469f-m87wz condition met + cat /tmp/tmp.pwNDXetp4d + rm /tmp/tmp.d2yyjqfaYE /tmp/tmp.pwNDXetp4d + return 0 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ grep -c percona-xtradb-cluster-operator +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ jq -r '.items[] | select(.metadata.deletionTimestamp == null) | .metadata.name' ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator --field-selector=status.phase=Running -o json -n pxc-operator ++ head -1 +++ mktemp ++ local LAST_OUT=/tmp/tmp.VnXCW9sCJi +++ mktemp ++ local LAST_ERR=/tmp/tmp.BvhDRADQqS ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator --field-selector=status.phase=Running -o json -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.VnXCW9sCJi ++ cat /tmp/tmp.BvhDRADQqS ++ rm /tmp/tmp.VnXCW9sCJi /tmp/tmp.BvhDRADQqS ++ return 0 + wait_pod percona-xtradb-cluster-operator-54849d469f-m87wz 480 pxc-operator + local pod=percona-xtradb-cluster-operator-54849d469f-m87wz + local max_retry=480 + local ns=pxc-operator ++ echo percona-xtradb-cluster-operator-54849d469f-m87wz ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-54849d469f-m87wz condition met waiting for pod/percona-xtradb-cluster-operator-54849d469f-m87wz to become Ready.Ok + sleep 3 + create_namespace demand-backup-28441 + local namespace=demand-backup-28441 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl get clusterrolebinding ++ grep chaos-mesh + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + '[' -n '' ']' + desc 'cleaned up old namespaces demand-backup-28441' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces demand-backup-28441++ mktemp ----------------------------------------------------------------------------------- + kubectl_bin delete namespace demand-backup-28441 ++ mktemp + xargs kubectl delete ns + local LAST_OUT=/tmp/tmp.C8napCJfS4 + local LAST_OUT=/tmp/tmp.1Ee235GonD ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.I3ljpa9jsu + local exit_status=0 + local LAST_ERR=/tmp/tmp.3tOX8iwgdY + local exit_status=0 ++ seq 0 2 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace demand-backup-28441 + for i in '$(seq 0 2)' + set +e + kubectl get ns + grep -E -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + awk '{print$1}' + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace demand-backup-28441 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.1Ee235GonD + cat /tmp/tmp.3tOX8iwgdY + rm /tmp/tmp.1Ee235GonD /tmp/tmp.3tOX8iwgdY + return 0 error: resource(s) were provided, but no name was specified + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace demand-backup-28441 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.C8napCJfS4 + cat /tmp/tmp.I3ljpa9jsu Error from server (NotFound): namespaces "demand-backup-28441" not found + rm /tmp/tmp.C8napCJfS4 /tmp/tmp.I3ljpa9jsu + return 1 + : + wait_for_delete namespace/demand-backup-28441 + local res=namespace/demand-backup-28441 + echo -n 'waiting for namespace/demand-backup-28441 to be deleted' waiting for namespace/demand-backup-28441 to be deleted+ set +o xtrace Error from server (NotFound): namespaces "demand-backup-28441" not found + desc 'create namespace demand-backup-28441' + set +o xtrace ----------------------------------------------------------------------------------- create namespace demand-backup-28441 ----------------------------------------------------------------------------------- + kubectl_bin create namespace demand-backup-28441 ++ mktemp + local LAST_OUT=/tmp/tmp.X9PmS9LQ0j ++ mktemp + local LAST_ERR=/tmp/tmp.G0JXpSAuny + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace demand-backup-28441 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.X9PmS9LQ0j namespace/demand-backup-28441 created + cat /tmp/tmp.G0JXpSAuny + rm /tmp/tmp.X9PmS9LQ0j /tmp/tmp.G0JXpSAuny + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.nAtHZiRHHj +++ mktemp ++ local LAST_ERR=/tmp/tmp.QeT5NN7zT0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.nAtHZiRHHj ++ cat /tmp/tmp.QeT5NN7zT0 ++ rm /tmp/tmp.nAtHZiRHHj /tmp/tmp.QeT5NN7zT0 ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2382-13dc5c0a-12-cluster1 --namespace=demand-backup-28441 ++ mktemp + local LAST_OUT=/tmp/tmp.Jgf3VYBVll ++ mktemp + local LAST_ERR=/tmp/tmp.pYzv1Hg8xt + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2382-13dc5c0a-12-cluster1 --namespace=demand-backup-28441 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Jgf3VYBVll Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2382-13dc5c0a-12-cluster1" modified. + cat /tmp/tmp.pYzv1Hg8xt + rm /tmp/tmp.Jgf3VYBVll /tmp/tmp.pYzv1Hg8xt + return 0 + apply_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.yNj8rpjFpl ++ mktemp + local LAST_ERR=/tmp/tmp.3eg3feIK6u + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.yNj8rpjFpl secret/minio-secret created secret/aws-s3-secret created secret/do-spaces-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.3eg3feIK6u + rm /tmp/tmp.yNj8rpjFpl /tmp/tmp.3eg3feIK6u + return 0 + cluster=demand-backup + spinup_pxc demand-backup /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/demand-backup.yml + local cluster=demand-backup + local config=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/demand-backup.yml + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/conf/secrets.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/conf/client.yml + local port=3306 + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.Yprh7sgSqf ++ mktemp + local LAST_ERR=/tmp/tmp.WXlZBBlsQn + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/conf/secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Yprh7sgSqf secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created + cat /tmp/tmp.WXlZBBlsQn + rm /tmp/tmp.Yprh7sgSqf /tmp/tmp.WXlZBBlsQn + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/conf/client.yml + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/conf/client.yml + local pvc_name= + '[' -z '' ']' + kubectl_bin apply -f - + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/conf/client.yml '' + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/conf/client.yml + local pvc_name= + /usr/bin/sed -e s~minio-service.#namespace~minio-service.demand-backup-28441~ + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/fluentbit:main-logcollector#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2382-13dc5c0a#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' ++ mktemp + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/conf/client.yml + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + local LAST_OUT=/tmp/tmp.Ei2ESKRs8J ++ mktemp + local LAST_ERR=/tmp/tmp.QSCkztv8jI + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Ei2ESKRs8J deployment.apps/pxc-client created + cat /tmp/tmp.QSCkztv8jI + rm /tmp/tmp.Ei2ESKRs8J /tmp/tmp.QSCkztv8jI + return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/demand-backup.yml + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/demand-backup.yml + local pvc_name= + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/demand-backup.yml '' + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/demand-backup.yml + local pvc_name= + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/fluentbit:main-logcollector#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.demand-backup-28441~ + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/demand-backup.yml + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2382-13dc5c0a#' ++ mktemp + local LAST_OUT=/tmp/tmp.0Ks91rE2g9 ++ mktemp + local LAST_ERR=/tmp/tmp.HDUyKt01Jg + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.0Ks91rE2g9 perconaxtradbcluster.pxc.percona.com/demand-backup created + cat /tmp/tmp.HDUyKt01Jg + rm /tmp/tmp.0Ks91rE2g9 /tmp/tmp.HDUyKt01Jg + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- ++ get_proxy demand-backup ++ local target_cluster=demand-backup +++ kubectl_bin get pxc demand-backup -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.FEoaFsYJRE ++++ mktemp +++ local LAST_ERR=/tmp/tmp.kzr8hK8Wz1 +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc demand-backup -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.FEoaFsYJRE +++ cat /tmp/tmp.kzr8hK8Wz1 +++ rm /tmp/tmp.FEoaFsYJRE /tmp/tmp.kzr8hK8Wz1 +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc demand-backup -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Pmqz4s44JQ ++++ mktemp +++ local LAST_ERR=/tmp/tmp.bT94YVwcrb +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc demand-backup -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.Pmqz4s44JQ +++ cat /tmp/tmp.bT94YVwcrb +++ rm /tmp/tmp.Pmqz4s44JQ /tmp/tmp.bT94YVwcrb +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo demand-backup-proxysql ++ return + local proxy=demand-backup-proxysql + kubectl_bin wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n demand-backup-28441 ++ mktemp + local LAST_OUT=/tmp/tmp.RFx0GaTQ31 ++ mktemp + local LAST_ERR=/tmp/tmp.uQu0AUkYGC + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n demand-backup-28441 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n demand-backup-28441 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n demand-backup-28441 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.RFx0GaTQ31 + cat /tmp/tmp.uQu0AUkYGC error: no matching resources found + rm /tmp/tmp.RFx0GaTQ31 /tmp/tmp.uQu0AUkYGC + return 1 + true + wait_for_running demand-backup-proxysql 1 + local name=demand-backup-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod demand-backup-proxysql-0 480 + local pod=demand-backup-proxysql-0 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ echo demand-backup-proxysql-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=proxysql + set +o xtrace pod/demand-backup-proxysql-0 condition met waiting for pod/demand-backup-proxysql-0 to become Ready.Ok + wait_for_running demand-backup-pxc 3 + local name=demand-backup-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod demand-backup-pxc-0 480 + local pod=demand-backup-pxc-0 + local max_retry=480 + local ns= ++ echo demand-backup-pxc-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/demand-backup-pxc-0 condition met waiting for pod/demand-backup-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod demand-backup-pxc-1 480 + local pod=demand-backup-pxc-1 + local max_retry=480 + local ns= ++ echo demand-backup-pxc-1 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/demand-backup-pxc-1 condition met waiting for pod/demand-backup-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod demand-backup-pxc-2 480 + local pod=demand-backup-pxc-2 + local max_retry=480 + local ns= ++ echo demand-backup-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/demand-backup-pxc-2 condition met waiting for pod/demand-backup-pxc-2 to become Ready.Ok + sleep 10 ++ kubectl get pxc demand-backup -o 'jsonpath={.spec.secretsName}' + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.LowRj3K8cq +++ mktemp ++ local LAST_ERR=/tmp/tmp.axl2maATOl ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.LowRj3K8cq ++ cat /tmp/tmp.axl2maATOl ++ rm /tmp/tmp.LowRj3K8cq /tmp/tmp.axl2maATOl ++ return 0 + local root_pass=root_password + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' '-h demand-backup-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' + local 'uri=-h demand-backup-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uod24spUPZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.RtP8jNzyJ7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.uod24spUPZ ++ cat /tmp/tmp.RtP8jNzyJ7 ++ rm /tmp/tmp.uod24spUPZ /tmp/tmp.RtP8jNzyJ7 ++ return 0 + client_pod=pxc-client-67fc4995bb-jwx2s + wait_pod pxc-client-67fc4995bb-jwx2s + local pod=pxc-client-67fc4995bb-jwx2s + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-jwx2s ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-jwx2s condition met waiting for pod/pxc-client-67fc4995bb-jwx2s to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h demand-backup-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h demand-backup-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VDiaInvHCc +++ mktemp ++ local LAST_ERR=/tmp/tmp.y5Z4YVodCX ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.VDiaInvHCc ++ cat /tmp/tmp.y5Z4YVodCX ++ rm /tmp/tmp.VDiaInvHCc /tmp/tmp.y5Z4YVodCX ++ return 0 + client_pod=pxc-client-67fc4995bb-jwx2s + wait_pod pxc-client-67fc4995bb-jwx2s + local pod=pxc-client-67fc4995bb-jwx2s + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-67fc4995bb-jwx2s ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-jwx2s condition met waiting for pod/pxc-client-67fc4995bb-jwx2s to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-0.demand-backup-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-0.demand-backup-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-0.demand-backup-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-0.demand-backup-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rc2wJGb0A2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.IacASN3sSO ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.rc2wJGb0A2 ++ cat /tmp/tmp.IacASN3sSO ++ rm /tmp/tmp.rc2wJGb0A2 /tmp/tmp.IacASN3sSO ++ return 0 + client_pod=pxc-client-67fc4995bb-jwx2s + wait_pod pxc-client-67fc4995bb-jwx2s + local pod=pxc-client-67fc4995bb-jwx2s + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-jwx2s ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-jwx2s condition met waiting for pod/pxc-client-67fc4995bb-jwx2s to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.14SIjS2s6f/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.14SIjS2s6f/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1.sql /tmp/tmp.14SIjS2s6f/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-1.demand-backup-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-1.demand-backup-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-1.demand-backup-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-1.demand-backup-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.G5JNIPXjzM +++ mktemp ++ local LAST_ERR=/tmp/tmp.yVKGFbMnCp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.G5JNIPXjzM ++ cat /tmp/tmp.yVKGFbMnCp ++ rm /tmp/tmp.G5JNIPXjzM /tmp/tmp.yVKGFbMnCp ++ return 0 + client_pod=pxc-client-67fc4995bb-jwx2s + wait_pod pxc-client-67fc4995bb-jwx2s + local pod=pxc-client-67fc4995bb-jwx2s + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-jwx2s ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-jwx2s condition met waiting for pod/pxc-client-67fc4995bb-jwx2s to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.14SIjS2s6f/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.14SIjS2s6f/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1.sql /tmp/tmp.14SIjS2s6f/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-2.demand-backup-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-2.demand-backup-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-2.demand-backup-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-2.demand-backup-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4omehS8VH0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.sF8HE5Us2Y ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.4omehS8VH0 ++ cat /tmp/tmp.sF8HE5Us2Y ++ rm /tmp/tmp.4omehS8VH0 /tmp/tmp.sF8HE5Us2Y ++ return 0 + client_pod=pxc-client-67fc4995bb-jwx2s + wait_pod pxc-client-67fc4995bb-jwx2s + local pod=pxc-client-67fc4995bb-jwx2s + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-jwx2s ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-jwx2s condition met waiting for pod/pxc-client-67fc4995bb-jwx2s to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.14SIjS2s6f/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.14SIjS2s6f/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1.sql /tmp/tmp.14SIjS2s6f/select-1.sql + is_keyring_plugin_in_use demand-backup + local cluster=demand-backup + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + kubectl exec demand-backup-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' + grep -E -o 'early-plugin-load=keyring_\w+.so' + return 1 + start_minio demand-backup-ssl + deploy_helm demand-backup-28441 + helm repo add hashicorp https://helm.releases.hashicorp.com "hashicorp" already exists with the same configuration, skipping + helm repo add minio https://charts.min.io/ "minio" already exists with the same configuration, skipping + helm repo update Hang tight while we grab the latest from your chart repositories... ...Successfully got an update from the "hashicorp" chart repository ...Successfully got an update from the "minio" chart repository ...Successfully got an update from the "percona" chart repository ...Successfully got an update from the "chaos-mesh" chart repository Update Complete. ⎈Happy Helming!⎈ + local cert_secret=demand-backup-ssl + local endpoint=http://minio-service:9000 + minio_args=(--version $MINIO_VER --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set "users[0].accessKey=some-access-key" --set "users[0].secretKey=some-secret-key" --set "users[0].policy=consoleAdmin" --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G) + local minio_args + [[ -n demand-backup-ssl ]] + endpoint=https://minio-service:9000 + minio_args+=(--set tls.enabled=true --set tls.certSecret="$cert_secret" --set tls.publicCrt=tls.crt --set tls.privateKey=tls.key) + desc 'install Minio' + set +o xtrace ----------------------------------------------------------------------------------- install Minio ----------------------------------------------------------------------------------- + helm uninstall minio-service Error: uninstall: Release not loaded: minio-service: release: not found + : + retry 10 60 helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G --set tls.enabled=true --set tls.certSecret=demand-backup-ssl --set tls.publicCrt=tls.crt --set tls.privateKey=tls.key minio/minio + local max=10 + local delay=60 + shift 2 + local n=1 + helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G --set tls.enabled=true --set tls.certSecret=demand-backup-ssl --set tls.publicCrt=tls.crt --set tls.privateKey=tls.key minio/minio NAME: minio-service LAST DEPLOYED: Mon Mar 30 10:11:42 2026 NAMESPACE: demand-backup-28441 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: MinIO can be accessed via port 9000 on the following DNS name from within your cluster: minio-service.demand-backup-28441.cluster.local To access MinIO from localhost, run the below commands: 1. export POD_NAME=$(kubectl get pods --namespace demand-backup-28441 -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}") 2. kubectl port-forward $POD_NAME 9000 --namespace demand-backup-28441 Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/ You can now access MinIO server on http://localhost:9000. Follow the below steps to connect to MinIO server with mc client: 1. Download the MinIO mc client - https://min.io/docs/minio/linux/reference/minio-mc.html#quickstart 2. export MC_HOST_minio-service-local=http://$(kubectl get secret --namespace demand-backup-28441 minio-service -o jsonpath="{.data.rootUser}" | base64 --decode):$(kubectl get secret --namespace demand-backup-28441 minio-service -o jsonpath="{.data.rootPassword}" | base64 --decode)@localhost:9000 3. mc ls minio-service-local + sleep 30 ++ kubectl_bin get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2iCMlCPllz +++ mktemp ++ local LAST_ERR=/tmp/tmp.CJkpr4cmQR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2iCMlCPllz ++ cat /tmp/tmp.CJkpr4cmQR ++ rm /tmp/tmp.2iCMlCPllz /tmp/tmp.CJkpr4cmQR ++ return 0 + MINIO_POD=minio-service-b58fd8fbf-9glwz + wait_pod minio-service-b58fd8fbf-9glwz + local pod=minio-service-b58fd8fbf-9glwz + local max_retry=480 + local ns= ++ echo minio-service-b58fd8fbf-9glwz ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/minio-service-b58fd8fbf-9glwz condition met waiting for pod/minio-service-b58fd8fbf-9glwz to become Ready.Ok + kubectl_bin run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- /usr/bin/env AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url https://minio-service:9000 --no-verify-ssl s3 mb s3://operator-testing ++ mktemp + local LAST_OUT=/tmp/tmp.hueS7E0jbt ++ mktemp + local LAST_ERR=/tmp/tmp.qaXxjOyz6n + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- /usr/bin/env AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url https://minio-service:9000 --no-verify-ssl s3 mb s3://operator-testing + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.hueS7E0jbt /usr/lib/python2.7/site-packages/botocore/vendored/requests/packages/urllib3/connectionpool.py:768: InsecureRequestWarning: Unverified HTTPS request is being made. Adding certificate verification is strongly advised. See: https://urllib3.readthedocs.org/en/latest/security.html InsecureRequestWarning) make_bucket: operator-testing pod "aws-cli" deleted from demand-backup-28441 namespace + cat /tmp/tmp.qaXxjOyz6n All commands and output from this session will be recorded in container logs, including credentials and sensitive information passed through the command prompt. If you don't see a command prompt, try pressing enter. warning: couldn't attach to pod/aws-cli, falling back to streaming logs: Internal error occurred: unable to upgrade connection: container aws-cli not found in pod aws-cli_demand-backup-28441 + rm /tmp/tmp.hueS7E0jbt /tmp/tmp.qaXxjOyz6n + return 0 + run_backup demand-backup on-demand-backup-pvc + local cluster=demand-backup + local backup=on-demand-backup-pvc + log 'run pxc-backup/on-demand-backup-pvc' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-03-30T10:12:51+0000]' run pxc-backup/on-demand-backup-pvc [2026-03-30T10:12:51+0000] run pxc-backup/on-demand-backup-pvc + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/on-demand-backup-pvc.yml ++ mktemp + local LAST_OUT=/tmp/tmp.uld2dVZbdw ++ mktemp + local LAST_ERR=/tmp/tmp.yqI7utKli3 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/on-demand-backup-pvc.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.uld2dVZbdw perconaxtradbclusterbackup.pxc.percona.com/on-demand-backup-pvc created + cat /tmp/tmp.yqI7utKli3 + rm /tmp/tmp.uld2dVZbdw /tmp/tmp.yqI7utKli3 + return 0 + wait_backup on-demand-backup-pvc + local backup=on-demand-backup-pvc + local status=Succeeded + set +o xtrace waiting for pxc-backup/on-demand-backup-pvc to reach Succeeded state.........................Succeeded + run_recovery_check demand-backup on-demand-backup-pvc + local cluster=demand-backup + local backup=on-demand-backup-pvc ++ get_proxy_engine demand-backup ++ local cluster_name=demand-backup +++ get_proxy demand-backup +++ local target_cluster=demand-backup ++++ kubectl_bin get pxc demand-backup -o 'jsonpath={.spec.haproxy.enabled}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.Fe00PFcV5Q +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.brsWHPbxNp ++++ local exit_status=0 +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ set +e ++++ kubectl get pxc demand-backup -o 'jsonpath={.spec.haproxy.enabled}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 ']' ++++ break ++++ cat /tmp/tmp.Fe00PFcV5Q ++++ cat /tmp/tmp.brsWHPbxNp ++++ rm /tmp/tmp.Fe00PFcV5Q /tmp/tmp.brsWHPbxNp ++++ return 0 +++ [[ '' == \t\r\u\e ]] ++++ kubectl_bin get pxc demand-backup -o 'jsonpath={.spec.proxysql.enabled}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.w5G6o6EPf4 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.NS40Jd6EhT ++++ local exit_status=0 +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ set +e ++++ kubectl get pxc demand-backup -o 'jsonpath={.spec.proxysql.enabled}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 ']' ++++ break ++++ cat /tmp/tmp.w5G6o6EPf4 ++++ cat /tmp/tmp.NS40Jd6EhT ++++ rm /tmp/tmp.w5G6o6EPf4 /tmp/tmp.NS40Jd6EhT ++++ return 0 +++ [[ true == \t\r\u\e ]] +++ echo demand-backup-proxysql +++ return ++ local cluster_proxy=demand-backup-proxysql ++ echo proxysql + local proxy=proxysql ++ get_pvc_name_for_backup on-demand-backup-pvc ++ local backup_name=on-demand-backup-pvc +++ kubectl_bin get pxc-backup on-demand-backup-pvc -o 'jsonpath={.status.destination}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.HkwNjLgt0n ++++ mktemp +++ local LAST_ERR=/tmp/tmp.uPnmx5SeLm +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc-backup on-demand-backup-pvc -o 'jsonpath={.status.destination}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.HkwNjLgt0n +++ cat /tmp/tmp.uPnmx5SeLm +++ rm /tmp/tmp.HkwNjLgt0n /tmp/tmp.uPnmx5SeLm +++ return 0 ++ local destination=pvc/xb-on-demand-backup-pvc-20260330101253-2d0422b9 ++ [[ -z pvc/xb-on-demand-backup-pvc-20260330101253-2d0422b9 ]] ++ local pvc_name=xb-on-demand-backup-pvc-20260330101253-2d0422b9 ++ echo xb-on-demand-backup-pvc-20260330101253-2d0422b9 + pvc_name=xb-on-demand-backup-pvc-20260330101253-2d0422b9 + dest=pvc/xb-on-demand-backup-pvc-20260330101253-2d0422b9 + log 'run pxc-restore/on-demand-backup-pvc' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-03-30T10:13:38+0000]' run pxc-restore/on-demand-backup-pvc [2026-03-30T10:13:38+0000] run pxc-restore/on-demand-backup-pvc + YAML_INPUT_FILE=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/restore-on-demand-backup-pvc.yaml ++ cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/restore-on-demand-backup-pvc.yaml ++ /usr/bin/sed -e s~minio-service.#namespace~minio-service.demand-backup-28441~ ++ /usr/bin/sed -e 's##pvc/xb-on-demand-backup-pvc-20260330101253-2d0422b9#g' + MODIFIED_YAML='apiVersion: pxc.percona.com/v1 kind: PerconaXtraDBClusterRestore metadata: name: on-demand-backup-pvc spec: pxcCluster: demand-backup backupName: on-demand-backup-pvc resources: requests: memory: 0.5G cpu: 500m limits: memory: "3G" cpu: "2"' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + FINAL_YAML='apiVersion: pxc.percona.com/v1 kind: PerconaXtraDBClusterRestore metadata: name: on-demand-backup-pvc spec: pxcCluster: demand-backup backupName: on-demand-backup-pvc resources: requests: memory: 0.5G cpu: 500m limits: memory: "3G" cpu: "2"' + echo 'apiVersion: pxc.percona.com/v1 kind: PerconaXtraDBClusterRestore metadata: name: on-demand-backup-pvc spec: pxcCluster: demand-backup backupName: on-demand-backup-pvc resources: requests: memory: 0.5G cpu: 500m limits: memory: "3G" cpu: "2"' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.YUOXW1s5Ln ++ mktemp + local LAST_ERR=/tmp/tmp.l3v1qfigIc + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.YUOXW1s5Ln perconaxtradbclusterrestore.pxc.percona.com/on-demand-backup-pvc created + cat /tmp/tmp.l3v1qfigIc + rm /tmp/tmp.YUOXW1s5Ln /tmp/tmp.l3v1qfigIc + return 0 + wait_backup_restore on-demand-backup-pvc + local backup_name=on-demand-backup-pvc + local target_state=Succeeded + local wait_time=720 + set +o xtrace waiting for pxc-restore/on-demand-backup-pvc to reach Succeeded state 2026-03-30T10:13:42 pxc-restore/on-demand-backup-pvc state: Starting 2026-03-30T10:13:44 pxc-restore/on-demand-backup-pvc state: Starting 2026-03-30T10:13:47 pxc-restore/on-demand-backup-pvc state: Starting 2026-03-30T10:13:50 pxc-restore/on-demand-backup-pvc state: Starting 2026-03-30T10:13:53 pxc-restore/on-demand-backup-pvc state: Starting 2026-03-30T10:13:56 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:13:59 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:14:01 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:14:04 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:14:06 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:14:08 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:14:10 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:14:12 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:14:15 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:14:17 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:14:19 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:14:21 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:14:24 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:14:26 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:14:28 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:14:30 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:14:33 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:14:35 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:14:37 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:14:40 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:14:42 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:14:45 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:14:48 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:14:50 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:14:52 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:14:54 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2026-03-30T10:14:57 pxc-restore/on-demand-backup-pvc state: Restoring 2026-03-30T10:14:59 pxc-restore/on-demand-backup-pvc state: Restoring 2026-03-30T10:15:01 pxc-restore/on-demand-backup-pvc state: Restoring 2026-03-30T10:15:03 pxc-restore/on-demand-backup-pvc state: Restoring 2026-03-30T10:15:05 pxc-restore/on-demand-backup-pvc state: Restoring 2026-03-30T10:15:07 pxc-restore/on-demand-backup-pvc state: Restoring 2026-03-30T10:15:09 pxc-restore/on-demand-backup-pvc state: Restoring 2026-03-30T10:15:11 pxc-restore/on-demand-backup-pvc state: Restoring 2026-03-30T10:15:14 pxc-restore/on-demand-backup-pvc state: Restoring 2026-03-30T10:15:16 pxc-restore/on-demand-backup-pvc state: Restoring 2026-03-30T10:15:18 pxc-restore/on-demand-backup-pvc state: Restoring 2026-03-30T10:15:20 pxc-restore/on-demand-backup-pvc state: Restoring 2026-03-30T10:15:22 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:15:24 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:15:26 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:15:28 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:15:30 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:15:32 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:15:34 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:15:37 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:15:39 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:15:41 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:15:43 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:15:45 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:15:47 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:15:49 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:15:51 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:15:53 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:15:56 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:15:58 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:16:01 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:16:03 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:16:06 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:16:08 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:16:10 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:16:12 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:16:15 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:16:17 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:16:19 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:16:22 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:16:24 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:16:26 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:16:28 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:16:31 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:16:33 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:16:35 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2026-03-30T10:16:37 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:16:40 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:16:42 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:16:44 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:16:46 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:16:49 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:16:51 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:16:54 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:16:57 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:16:59 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:17:01 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:17:04 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:17:06 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:17:09 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:17:12 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:17:14 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:17:17 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:17:19 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:17:22 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:17:25 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:17:27 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:17:29 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:17:31 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:17:34 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:17:37 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:17:39 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:17:42 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:17:44 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:17:47 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:17:49 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:17:51 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:17:54 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:17:56 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:17:58 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:18:01 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:18:03 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:18:06 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:18:09 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:18:12 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:18:14 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:18:17 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:18:20 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:18:22 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:18:25 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:18:27 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:18:30 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:18:32 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:18:35 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:18:37 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:18:39 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:18:42 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:18:43 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:18:45 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:18:48 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:18:50 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:18:52 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:18:55 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:18:57 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:18:59 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:01 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:03 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:05 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:07 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:09 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:11 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:14 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:16 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:18 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:20 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:22 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:24 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:26 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:28 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:30 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:32 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:34 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:36 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:39 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:41 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:43 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:45 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:47 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:49 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:51 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:53 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:55 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:19:58 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:20:00 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:20:02 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2026-03-30T10:20:04 pxc-restore/on-demand-backup-pvc state: Succeeded + kubectl_bin logs job/restore-job-on-demand-backup-pvc-demand-backup ++ mktemp + local LAST_OUT=/tmp/tmp.sJIGcUzX4i ++ mktemp + local LAST_ERR=/tmp/tmp.r1Owkol9aM + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl logs job/restore-job-on-demand-backup-pvc-demand-backup + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.sJIGcUzX4i + LIB_PATH=/opt/percona/backup/lib/pxc + . /opt/percona/backup/lib/pxc/check-version.sh + . /opt/percona/backup/lib/pxc/vault.sh ++ set -o errexit ++ keyring_vault=/etc/mysql/vault-keyring-secret/keyring_vault.conf + SOCAT_OPTS=TCP:restore-src-on-demand-backup-pvc-demand-backup:3307,retry=30 + check_ssl + CA=/var/run/secrets/kubernetes.io/serviceaccount/ca.crt + '[' -f /var/run/secrets/kubernetes.io/serviceaccount/service-ca.crt ']' + SSL_DIR=/etc/mysql/ssl + '[' -f /etc/mysql/ssl/ca.crt ']' + CA=/etc/mysql/ssl/ca.crt + SSL_INTERNAL_DIR=/etc/mysql/ssl-internal + '[' -f /etc/mysql/ssl-internal/ca.crt ']' + CA=/etc/mysql/ssl-internal/ca.crt + KEY=/etc/mysql/ssl/tls.key + CERT=/etc/mysql/ssl/tls.crt + '[' -f /etc/mysql/ssl-internal/tls.key ']' + '[' -f /etc/mysql/ssl-internal/tls.crt ']' + KEY=/etc/mysql/ssl-internal/tls.key + CERT=/etc/mysql/ssl-internal/tls.crt + '[' -f /etc/mysql/ssl-internal/ca.crt ']' + '[' -f /etc/mysql/ssl-internal/tls.key ']' + '[' -f /etc/mysql/ssl-internal/tls.crt ']' + SOCAT_OPTS='openssl-connect:restore-src-on-demand-backup-pvc-demand-backup:3307,reuseaddr,cert=/etc/mysql/ssl-internal/tls.crt,key=/etc/mysql/ssl-internal/tls.key,cafile=/etc/mysql/ssl-internal/ca.crt,verify=1,commonname='\'''\'',retry=30,no-sni=1' + ping -c1 restore-src-on-demand-backup-pvc-demand-backup /opt/percona/backup/recovery-pvc-joiner.sh: line 40: ping: command not found + : + rm -rf /datadir/#ib_16384_0.dblwr /datadir/#ib_16384_1.dblwr /datadir/#ib_16384_10.dblwr /datadir/#ib_16384_11.dblwr /datadir/#ib_16384_12.dblwr /datadir/#ib_16384_13.dblwr /datadir/#ib_16384_14.dblwr /datadir/#ib_16384_15.dblwr /datadir/#ib_16384_2.dblwr /datadir/#ib_16384_3.dblwr /datadir/#ib_16384_4.dblwr /datadir/#ib_16384_5.dblwr /datadir/#ib_16384_6.dblwr /datadir/#ib_16384_7.dblwr /datadir/#ib_16384_8.dblwr /datadir/#ib_16384_9.dblwr /datadir/#innodb_redo /datadir/#innodb_temp /datadir/auth_plugin /datadir/auto.cnf /datadir/binlog.000001 /datadir/binlog.000002 /datadir/binlog.000003 /datadir/binlog.index /datadir/galera.cache /datadir/get-pxc-state /datadir/grastate.dat /datadir/ib_buffer_pool /datadir/ibdata1 /datadir/innobackup.backup.log /datadir/liveness-check.sh /datadir/myApp /datadir/mysql /datadir/mysql-state-monitor /datadir/mysql-state-monitor.log /datadir/mysql.ibd /datadir/mysql.state /datadir/notify.sock /datadir/peer-list /datadir/performance_schema /datadir/pmm-prerun.sh /datadir/private_key.pem /datadir/public_key.pem /datadir/pxc-configure-pxc.sh /datadir/pxc-entrypoint.sh /datadir/readiness-check.sh /datadir/sys /datadir/undo_001 /datadir/undo_002 /datadir/version_info /datadir/wsrep_cmd_notify_handler.sh ++ mktemp --directory /datadir/pxc_sst_XXXX + tmp=/datadir/pxc_sst_g1Ta + socat -u 'openssl-connect:restore-src-on-demand-backup-pvc-demand-backup:3307,reuseaddr,cert=/etc/mysql/ssl-internal/tls.crt,key=/etc/mysql/ssl-internal/tls.key,cafile=/etc/mysql/ssl-internal/ca.crt,verify=1,commonname='\'''\'',retry=30,no-sni=1' stdio ++ get_xtrabackup_version +++ xtrabackup --version +++ grep '^xtrabackup version' +++ awk '{print $3}' +++ sed 's/-.*//' ++ xtrabackup_ver=8.0.35 ++ echo 8.0.35 + XTRABACKUP_VERSION=8.0.35 + check_for_version 8.0.35 8.0.0 + '[' -z 8.0.35 ']' + '[' -z 8.0.0 ']' + local local_version_str + local required_version_str ++ normalize_version 8.0.35 ++ local major=0 ++ local minor=0 ++ local patch=0 ++ [[ 8.0.35 =~ ^([0-9]+)\.([0-9]+)\.?([0-9]*)([^ ])* ]] ++ major=8 ++ minor=0 ++ patch=35 ++ printf %02d%02d%02d 8 0 35 + local_version_str=080035 ++ normalize_version 8.0.0 ++ local major=0 ++ local minor=0 ++ local patch=0 ++ [[ 8.0.0 =~ ^([0-9]+)\.([0-9]+)\.?([0-9]*)([^ ])* ]] ++ major=8 ++ minor=0 ++ patch=0 ++ printf %02d%02d%02d 8 0 0 + required_version_str=080000 + [[ 080035 < 080000 ]] + return 0 + XBSTREAM_EXTRA_ARGS=' --decompress' + socat -u 'openssl-connect:restore-src-on-demand-backup-pvc-demand-backup:3307,reuseaddr,cert=/etc/mysql/ssl-internal/tls.crt,key=/etc/mysql/ssl-internal/tls.key,cafile=/etc/mysql/ssl-internal/ca.crt,verify=1,commonname='\'''\'',retry=30,no-sni=1' stdio ++ grep -c processor /proc/cpuinfo + xbstream -x -C /datadir/pxc_sst_g1Ta --parallel=4 --decompress + PXB_VAULT_PREPARE_ARGS= + PXB_VAULT_MOVEBACK_ARGS= + VAULT_CONFIG_FILE=/etc/mysql/vault-keyring-secret/keyring_vault.conf + VAULT_KEYRING_COMPONENT=/opt/percona/component_keyring_vault.cnf + [[ -f /etc/mysql/vault-keyring-secret/keyring_vault.conf ]] + set +o xtrace vault configuration not found + xtrabackup --use-memory=2GB --prepare --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_g1Ta 2026-03-30T10:15:11.456398-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized server arguments: --innodb_checksum_algorithm=crc32 --innodb_log_checksums=1 --innodb_data_file_path=ibdata1:12M:autoextend --innodb_log_file_size=50331648 --innodb_page_size=16384 --innodb_undo_directory=./ --innodb_undo_tablespaces=2 --server-id=24613582 --innodb_log_checksums=ON --innodb_redo_log_encrypt=0 --innodb_undo_log_encrypt=0 2026-03-30T10:15:11.456473-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized client arguments: --use-memory=2GB --prepare=1 --rollback-prepared-trx=1 --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_g1Ta xtrabackup version 8.0.35-34 based on MySQL server 8.0.35 Linux (x86_64) (revision id: c8a25ff9) 2026-03-30T10:15:11.456510-00:00 0 [Note] [MY-011825] [Xtrabackup] cd to /datadir/pxc_sst_g1Ta/ 2026-03-30T10:15:11.456573-00:00 0 [Note] [MY-011825] [Xtrabackup] This target seems to be not prepared yet. 2026-03-30T10:15:11.466254-00:00 0 [Note] [MY-011825] [Xtrabackup] xtrabackup_logfile detected: size=8388608, start_lsn=(30354336) 2026-03-30T10:15:11.467043-00:00 0 [Note] [MY-011825] [Xtrabackup] using the following InnoDB configuration for recovery: 2026-03-30T10:15:11.467056-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_home_dir = . 2026-03-30T10:15:11.467060-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_file_path = ibdata1:12M:autoextend 2026-03-30T10:15:11.467082-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_group_home_dir = . 2026-03-30T10:15:11.467092-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_files_in_group = 1 2026-03-30T10:15:11.467099-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_file_size = 8388608 2026-03-30T10:15:11.467263-00:00 0 [Note] [MY-011825] [Xtrabackup] inititialize_service_handles suceeded 2026-03-30T10:15:11.467418-00:00 0 [Note] [MY-011825] [Xtrabackup] using the following InnoDB configuration for recovery: 2026-03-30T10:15:11.467430-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_home_dir = . 2026-03-30T10:15:11.467433-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_file_path = ibdata1:12M:autoextend 2026-03-30T10:15:11.467439-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_group_home_dir = . 2026-03-30T10:15:11.467443-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_files_in_group = 1 2026-03-30T10:15:11.467449-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_file_size = 8388608 2026-03-30T10:15:11.467460-00:00 0 [Note] [MY-011825] [Xtrabackup] Starting InnoDB instance for recovery. 2026-03-30T10:15:11.467470-00:00 0 [Note] [MY-011825] [Xtrabackup] Using 2147483648 bytes for buffer pool (set by --use-memory parameter) 2026-03-30T10:15:11.467499-00:00 0 [Note] [MY-012932] [InnoDB] PUNCH HOLE support available 2026-03-30T10:15:11.467518-00:00 0 [Note] [MY-012944] [InnoDB] Uses event mutexes 2026-03-30T10:15:11.467522-00:00 0 [Note] [MY-012945] [InnoDB] GCC builtin __atomic_thread_fence() is used for memory barrier 2026-03-30T10:15:11.467527-00:00 0 [Note] [MY-012948] [InnoDB] Compressed tables use zlib 1.2.13 2026-03-30T10:15:11.467680-00:00 0 [Note] [MY-012951] [InnoDB] Using hardware accelerated crc32 and polynomial multiplication. 2026-03-30T10:15:11.468024-00:00 0 [Note] [MY-012203] [InnoDB] Directories to scan './' 2026-03-30T10:15:11.468060-00:00 0 [Note] [MY-012204] [InnoDB] Scanning './' 2026-03-30T10:15:11.477264-00:00 0 [Note] [MY-012208] [InnoDB] Completed space ID check of 8 files. 2026-03-30T10:15:11.477824-00:00 0 [Note] [MY-012955] [InnoDB] Initializing buffer pool, total size = 2.000000G, instances = 1, chunk size =128.000000M 2026-03-30T10:15:11.582560-00:00 0 [Note] [MY-012957] [InnoDB] Completed initialization of buffer pool 2026-03-30T10:15:11.592091-00:00 0 [Note] [MY-011952] [InnoDB] If the mysqld execution user is authorized, page cleaner thread priority can be changed. See the man page of setpriority(). 2026-03-30T10:15:11.620529-00:00 0 [Note] [MY-013883] [InnoDB] The latest found checkpoint is at lsn = 30354336 in redo log file ./#innodb_redo/#ib_redo0. 2026-03-30T10:15:11.620572-00:00 0 [Note] [MY-012560] [InnoDB] The log sequence number 30235607 in the system tablespace does not match the log sequence number 30354336 in the redo log files! 2026-03-30T10:15:11.620581-00:00 0 [Note] [MY-012551] [InnoDB] Database was not shutdown normally! 2026-03-30T10:15:11.620587-00:00 0 [Note] [MY-012552] [InnoDB] Starting crash recovery. 2026-03-30T10:15:11.620772-00:00 0 [Note] [MY-013086] [InnoDB] Starting to parse redo log at lsn = 30354017, whereas checkpoint_lsn = 30354336 and start_lsn = 30353920 2026-03-30T10:15:11.620783-00:00 0 [Note] [MY-012550] [InnoDB] Doing recovery: scanned up to log sequence number 30354346 2026-03-30T10:15:11.640125-00:00 0 [Note] [MY-013083] [InnoDB] Log background threads are being started... 2026-03-30T10:15:11.644903-00:00 0 [Note] [MY-012532] [InnoDB] Applying a batch of 1 redo log records ... 2026-03-30T10:15:11.644957-00:00 0 [Note] [MY-012533] [InnoDB] 100% 2026-03-30T10:15:11.644969-00:00 0 [Note] [MY-012535] [InnoDB] Apply batch completed! 2026-03-30T10:15:11.748317-00:00 0 [Note] [MY-013084] [InnoDB] Log background threads are being closed... 2026-03-30T10:15:11.749746-00:00 0 [Note] [MY-013888] [InnoDB] Upgrading redo log: 1032M, LSN=30354346. 2026-03-30T10:15:11.749757-00:00 0 [Note] [MY-012968] [InnoDB] Starting to delete and rewrite redo log files. 2026-03-30T10:15:11.749802-00:00 0 [Note] [MY-011825] [InnoDB] Removing redo log file: ./#innodb_redo/#ib_redo0 2026-03-30T10:15:11.779009-00:00 0 [Note] [MY-011825] [InnoDB] Creating redo log file at ./#innodb_redo/#ib_redo0_tmp with file_id 0 with size 33554432 bytes 2026-03-30T10:15:11.784950-00:00 0 [Note] [MY-011825] [InnoDB] Renaming redo log file from ./#innodb_redo/#ib_redo0_tmp to ./#innodb_redo/#ib_redo0 2026-03-30T10:15:11.790164-00:00 0 [Note] [MY-012893] [InnoDB] New redo log files created, LSN=30354444 2026-03-30T10:15:11.790221-00:00 0 [Note] [MY-013083] [InnoDB] Log background threads are being started... 2026-03-30T10:15:11.792875-00:00 0 [Note] [MY-013252] [InnoDB] Using undo tablespace './undo_001'. 2026-03-30T10:15:11.794196-00:00 0 [Note] [MY-013252] [InnoDB] Using undo tablespace './undo_002'. 2026-03-30T10:15:11.795764-00:00 0 [Note] [MY-012910] [InnoDB] Opened 2 existing undo tablespaces. 2026-03-30T10:15:11.795814-00:00 0 [Note] [MY-011980] [InnoDB] GTID recovery trx_no: 5200 2026-03-30T10:15:11.933808-00:00 0 [Note] [MY-013776] [InnoDB] Parallel initialization of rseg complete 2026-03-30T10:15:11.933828-00:00 0 [Note] [MY-013777] [InnoDB] Time taken to initialize rseg using 4 thread: 138016 ms. 2026-03-30T10:15:11.933875-00:00 0 [Note] [MY-012923] [InnoDB] Creating shared tablespace for temporary tables 2026-03-30T10:15:11.933938-00:00 0 [Note] [MY-012265] [InnoDB] Setting file './ibtmp1' size to 12 MB. Physically writing the file full; Please wait ... 2026-03-30T10:15:11.966528-00:00 0 [Note] [MY-012266] [InnoDB] File './ibtmp1' size is now 12 MB. 2026-03-30T10:15:11.966655-00:00 0 [Note] [MY-013627] [InnoDB] Scanning temp tablespace dir:'./#innodb_temp/' 2026-03-30T10:15:12.000330-00:00 0 [Note] [MY-013018] [InnoDB] Created 128 and tracked 128 new rollback segment(s) in the temporary tablespace. 128 are now active. 2026-03-30T10:15:12.000840-00:00 0 [Note] [MY-012976] [InnoDB] 8.0.35 started; log sequence number 30354454 2026-03-30T10:15:12.002057-00:00 0 [Warning] [MY-012091] [InnoDB] Allocated tablespace ID 1 for sys/sys_config, old maximum was 0 2026-03-30T10:15:12.007668-00:00 0 [Note] [MY-011825] [Xtrabackup] Completed loading of 6 tablespaces into cache in 0.00679878 seconds 2026-03-30T10:15:12.028626-00:00 0 [Note] [MY-011825] [Xtrabackup] Time taken to build dictionary: 0.0209407 seconds 2026-03-30T10:15:13.029075-00:00 0 [Note] [MY-011825] [Xtrabackup] Recovered WSREP position: 26c347d0-2c20-11f1-bbe6-5fc7a544dde9:42 2026-03-30T10:15:13.029122-00:00 0 [Note] [MY-011825] [Xtrabackup] starting shutdown with innodb_fast_shutdown = 1 2026-03-30T10:15:13.029195-00:00 0 [Note] [MY-012330] [InnoDB] FTS optimize thread exiting. 2026-03-30T10:15:14.029048-00:00 0 [Note] [MY-013072] [InnoDB] Starting shutdown... 2026-03-30T10:15:14.129906-00:00 0 [Note] [MY-013084] [InnoDB] Log background threads are being closed... 2026-03-30T10:15:14.165959-00:00 0 [Note] [MY-012980] [InnoDB] Shutdown completed; log sequence number 30354454 2026-03-30T10:15:14.169034-00:00 0 [Note] [MY-011825] [Xtrabackup] completed OK! + xtrabackup --defaults-group=mysqld --datadir=/datadir --move-back --force-non-empty-directories --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_g1Ta 2026-03-30T10:15:14.192481-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized server arguments: --defaults_group=mysqld --datadir=/datadir 2026-03-30T10:15:14.192536-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized client arguments: --move-back=1 --force-non-empty-directories=1 --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_g1Ta xtrabackup version 8.0.35-34 based on MySQL server 8.0.35 Linux (x86_64) (revision id: c8a25ff9) 2026-03-30T10:15:14.192567-00:00 0 [Note] [MY-011825] [Xtrabackup] cd to /datadir/pxc_sst_g1Ta/ 2026-03-30T10:15:14.193106-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving undo_001 to /datadir/undo_001 2026-03-30T10:15:14.193146-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file undo_001 to /datadir/undo_001 2026-03-30T10:15:14.193179-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving undo_002 to /datadir/undo_002 2026-03-30T10:15:14.193200-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file undo_002 to /datadir/undo_002 2026-03-30T10:15:14.193340-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving ibdata1 to /datadir/ibdata1 2026-03-30T10:15:14.193365-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file ibdata1 to /datadir/ibdata1 2026-03-30T10:15:14.193591-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving binlog.000009 to /datadir//binlog.000009 2026-03-30T10:15:14.193618-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file binlog.000009 to /datadir//binlog.000009 2026-03-30T10:15:14.193690-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving binlog.index to /datadir//binlog.index 2026-03-30T10:15:14.193716-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file binlog.index to /datadir//binlog.index 2026-03-30T10:15:14.194137-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./myApp/myApp.ibd to /datadir/myApp/myApp.ibd 2026-03-30T10:15:14.194186-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./myApp/myApp.ibd to /datadir/myApp/myApp.ibd 2026-03-30T10:15:14.194200-00:00 1 [Note] [MY-011825] [Xtrabackup] Creating directory ./#innodb_redo 2026-03-30T10:15:14.194214-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: creating directory ./#innodb_redo 2026-03-30T10:15:14.194241-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql.ibd to /datadir/mysql.ibd 2026-03-30T10:15:14.194264-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql.ibd to /datadir/mysql.ibd 2026-03-30T10:15:14.194290-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./xtrabackup_galera_info to /datadir/xtrabackup_galera_info 2026-03-30T10:15:14.194309-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./xtrabackup_galera_info to /datadir/xtrabackup_galera_info 2026-03-30T10:15:14.194327-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./ib_buffer_pool to /datadir/ib_buffer_pool 2026-03-30T10:15:14.194342-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./ib_buffer_pool to /datadir/ib_buffer_pool 2026-03-30T10:15:14.194359-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./ibtmp1 to /datadir/ibtmp1 2026-03-30T10:15:14.194374-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./ibtmp1 to /datadir/ibtmp1 2026-03-30T10:15:14.194429-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_cluster.ibd to /datadir/mysql/wsrep_cluster.ibd 2026-03-30T10:15:14.194448-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_cluster.ibd to /datadir/mysql/wsrep_cluster.ibd 2026-03-30T10:15:14.194466-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log_225.sdi to /datadir/mysql/general_log_225.sdi 2026-03-30T10:15:14.194481-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log_225.sdi to /datadir/mysql/general_log_225.sdi 2026-03-30T10:15:14.194499-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_streaming_log.ibd to /datadir/mysql/wsrep_streaming_log.ibd 2026-03-30T10:15:14.194513-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_streaming_log.ibd to /datadir/mysql/wsrep_streaming_log.ibd 2026-03-30T10:15:14.194530-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log.CSV to /datadir/mysql/general_log.CSV 2026-03-30T10:15:14.194544-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log.CSV to /datadir/mysql/general_log.CSV 2026-03-30T10:15:14.194562-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log.CSM to /datadir/mysql/slow_log.CSM 2026-03-30T10:15:14.194576-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log.CSM to /datadir/mysql/slow_log.CSM 2026-03-30T10:15:14.194592-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log_226.sdi to /datadir/mysql/slow_log_226.sdi 2026-03-30T10:15:14.194605-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log_226.sdi to /datadir/mysql/slow_log_226.sdi 2026-03-30T10:15:14.194620-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log.CSV to /datadir/mysql/slow_log.CSV 2026-03-30T10:15:14.194634-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log.CSV to /datadir/mysql/slow_log.CSV 2026-03-30T10:15:14.194650-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_cluster_members.ibd to /datadir/mysql/wsrep_cluster_members.ibd 2026-03-30T10:15:14.194663-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_cluster_members.ibd to /datadir/mysql/wsrep_cluster_members.ibd 2026-03-30T10:15:14.194679-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log.CSM to /datadir/mysql/general_log.CSM 2026-03-30T10:15:14.194692-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log.CSM to /datadir/mysql/general_log.CSM 2026-03-30T10:15:14.194708-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./sst_info to /datadir/sst_info 2026-03-30T10:15:14.194763-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./sst_info to /datadir/sst_info 2026-03-30T10:15:14.194842-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_host_187.sdi to /datadir/performance_schema/status_by_host_187.sdi 2026-03-30T10:15:14.194916-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_host_187.sdi to /datadir/performance_schema/status_by_host_187.sdi 2026-03-30T10:15:14.194951-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_101.sdi to /datadir/performance_schema/events_waits_sum_101.sdi 2026-03-30T10:15:14.194975-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_101.sdi to /datadir/performance_schema/events_waits_sum_101.sdi 2026-03-30T10:15:14.195037-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_125.sdi to /datadir/performance_schema/events_stages_su_125.sdi 2026-03-30T10:15:14.195060-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_125.sdi to /datadir/performance_schema/events_stages_su_125.sdi 2026-03-30T10:15:14.195089-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/metadata_locks_168.sdi to /datadir/performance_schema/metadata_locks_168.sdi 2026-03-30T10:15:14.195123-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/metadata_locks_168.sdi to /datadir/performance_schema/metadata_locks_168.sdi 2026-03-30T10:15:14.195155-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/accounts_154.sdi to /datadir/performance_schema/accounts_154.sdi 2026-03-30T10:15:14.195184-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/accounts_154.sdi to /datadir/performance_schema/accounts_154.sdi 2026-03-30T10:15:14.195235-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_136.sdi to /datadir/performance_schema/events_statement_136.sdi 2026-03-30T10:15:14.195270-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_136.sdi to /datadir/performance_schema/events_statement_136.sdi 2026-03-30T10:15:14.195302-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_grou_178.sdi to /datadir/performance_schema/replication_grou_178.sdi 2026-03-30T10:15:14.195326-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_grou_178.sdi to /datadir/performance_schema/replication_grou_178.sdi 2026-03-30T10:15:14.195385-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/rwlock_instances_110.sdi to /datadir/performance_schema/rwlock_instances_110.sdi 2026-03-30T10:15:14.195438-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/rwlock_instances_110.sdi to /datadir/performance_schema/rwlock_instances_110.sdi 2026-03-30T10:15:14.195463-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_137.sdi to /datadir/performance_schema/events_statement_137.sdi 2026-03-30T10:15:14.195487-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_137.sdi to /datadir/performance_schema/events_statement_137.sdi 2026-03-30T10:15:14.195521-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_151.sdi to /datadir/performance_schema/events_errors_su_151.sdi 2026-03-30T10:15:14.195552-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_151.sdi to /datadir/performance_schema/events_errors_su_151.sdi 2026-03-30T10:15:14.195583-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_145.sdi to /datadir/performance_schema/events_transacti_145.sdi 2026-03-30T10:15:14.195613-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_145.sdi to /datadir/performance_schema/events_transacti_145.sdi 2026-03-30T10:15:14.195646-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_cur_93.sdi to /datadir/performance_schema/events_waits_cur_93.sdi 2026-03-30T10:15:14.195678-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_cur_93.sdi to /datadir/performance_schema/events_waits_cur_93.sdi 2026-03-30T10:15:14.195701-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/variables_by_thr_192.sdi to /datadir/performance_schema/variables_by_thr_192.sdi 2026-03-30T10:15:14.195746-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/variables_by_thr_192.sdi to /datadir/performance_schema/variables_by_thr_192.sdi 2026-03-30T10:15:14.195779-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_97.sdi to /datadir/performance_schema/events_waits_sum_97.sdi 2026-03-30T10:15:14.195804-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_97.sdi to /datadir/performance_schema/events_waits_sum_97.sdi 2026-03-30T10:15:14.195827-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/prepared_stateme_184.sdi to /datadir/performance_schema/prepared_stateme_184.sdi 2026-03-30T10:15:14.195851-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/prepared_stateme_184.sdi to /datadir/performance_schema/prepared_stateme_184.sdi 2026-03-30T10:15:14.195912-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_133.sdi to /datadir/performance_schema/events_statement_133.sdi 2026-03-30T10:15:14.195937-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_133.sdi to /datadir/performance_schema/events_statement_133.sdi 2026-03-30T10:15:14.195960-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_139.sdi to /datadir/performance_schema/events_statement_139.sdi 2026-03-30T10:15:14.195989-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_139.sdi to /datadir/performance_schema/events_statement_139.sdi 2026-03-30T10:15:14.196016-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_123.sdi to /datadir/performance_schema/events_stages_su_123.sdi 2026-03-30T10:15:14.196035-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_123.sdi to /datadir/performance_schema/events_stages_su_123.sdi 2026-03-30T10:15:14.196057-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_174.sdi to /datadir/performance_schema/replication_appl_174.sdi 2026-03-30T10:15:14.196118-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_174.sdi to /datadir/performance_schema/replication_appl_174.sdi 2026-03-30T10:15:14.196149-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_variable_194.sdi to /datadir/performance_schema/session_variable_194.sdi 2026-03-30T10:15:14.196178-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_variable_194.sdi to /datadir/performance_schema/session_variable_194.sdi 2026-03-30T10:15:14.196207-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_127.sdi to /datadir/performance_schema/events_stages_su_127.sdi 2026-03-30T10:15:14.196232-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_127.sdi to /datadir/performance_schema/events_stages_su_127.sdi 2026-03-30T10:15:14.196262-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_146.sdi to /datadir/performance_schema/events_transacti_146.sdi 2026-03-30T10:15:14.196290-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_146.sdi to /datadir/performance_schema/events_transacti_146.sdi 2026-03-30T10:15:14.196314-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_166.sdi to /datadir/performance_schema/memory_summary_b_166.sdi 2026-03-30T10:15:14.196336-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_166.sdi to /datadir/performance_schema/memory_summary_b_166.sdi 2026-03-30T10:15:14.196362-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_180.sdi to /datadir/performance_schema/replication_appl_180.sdi 2026-03-30T10:15:14.196415-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_180.sdi to /datadir/performance_schema/replication_appl_180.sdi 2026-03-30T10:15:14.196442-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_147.sdi to /datadir/performance_schema/events_transacti_147.sdi 2026-03-30T10:15:14.196476-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_147.sdi to /datadir/performance_schema/events_transacti_147.sdi 2026-03-30T10:15:14.196507-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_io_waits_s_116.sdi to /datadir/performance_schema/table_io_waits_s_116.sdi 2026-03-30T10:15:14.196538-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_io_waits_s_116.sdi to /datadir/performance_schema/table_io_waits_s_116.sdi 2026-03-30T10:15:14.196569-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_asyn_181.sdi to /datadir/performance_schema/replication_asyn_181.sdi 2026-03-30T10:15:14.196599-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_asyn_181.sdi to /datadir/performance_schema/replication_asyn_181.sdi 2026-03-30T10:15:14.196632-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/log_status_183.sdi to /datadir/performance_schema/log_status_183.sdi 2026-03-30T10:15:14.196664-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/log_status_183.sdi to /datadir/performance_schema/log_status_183.sdi 2026-03-30T10:15:14.196691-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/global_variables_193.sdi to /datadir/performance_schema/global_variables_193.sdi 2026-03-30T10:15:14.196757-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/global_variables_193.sdi to /datadir/performance_schema/global_variables_193.sdi 2026-03-30T10:15:14.196790-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_summary_by__104.sdi to /datadir/performance_schema/file_summary_by__104.sdi 2026-03-30T10:15:14.196816-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_summary_by__104.sdi to /datadir/performance_schema/file_summary_by__104.sdi 2026-03-30T10:15:14.196851-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_conn_173.sdi to /datadir/performance_schema/replication_conn_173.sdi 2026-03-30T10:15:14.196913-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_conn_173.sdi to /datadir/performance_schema/replication_conn_173.sdi 2026-03-30T10:15:14.196939-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_consumers_112.sdi to /datadir/performance_schema/setup_consumers_112.sdi 2026-03-30T10:15:14.196963-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_consumers_112.sdi to /datadir/performance_schema/setup_consumers_112.sdi 2026-03-30T10:15:14.196990-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_131.sdi to /datadir/performance_schema/events_statement_131.sdi 2026-03-30T10:15:14.197049-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_131.sdi to /datadir/performance_schema/events_statement_131.sdi 2026-03-30T10:15:14.197082-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_summary_b_158.sdi to /datadir/performance_schema/socket_summary_b_158.sdi 2026-03-30T10:15:14.197141-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_summary_b_158.sdi to /datadir/performance_schema/socket_summary_b_158.sdi 2026-03-30T10:15:14.197170-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_128.sdi to /datadir/performance_schema/events_statement_128.sdi 2026-03-30T10:15:14.197222-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_128.sdi to /datadir/performance_schema/events_statement_128.sdi 2026-03-30T10:15:14.197249-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_175.sdi to /datadir/performance_schema/replication_appl_175.sdi 2026-03-30T10:15:14.197280-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_175.sdi to /datadir/performance_schema/replication_appl_175.sdi 2026-03-30T10:15:14.197312-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_grou_172.sdi to /datadir/performance_schema/replication_grou_172.sdi 2026-03-30T10:15:14.197340-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_grou_172.sdi to /datadir/performance_schema/replication_grou_172.sdi 2026-03-30T10:15:14.197367-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_124.sdi to /datadir/performance_schema/events_stages_su_124.sdi 2026-03-30T10:15:14.197393-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_124.sdi to /datadir/performance_schema/events_stages_su_124.sdi 2026-03-30T10:15:14.197425-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/keyring_keys_161.sdi to /datadir/performance_schema/keyring_keys_161.sdi 2026-03-30T10:15:14.197452-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/keyring_keys_161.sdi to /datadir/performance_schema/keyring_keys_161.sdi 2026-03-30T10:15:14.197484-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/tls_channel_stat_199.sdi to /datadir/performance_schema/tls_channel_stat_199.sdi 2026-03-30T10:15:14.197515-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/tls_channel_stat_199.sdi to /datadir/performance_schema/tls_channel_stat_199.sdi 2026-03-30T10:15:14.197549-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_user_189.sdi to /datadir/performance_schema/status_by_user_189.sdi 2026-03-30T10:15:14.197577-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_user_189.sdi to /datadir/performance_schema/status_by_user_189.sdi 2026-03-30T10:15:14.197632-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_g_162.sdi to /datadir/performance_schema/memory_summary_g_162.sdi 2026-03-30T10:15:14.197657-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_g_162.sdi to /datadir/performance_schema/memory_summary_g_162.sdi 2026-03-30T10:15:14.197685-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_instrument_113.sdi to /datadir/performance_schema/setup_instrument_113.sdi 2026-03-30T10:15:14.197711-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_instrument_113.sdi to /datadir/performance_schema/setup_instrument_113.sdi 2026-03-30T10:15:14.197752-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_149.sdi to /datadir/performance_schema/events_errors_su_149.sdi 2026-03-30T10:15:14.197781-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_149.sdi to /datadir/performance_schema/events_errors_su_149.sdi 2026-03-30T10:15:14.197814-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_98.sdi to /datadir/performance_schema/events_waits_sum_98.sdi 2026-03-30T10:15:14.197845-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_98.sdi to /datadir/performance_schema/events_waits_sum_98.sdi 2026-03-30T10:15:14.197877-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/keyring_componen_202.sdi to /datadir/performance_schema/keyring_componen_202.sdi 2026-03-30T10:15:14.197908-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/keyring_componen_202.sdi to /datadir/performance_schema/keyring_componen_202.sdi 2026-03-30T10:15:14.197940-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_140.sdi to /datadir/performance_schema/events_transacti_140.sdi 2026-03-30T10:15:14.197971-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_140.sdi to /datadir/performance_schema/events_transacti_140.sdi 2026-03-30T10:15:14.198004-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_account__160.sdi to /datadir/performance_schema/session_account__160.sdi 2026-03-30T10:15:14.198064-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_account__160.sdi to /datadir/performance_schema/session_account__160.sdi 2026-03-30T10:15:14.198094-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_100.sdi to /datadir/performance_schema/events_waits_sum_100.sdi 2026-03-30T10:15:14.198118-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_100.sdi to /datadir/performance_schema/events_waits_sum_100.sdi 2026-03-30T10:15:14.198152-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_summary_by__103.sdi to /datadir/performance_schema/file_summary_by__103.sdi 2026-03-30T10:15:14.198180-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_summary_by__103.sdi to /datadir/performance_schema/file_summary_by__103.sdi 2026-03-30T10:15:14.198209-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_his_94.sdi to /datadir/performance_schema/events_waits_his_94.sdi 2026-03-30T10:15:14.198233-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_his_94.sdi to /datadir/performance_schema/events_waits_his_94.sdi 2026-03-30T10:15:14.198260-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_hi_121.sdi to /datadir/performance_schema/events_stages_hi_121.sdi 2026-03-30T10:15:14.198294-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_hi_121.sdi to /datadir/performance_schema/events_stages_hi_121.sdi 2026-03-30T10:15:14.198328-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_126.sdi to /datadir/performance_schema/events_stages_su_126.sdi 2026-03-30T10:15:14.198382-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_126.sdi to /datadir/performance_schema/events_stages_su_126.sdi 2026-03-30T10:15:14.198420-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_165.sdi to /datadir/performance_schema/memory_summary_b_165.sdi 2026-03-30T10:15:14.198449-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_165.sdi to /datadir/performance_schema/memory_summary_b_165.sdi 2026-03-30T10:15:14.198475-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_142.sdi to /datadir/performance_schema/events_transacti_142.sdi 2026-03-30T10:15:14.198503-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_142.sdi to /datadir/performance_schema/events_transacti_142.sdi 2026-03-30T10:15:14.198522-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_141.sdi to /datadir/performance_schema/events_transacti_141.sdi 2026-03-30T10:15:14.198538-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_141.sdi to /datadir/performance_schema/events_transacti_141.sdi 2026-03-30T10:15:14.198556-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_152.sdi to /datadir/performance_schema/events_errors_su_152.sdi 2026-03-30T10:15:14.198580-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_152.sdi to /datadir/performance_schema/events_errors_su_152.sdi 2026-03-30T10:15:14.198599-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_177.sdi to /datadir/performance_schema/replication_appl_177.sdi 2026-03-30T10:15:14.198616-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_177.sdi to /datadir/performance_schema/replication_appl_177.sdi 2026-03-30T10:15:14.198632-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/cond_instances_91.sdi to /datadir/performance_schema/cond_instances_91.sdi 2026-03-30T10:15:14.198647-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/cond_instances_91.sdi to /datadir/performance_schema/cond_instances_91.sdi 2026-03-30T10:15:14.198662-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/variables_info_195.sdi to /datadir/performance_schema/variables_info_195.sdi 2026-03-30T10:15:14.198677-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/variables_info_195.sdi to /datadir/performance_schema/variables_info_195.sdi 2026-03-30T10:15:14.198693-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_176.sdi to /datadir/performance_schema/replication_appl_176.sdi 2026-03-30T10:15:14.198708-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_176.sdi to /datadir/performance_schema/replication_appl_176.sdi 2026-03-30T10:15:14.198744-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_148.sdi to /datadir/performance_schema/events_errors_su_148.sdi 2026-03-30T10:15:14.198770-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_148.sdi to /datadir/performance_schema/events_errors_su_148.sdi 2026-03-30T10:15:14.198804-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/processlist_109.sdi to /datadir/performance_schema/processlist_109.sdi 2026-03-30T10:15:14.198839-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/processlist_109.sdi to /datadir/performance_schema/processlist_109.sdi 2026-03-30T10:15:14.198871-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/global_status_190.sdi to /datadir/performance_schema/global_status_190.sdi 2026-03-30T10:15:14.198900-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/global_status_190.sdi to /datadir/performance_schema/global_status_190.sdi 2026-03-30T10:15:14.198932-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/malloc_stats_201.sdi to /datadir/performance_schema/malloc_stats_201.sdi 2026-03-30T10:15:14.198963-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/malloc_stats_201.sdi to /datadir/performance_schema/malloc_stats_201.sdi 2026-03-30T10:15:14.198996-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_132.sdi to /datadir/performance_schema/events_statement_132.sdi 2026-03-30T10:15:14.199029-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_132.sdi to /datadir/performance_schema/events_statement_132.sdi 2026-03-30T10:15:14.199093-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/host_cache_105.sdi to /datadir/performance_schema/host_cache_105.sdi 2026-03-30T10:15:14.199119-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/host_cache_105.sdi to /datadir/performance_schema/host_cache_105.sdi 2026-03-30T10:15:14.199149-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/hosts_155.sdi to /datadir/performance_schema/hosts_155.sdi 2026-03-30T10:15:14.199174-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/hosts_155.sdi to /datadir/performance_schema/hosts_155.sdi 2026-03-30T10:15:14.199201-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_handles_167.sdi to /datadir/performance_schema/table_handles_167.sdi 2026-03-30T10:15:14.199232-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_handles_167.sdi to /datadir/performance_schema/table_handles_167.sdi 2026-03-30T10:15:14.199262-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_objects_114.sdi to /datadir/performance_schema/setup_objects_114.sdi 2026-03-30T10:15:14.199320-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_objects_114.sdi to /datadir/performance_schema/setup_objects_114.sdi 2026-03-30T10:15:14.199354-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_144.sdi to /datadir/performance_schema/events_transacti_144.sdi 2026-03-30T10:15:14.199383-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_144.sdi to /datadir/performance_schema/events_transacti_144.sdi 2026-03-30T10:15:14.199407-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/data_locks_169.sdi to /datadir/performance_schema/data_locks_169.sdi 2026-03-30T10:15:14.199433-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/data_locks_169.sdi to /datadir/performance_schema/data_locks_169.sdi 2026-03-30T10:15:14.199462-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/threads_119.sdi to /datadir/performance_schema/threads_119.sdi 2026-03-30T10:15:14.199490-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/threads_119.sdi to /datadir/performance_schema/threads_119.sdi 2026-03-30T10:15:14.199522-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_150.sdi to /datadir/performance_schema/events_errors_su_150.sdi 2026-03-30T10:15:14.199550-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_150.sdi to /datadir/performance_schema/events_errors_su_150.sdi 2026-03-30T10:15:14.199580-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/objects_summary__107.sdi to /datadir/performance_schema/objects_summary__107.sdi 2026-03-30T10:15:14.199608-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/objects_summary__107.sdi to /datadir/performance_schema/objects_summary__107.sdi 2026-03-30T10:15:14.199637-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/user_defined_fun_197.sdi to /datadir/performance_schema/user_defined_fun_197.sdi 2026-03-30T10:15:14.199667-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/user_defined_fun_197.sdi to /datadir/performance_schema/user_defined_fun_197.sdi 2026-03-30T10:15:14.199699-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_conn_171.sdi to /datadir/performance_schema/replication_conn_171.sdi 2026-03-30T10:15:14.199741-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_conn_171.sdi to /datadir/performance_schema/replication_conn_171.sdi 2026-03-30T10:15:14.199771-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_cu_120.sdi to /datadir/performance_schema/events_stages_cu_120.sdi 2026-03-30T10:15:14.199804-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_cu_120.sdi to /datadir/performance_schema/events_stages_cu_120.sdi 2026-03-30T10:15:14.199839-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_134.sdi to /datadir/performance_schema/events_statement_134.sdi 2026-03-30T10:15:14.199863-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_134.sdi to /datadir/performance_schema/events_statement_134.sdi 2026-03-30T10:15:14.199895-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_135.sdi to /datadir/performance_schema/events_statement_135.sdi 2026-03-30T10:15:14.199960-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_135.sdi to /datadir/performance_schema/events_statement_135.sdi 2026-03-30T10:15:14.199990-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_138.sdi to /datadir/performance_schema/events_statement_138.sdi 2026-03-30T10:15:14.200022-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_138.sdi to /datadir/performance_schema/events_statement_138.sdi 2026-03-30T10:15:14.200089-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_lock_waits_118.sdi to /datadir/performance_schema/table_lock_waits_118.sdi 2026-03-30T10:15:14.200116-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_lock_waits_118.sdi to /datadir/performance_schema/table_lock_waits_118.sdi 2026-03-30T10:15:14.200141-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/data_lock_waits_170.sdi to /datadir/performance_schema/data_lock_waits_170.sdi 2026-03-30T10:15:14.200170-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/data_lock_waits_170.sdi to /datadir/performance_schema/data_lock_waits_170.sdi 2026-03-30T10:15:14.200201-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_hi_122.sdi to /datadir/performance_schema/events_stages_hi_122.sdi 2026-03-30T10:15:14.200263-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_hi_122.sdi to /datadir/performance_schema/events_stages_hi_122.sdi 2026-03-30T10:15:14.200295-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/performance_time_108.sdi to /datadir/performance_schema/performance_time_108.sdi 2026-03-30T10:15:14.200320-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/performance_time_108.sdi to /datadir/performance_schema/performance_time_108.sdi 2026-03-30T10:15:14.200350-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_164.sdi to /datadir/performance_schema/memory_summary_b_164.sdi 2026-03-30T10:15:14.200379-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_164.sdi to /datadir/performance_schema/memory_summary_b_164.sdi 2026-03-30T10:15:14.200408-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_summary_b_157.sdi to /datadir/performance_schema/socket_summary_b_157.sdi 2026-03-30T10:15:14.200457-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_summary_b_157.sdi to /datadir/performance_schema/socket_summary_b_157.sdi 2026-03-30T10:15:14.200491-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/malloc_stats_tot_200.sdi to /datadir/performance_schema/malloc_stats_tot_200.sdi 2026-03-30T10:15:14.200526-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/malloc_stats_tot_200.sdi to /datadir/performance_schema/malloc_stats_tot_200.sdi 2026-03-30T10:15:14.200546-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_io_waits_s_117.sdi to /datadir/performance_schema/table_io_waits_s_117.sdi 2026-03-30T10:15:14.200563-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_io_waits_s_117.sdi to /datadir/performance_schema/table_io_waits_s_117.sdi 2026-03-30T10:15:14.200579-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_his_95.sdi to /datadir/performance_schema/events_waits_his_95.sdi 2026-03-30T10:15:14.200595-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_his_95.sdi to /datadir/performance_schema/events_waits_his_95.sdi 2026-03-30T10:15:14.200610-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_threads_115.sdi to /datadir/performance_schema/setup_threads_115.sdi 2026-03-30T10:15:14.200626-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_threads_115.sdi to /datadir/performance_schema/setup_threads_115.sdi 2026-03-30T10:15:14.200641-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/users_153.sdi to /datadir/performance_schema/users_153.sdi 2026-03-30T10:15:14.200656-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/users_153.sdi to /datadir/performance_schema/users_153.sdi 2026-03-30T10:15:14.200672-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_163.sdi to /datadir/performance_schema/memory_summary_b_163.sdi 2026-03-30T10:15:14.200689-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_163.sdi to /datadir/performance_schema/memory_summary_b_163.sdi 2026-03-30T10:15:14.200705-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/binary_log_trans_198.sdi to /datadir/performance_schema/binary_log_trans_198.sdi 2026-03-30T10:15:14.200721-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/binary_log_trans_198.sdi to /datadir/performance_schema/binary_log_trans_198.sdi 2026-03-30T10:15:14.200766-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_instances_156.sdi to /datadir/performance_schema/socket_instances_156.sdi 2026-03-30T10:15:14.200791-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_instances_156.sdi to /datadir/performance_schema/socket_instances_156.sdi 2026-03-30T10:15:14.200822-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_thread_188.sdi to /datadir/performance_schema/status_by_thread_188.sdi 2026-03-30T10:15:14.200851-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_thread_188.sdi to /datadir/performance_schema/status_by_thread_188.sdi 2026-03-30T10:15:14.200880-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_accoun_186.sdi to /datadir/performance_schema/status_by_accoun_186.sdi 2026-03-30T10:15:14.200909-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_accoun_186.sdi to /datadir/performance_schema/status_by_accoun_186.sdi 2026-03-30T10:15:14.200957-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_96.sdi to /datadir/performance_schema/events_waits_sum_96.sdi 2026-03-30T10:15:14.200986-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_96.sdi to /datadir/performance_schema/events_waits_sum_96.sdi 2026-03-30T10:15:14.201016-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_99.sdi to /datadir/performance_schema/events_waits_sum_99.sdi 2026-03-30T10:15:14.201042-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_99.sdi to /datadir/performance_schema/events_waits_sum_99.sdi 2026-03-30T10:15:14.201081-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/mutex_instances_106.sdi to /datadir/performance_schema/mutex_instances_106.sdi 2026-03-30T10:15:14.201108-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/mutex_instances_106.sdi to /datadir/performance_schema/mutex_instances_106.sdi 2026-03-30T10:15:14.201139-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/user_variables_b_185.sdi to /datadir/performance_schema/user_variables_b_185.sdi 2026-03-30T10:15:14.201170-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/user_variables_b_185.sdi to /datadir/performance_schema/user_variables_b_185.sdi 2026-03-30T10:15:14.201200-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_connect__159.sdi to /datadir/performance_schema/session_connect__159.sdi 2026-03-30T10:15:14.201227-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_connect__159.sdi to /datadir/performance_schema/session_connect__159.sdi 2026-03-30T10:15:14.201256-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/error_log_92.sdi to /datadir/performance_schema/error_log_92.sdi 2026-03-30T10:15:14.201285-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/error_log_92.sdi to /datadir/performance_schema/error_log_92.sdi 2026-03-30T10:15:14.201313-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_179.sdi to /datadir/performance_schema/replication_appl_179.sdi 2026-03-30T10:15:14.201336-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_179.sdi to /datadir/performance_schema/replication_appl_179.sdi 2026-03-30T10:15:14.201400-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_status_191.sdi to /datadir/performance_schema/session_status_191.sdi 2026-03-30T10:15:14.201429-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_status_191.sdi to /datadir/performance_schema/session_status_191.sdi 2026-03-30T10:15:14.201461-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_143.sdi to /datadir/performance_schema/events_transacti_143.sdi 2026-03-30T10:15:14.201485-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_143.sdi to /datadir/performance_schema/events_transacti_143.sdi 2026-03-30T10:15:14.201512-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_actors_111.sdi to /datadir/performance_schema/setup_actors_111.sdi 2026-03-30T10:15:14.201542-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_actors_111.sdi to /datadir/performance_schema/setup_actors_111.sdi 2026-03-30T10:15:14.201576-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/persisted_variab_196.sdi to /datadir/performance_schema/persisted_variab_196.sdi 2026-03-30T10:15:14.201609-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/persisted_variab_196.sdi to /datadir/performance_schema/persisted_variab_196.sdi 2026-03-30T10:15:14.201636-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_instances_102.sdi to /datadir/performance_schema/file_instances_102.sdi 2026-03-30T10:15:14.201665-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_instances_102.sdi to /datadir/performance_schema/file_instances_102.sdi 2026-03-30T10:15:14.201693-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_asyn_182.sdi to /datadir/performance_schema/replication_asyn_182.sdi 2026-03-30T10:15:14.201749-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_asyn_182.sdi to /datadir/performance_schema/replication_asyn_182.sdi 2026-03-30T10:15:14.201804-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_129.sdi to /datadir/performance_schema/events_statement_129.sdi 2026-03-30T10:15:14.201836-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_129.sdi to /datadir/performance_schema/events_statement_129.sdi 2026-03-30T10:15:14.201869-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/pxc_cluster_view_203.sdi to /datadir/performance_schema/pxc_cluster_view_203.sdi 2026-03-30T10:15:14.201902-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/pxc_cluster_view_203.sdi to /datadir/performance_schema/pxc_cluster_view_203.sdi 2026-03-30T10:15:14.201930-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_130.sdi to /datadir/performance_schema/events_statement_130.sdi 2026-03-30T10:15:14.201961-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_130.sdi to /datadir/performance_schema/events_statement_130.sdi 2026-03-30T10:15:14.202015-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./xtrabackup_info to /datadir/xtrabackup_info 2026-03-30T10:15:14.202039-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./xtrabackup_info to /datadir/xtrabackup_info 2026-03-30T10:15:14.202119-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./sys/sys_config.ibd to /datadir/sys/sys_config.ibd 2026-03-30T10:15:14.202144-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./sys/sys_config.ibd to /datadir/sys/sys_config.ibd 2026-03-30T10:15:14.294771-00:00 0 [Note] [MY-011825] [Xtrabackup] completed OK! + cat /tmp/tmp.r1Owkol9aM Defaulted container "xtrabackup" out of: xtrabackup, backup-init (init) + rm /tmp/tmp.sJIGcUzX4i /tmp/tmp.r1Owkol9aM + return 0 + wait_for_running demand-backup-proxysql 1 + local name=demand-backup-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod demand-backup-proxysql-0 480 + local pod=demand-backup-proxysql-0 + local max_retry=480 + local ns= ++ echo demand-backup-proxysql-0 ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=proxysql + set +o xtrace pod/demand-backup-proxysql-0 condition met waiting for pod/demand-backup-proxysql-0 to become Ready.Ok + wait_for_running demand-backup-pxc 3 + local name=demand-backup-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod demand-backup-pxc-0 480 + local pod=demand-backup-pxc-0 + local max_retry=480 + local ns= ++ echo demand-backup-pxc-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/demand-backup-pxc-0 condition met waiting for pod/demand-backup-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod demand-backup-pxc-1 480 + local pod=demand-backup-pxc-1 + local max_retry=480 + local ns= ++ echo demand-backup-pxc-1 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/demand-backup-pxc-1 condition met waiting for pod/demand-backup-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod demand-backup-pxc-2 480 + local pod=demand-backup-pxc-2 + local max_retry=480 + local ns= ++ echo demand-backup-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/demand-backup-pxc-2 condition met waiting for pod/demand-backup-pxc-2 to become Ready.Ok ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.spec.secretsName}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IWjAgakthi +++ mktemp ++ local LAST_ERR=/tmp/tmp.GOS1mwqlql ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.spec.secretsName}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.IWjAgakthi ++ cat /tmp/tmp.GOS1mwqlql ++ rm /tmp/tmp.IWjAgakthi /tmp/tmp.GOS1mwqlql ++ return 0 + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.45eyNxOQoc +++ mktemp ++ local LAST_ERR=/tmp/tmp.8gVj8h5bTv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.45eyNxOQoc ++ cat /tmp/tmp.8gVj8h5bTv ++ rm /tmp/tmp.45eyNxOQoc /tmp/tmp.8gVj8h5bTv ++ return 0 + local root_pass=root_password + sleep 35 + log 'check data after pxc-restore/on-demand-backup-pvc' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-03-30T10:21:02+0000]' check data after pxc-restore/on-demand-backup-pvc [2026-03-30T10:21:02+0000] check data after pxc-restore/on-demand-backup-pvc + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-0.demand-backup-pxc -uroot -p'\''root_password'\''' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-0.demand-backup-pxc -uroot -p'\''root_password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-0.demand-backup-pxc -uroot -p'\''root_password'\''' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-0.demand-backup-pxc -uroot -p'\''root_password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.olyszfCiOw +++ mktemp ++ local LAST_ERR=/tmp/tmp.KkohToaehb ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.olyszfCiOw ++ cat /tmp/tmp.KkohToaehb ++ rm /tmp/tmp.olyszfCiOw /tmp/tmp.KkohToaehb ++ return 0 + client_pod=pxc-client-67fc4995bb-jwx2s + wait_pod pxc-client-67fc4995bb-jwx2s + local pod=pxc-client-67fc4995bb-jwx2s + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-67fc4995bb-jwx2s + local container= + set +o xtrace pod/pxc-client-67fc4995bb-jwx2s condition met waiting for pod/pxc-client-67fc4995bb-jwx2s to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.14SIjS2s6f/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.14SIjS2s6f/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1.sql /tmp/tmp.14SIjS2s6f/select-1.sql + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-1.demand-backup-pxc -uroot -p'\''root_password'\''' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-1.demand-backup-pxc -uroot -p'\''root_password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-1.demand-backup-pxc -uroot -p'\''root_password'\''' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-1.demand-backup-pxc -uroot -p'\''root_password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HFnJTG5jeB +++ mktemp ++ local LAST_ERR=/tmp/tmp.PtplXfngJK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.HFnJTG5jeB ++ cat /tmp/tmp.PtplXfngJK ++ rm /tmp/tmp.HFnJTG5jeB /tmp/tmp.PtplXfngJK ++ return 0 + client_pod=pxc-client-67fc4995bb-jwx2s + wait_pod pxc-client-67fc4995bb-jwx2s + local pod=pxc-client-67fc4995bb-jwx2s + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-jwx2s ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-jwx2s condition met waiting for pod/pxc-client-67fc4995bb-jwx2s to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.14SIjS2s6f/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.14SIjS2s6f/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1.sql /tmp/tmp.14SIjS2s6f/select-1.sql + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-2.demand-backup-pxc -uroot -p'\''root_password'\''' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-2.demand-backup-pxc -uroot -p'\''root_password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-2.demand-backup-pxc -uroot -p'\''root_password'\''' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-2.demand-backup-pxc -uroot -p'\''root_password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4TS1jgQNA7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.q07SBcpJ72 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.4TS1jgQNA7 ++ cat /tmp/tmp.q07SBcpJ72 ++ rm /tmp/tmp.4TS1jgQNA7 /tmp/tmp.q07SBcpJ72 ++ return 0 + client_pod=pxc-client-67fc4995bb-jwx2s + wait_pod pxc-client-67fc4995bb-jwx2s + local pod=pxc-client-67fc4995bb-jwx2s + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-67fc4995bb-jwx2s ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-jwx2s condition met waiting for pod/pxc-client-67fc4995bb-jwx2s to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.14SIjS2s6f/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.14SIjS2s6f/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1.sql /tmp/tmp.14SIjS2s6f/select-1.sql + '[' on-demand-backup-pvc '!=' on-demand-backup-minio ']' + log 'copy backup' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-03-30T10:21:22+0000]' copy backup [2026-03-30T10:21:22+0000] copy backup + '[' -n '' ']' + bash /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/deploy/backup/copy-backup.sh on-demand-backup-pvc /tmp/tmp.14SIjS2s6f/backup Log: /tmp/tmp.1S750yJ2yH/log pvc/xb-on-demand-backup-pvc-20260330101253-2d0422b9pod/backup-access created Starting pod.....[done] Downloading started tar: Removing leading `/' from member names Downloading finished pod "backup-access" deleted from demand-backup-28441 namespace You can recover data locally with following commands: $ service mysqld stop $ rm -rf /var/lib/mysql/* $ cat /tmp/tmp.14SIjS2s6f/backup/xtrabackup.stream | xbstream --decompress -x -C /var/lib/mysql $ xtrabackup --prepare --target-dir=/var/lib/mysql $ chown -R mysql:mysql /var/lib/mysql $ service mysqld start + compare_kubectl job/restore-job-on-demand-backup-pvc-demand-backup + local resource=job/restore-job-on-demand-backup-pvc-demand-backup + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_restore-job-on-demand-backup-pvc-demand-backup.yml + local new_result=/tmp/tmp.14SIjS2s6f/job_restore-job-on-demand-backup-pvc-demand-backup.yml + desc 'compare job/restore-job-on-demand-backup-pvc-demand-backup-' + set +o xtrace ----------------------------------------------------------------------------------- compare job/restore-job-on-demand-backup-pvc-demand-backup- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_restore-job-on-demand-backup-pvc-demand-backup-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_restore-job-on-demand-backup-pvc-demand-backup-80.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.32 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.32 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_restore-job-on-demand-backup-pvc-demand-backup-k129.yml ']' + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_restore-job-on-demand-backup-pvc-demand-backup-k129.yml + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_restore-job-on-demand-backup-pvc-demand-backup-k129-oc.yml ']' + version_gt 1.29 ++ echo '1.32 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_restore-job-on-demand-backup-pvc-demand-backup-k129-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_restore-job-on-demand-backup-pvc-demand-backup-k129-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_restore-job-on-demand-backup-pvc-demand-backup-k129-aks.yml ']' + kubectl_bin get -o yaml job/restore-job-on-demand-backup-pvc-demand-backup ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.spec.template.spec.containers[].env[] | select(.name == "XTRABACKUP_ENABLED")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.metadata.annotations."kubernetes.digitalocean.com/load-balancer-id") | del(.metadata.annotations."service.beta.kubernetes.io/do-loadbalancer-type") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("demand-backup-28441", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.KCKcI6mwQb ++ mktemp + local LAST_ERR=/tmp/tmp.26sfB4bm1Y + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml job/restore-job-on-demand-backup-pvc-demand-backup + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.KCKcI6mwQb + cat /tmp/tmp.26sfB4bm1Y + rm /tmp/tmp.KCKcI6mwQb /tmp/tmp.26sfB4bm1Y + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_restore-job-on-demand-backup-pvc-demand-backup-k129.yml /tmp/tmp.14SIjS2s6f/job_restore-job-on-demand-backup-pvc-demand-backup.yml + log 'compare_kubectl: job/restore-job-on-demand-backup-pvc-demand-backup OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-03-30T10:22:14+0000]' compare_kubectl: job/restore-job-on-demand-backup-pvc-demand-backup OK [2026-03-30T10:22:14+0000] compare_kubectl: job/restore-job-on-demand-backup-pvc-demand-backup OK + check_pvc_md5 on-demand-backup-pvc + local backup_name=on-demand-backup-pvc + desc 'check backup file md5sum' + set +o xtrace ----------------------------------------------------------------------------------- check backup file md5sum ----------------------------------------------------------------------------------- ++ get_pvc_name_for_backup on-demand-backup-pvc ++ local backup_name=on-demand-backup-pvc +++ kubectl_bin get pxc-backup on-demand-backup-pvc -o 'jsonpath={.status.destination}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.eJ7AyriVhP ++++ mktemp +++ local LAST_ERR=/tmp/tmp.9UAnXkwXeT +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc-backup on-demand-backup-pvc -o 'jsonpath={.status.destination}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.eJ7AyriVhP +++ cat /tmp/tmp.9UAnXkwXeT +++ rm /tmp/tmp.eJ7AyriVhP /tmp/tmp.9UAnXkwXeT +++ return 0 ++ local destination=pvc/xb-on-demand-backup-pvc-20260330101253-2d0422b9 ++ [[ -z pvc/xb-on-demand-backup-pvc-20260330101253-2d0422b9 ]] ++ local pvc_name=xb-on-demand-backup-pvc-20260330101253-2d0422b9 ++ echo xb-on-demand-backup-pvc-20260330101253-2d0422b9 + pvc_name=xb-on-demand-backup-pvc-20260330101253-2d0422b9 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/client.yml xb-on-demand-backup-pvc-20260330101253-2d0422b9 + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/client.yml + local pvc_name=xb-on-demand-backup-pvc-20260330101253-2d0422b9 + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/client.yml xb-on-demand-backup-pvc-20260330101253-2d0422b9 + kubectl_bin apply -f - + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/client.yml + local pvc_name=xb-on-demand-backup-pvc-20260330101253-2d0422b9 + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/client.yml ++ mktemp + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + local LAST_OUT=/tmp/tmp.IjF0xSvAIW + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2382-13dc5c0a#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.demand-backup-28441~ ++ mktemp + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: xb-on-demand-backup-pvc-20260330101253-2d0422b9#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/fluentbit:main-logcollector#' + local LAST_ERR=/tmp/tmp.hLtmYtn9bo + local exit_status=0 + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' ++ seq 0 2 + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.IjF0xSvAIW deployment.apps/backup-client created + cat /tmp/tmp.hLtmYtn9bo + rm /tmp/tmp.IjF0xSvAIW /tmp/tmp.hLtmYtn9bo + return 0 + sleep 10 ++ kubectl_bin get pods --selector=name=backup-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yKFIXwwem7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.xnvGu7fdkz ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=backup-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.yKFIXwwem7 ++ cat /tmp/tmp.xnvGu7fdkz ++ rm /tmp/tmp.yKFIXwwem7 /tmp/tmp.xnvGu7fdkz ++ return 0 + bak_client_pod=backup-client-875cd9445-vcx22 + wait_pod backup-client-875cd9445-vcx22 + local pod=backup-client-875cd9445-vcx22 + local max_retry=480 + local ns= ++ echo backup-client-875cd9445-vcx22 ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/backup-client-875cd9445-vcx22 condition met waiting for pod/backup-client-875cd9445-vcx22 to become Ready.Ok + kubectl_bin exec backup-client-875cd9445-vcx22 -- bash -c 'cd /backup; md5sum -c md5sum.txt' ++ mktemp + local LAST_OUT=/tmp/tmp.ZG2hXaZcwq ++ mktemp + local LAST_ERR=/tmp/tmp.bbcKIlAwxp + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec backup-client-875cd9445-vcx22 -- bash -c 'cd /backup; md5sum -c md5sum.txt' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ZG2hXaZcwq xtrabackup.stream: OK + cat /tmp/tmp.bbcKIlAwxp + rm /tmp/tmp.ZG2hXaZcwq /tmp/tmp.bbcKIlAwxp + return 0 + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.6oBPOxhXpn ++ mktemp + local LAST_ERR=/tmp/tmp.v57ImkwXL5 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/client.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.6oBPOxhXpn deployment.apps "backup-client" deleted from demand-backup-28441 namespace + cat /tmp/tmp.v57ImkwXL5 + rm /tmp/tmp.6oBPOxhXpn /tmp/tmp.v57ImkwXL5 + return 0 + check_finalizer_for_fs on-demand-backup-pvc + local backup_name=on-demand-backup-pvc + local finalizer_flag=off + desc 'Checking finalizers on backup on-demand-backup-pvc' + set +o xtrace ----------------------------------------------------------------------------------- Checking finalizers on backup on-demand-backup-pvc ----------------------------------------------------------------------------------- + kubectl get pxc-backup on-demand-backup-pvc -o 'jsonpath={.metadata.finalizers}' + grep -q percona.com/delete-backup + echo 'Finalizer is off' Finalizer is off + echo 'Looking for PVCs containing '\''on-demand-backup-pvc'\''' Looking for PVCs containing 'on-demand-backup-pvc' ++ kubectl get pvc --no-headers -o custom-columns=:metadata.name ++ grep on-demand-backup-pvc + pvc=xb-on-demand-backup-pvc-20260330101253-2d0422b9 + [[ -z xb-on-demand-backup-pvc-20260330101253-2d0422b9 ]] + echo 'Deleting backup on-demand-backup-pvc ...' Deleting backup on-demand-backup-pvc ... + kubectl delete pxc-backup on-demand-backup-pvc perconaxtradbclusterbackup.pxc.percona.com "on-demand-backup-pvc" deleted from demand-backup-28441 namespace + sleep 5 ++ kubectl get pvc --no-headers -o custom-columns=:metadata.name ++ grep on-demand-backup-pvc + pvc=xb-on-demand-backup-pvc-20260330101253-2d0422b9 + echo 'Checking PVC xb-on-demand-backup-pvc-20260330101253-2d0422b9 existence and finalizers' Checking PVC xb-on-demand-backup-pvc-20260330101253-2d0422b9 existence and finalizers + [[ -z xb-on-demand-backup-pvc-20260330101253-2d0422b9 ]] + [[ -z xb-on-demand-backup-pvc-20260330101253-2d0422b9 ]] + [[ -n xb-on-demand-backup-pvc-20260330101253-2d0422b9 ]] + [[ off == \o\n ]] + echo 'PVC exists: xb-on-demand-backup-pvc-20260330101253-2d0422b9 (finalizer_flag=off)' PVC exists: xb-on-demand-backup-pvc-20260330101253-2d0422b9 (finalizer_flag=off) + run_backup demand-backup on-demand-backup-pvc-with-finalizer + local cluster=demand-backup + local backup=on-demand-backup-pvc-with-finalizer + log 'run pxc-backup/on-demand-backup-pvc-with-finalizer' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-03-30T10:22:40+0000]' run pxc-backup/on-demand-backup-pvc-with-finalizer [2026-03-30T10:22:40+0000] run pxc-backup/on-demand-backup-pvc-with-finalizer + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/on-demand-backup-pvc-with-finalizer.yml ++ mktemp + local LAST_OUT=/tmp/tmp.8Ay2ICtwCu ++ mktemp + local LAST_ERR=/tmp/tmp.QXzd9JGtfc + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/on-demand-backup-pvc-with-finalizer.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.8Ay2ICtwCu perconaxtradbclusterbackup.pxc.percona.com/on-demand-backup-pvc-with-finalizer created + cat /tmp/tmp.QXzd9JGtfc + rm /tmp/tmp.8Ay2ICtwCu /tmp/tmp.QXzd9JGtfc + return 0 + wait_backup on-demand-backup-pvc-with-finalizer + local backup=on-demand-backup-pvc-with-finalizer + local status=Succeeded + set +o xtrace waiting for pxc-backup/on-demand-backup-pvc-with-finalizer to reach Succeeded state..........................Succeeded + check_finalizer_for_fs on-demand-backup-pvc-with-finalizer + local backup_name=on-demand-backup-pvc-with-finalizer + local finalizer_flag=off + desc 'Checking finalizers on backup on-demand-backup-pvc-with-finalizer' + set +o xtrace ----------------------------------------------------------------------------------- Checking finalizers on backup on-demand-backup-pvc-with-finalizer ----------------------------------------------------------------------------------- + kubectl get pxc-backup on-demand-backup-pvc-with-finalizer -o 'jsonpath={.metadata.finalizers}' + grep -q percona.com/delete-backup + finalizer_flag=on + echo 'Finalizer is on' Finalizer is on + echo 'Looking for PVCs containing '\''on-demand-backup-pvc-with-finalizer'\''' Looking for PVCs containing 'on-demand-backup-pvc-with-finalizer' ++ grep on-demand-backup-pvc-with-finalizer ++ kubectl get pvc --no-headers -o custom-columns=:metadata.name + pvc=xb-on-demand-backup-pvc-with-finalizer-20260330102242-3ae22e3d + [[ -z xb-on-demand-backup-pvc-with-finalizer-20260330102242-3ae22e3d ]] + echo 'Deleting backup on-demand-backup-pvc-with-finalizer ...' Deleting backup on-demand-backup-pvc-with-finalizer ... + kubectl delete pxc-backup on-demand-backup-pvc-with-finalizer perconaxtradbclusterbackup.pxc.percona.com "on-demand-backup-pvc-with-finalizer" deleted from demand-backup-28441 namespace + sleep 5 ++ grep on-demand-backup-pvc-with-finalizer ++ kubectl get pvc --no-headers -o custom-columns=:metadata.name ++ true + pvc= + echo 'Checking PVC existence and finalizers' Checking PVC existence and finalizers + [[ -z '' ]] + [[ on == \o\f\f ]] + [[ -z '' ]] + [[ on == \o\n ]] + echo 'Correct: No PVCs found with substring '\''on-demand-backup-pvc-with-finalizer'\'' because finalizer is ON and PVC was deleted.' Correct: No PVCs found with substring 'on-demand-backup-pvc-with-finalizer' because finalizer is ON and PVC was deleted. + run_backup demand-backup on-demand-backup-pvc-bsource + local cluster=demand-backup + local backup=on-demand-backup-pvc-bsource + log 'run pxc-backup/on-demand-backup-pvc-bsource' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-03-30T10:23:33+0000]' run pxc-backup/on-demand-backup-pvc-bsource [2026-03-30T10:23:33+0000] run pxc-backup/on-demand-backup-pvc-bsource + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/on-demand-backup-pvc-bsource.yml ++ mktemp + local LAST_OUT=/tmp/tmp.TzBRddBDat ++ mktemp + local LAST_ERR=/tmp/tmp.D8K2yULSeT + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/on-demand-backup-pvc-bsource.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.TzBRddBDat perconaxtradbclusterbackup.pxc.percona.com/on-demand-backup-pvc-bsource created + cat /tmp/tmp.D8K2yULSeT + rm /tmp/tmp.TzBRddBDat /tmp/tmp.D8K2yULSeT + return 0 + wait_backup on-demand-backup-pvc-bsource + local backup=on-demand-backup-pvc-bsource + local status=Succeeded + set +o xtrace waiting for pxc-backup/on-demand-backup-pvc-bsource to reach Succeeded state....................Succeeded + run_recovery_check demand-backup on-demand-backup-pvc-bsource + local cluster=demand-backup + local backup=on-demand-backup-pvc-bsource ++ get_proxy_engine demand-backup ++ local cluster_name=demand-backup +++ get_proxy demand-backup +++ local target_cluster=demand-backup ++++ kubectl_bin get pxc demand-backup -o 'jsonpath={.spec.haproxy.enabled}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.tz3gOHJP59 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.yPxk2WkT2J ++++ local exit_status=0 +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ set +e ++++ kubectl get pxc demand-backup -o 'jsonpath={.spec.haproxy.enabled}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 ']' ++++ break ++++ cat /tmp/tmp.tz3gOHJP59 ++++ cat /tmp/tmp.yPxk2WkT2J ++++ rm /tmp/tmp.tz3gOHJP59 /tmp/tmp.yPxk2WkT2J ++++ return 0 +++ [[ '' == \t\r\u\e ]] ++++ kubectl_bin get pxc demand-backup -o 'jsonpath={.spec.proxysql.enabled}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.1I9KyAJzKa +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.bRGiVsdKty ++++ local exit_status=0 +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ set +e ++++ kubectl get pxc demand-backup -o 'jsonpath={.spec.proxysql.enabled}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 ']' ++++ break ++++ cat /tmp/tmp.1I9KyAJzKa ++++ cat /tmp/tmp.bRGiVsdKty ++++ rm /tmp/tmp.1I9KyAJzKa /tmp/tmp.bRGiVsdKty ++++ return 0 +++ [[ true == \t\r\u\e ]] +++ echo demand-backup-proxysql +++ return ++ local cluster_proxy=demand-backup-proxysql ++ echo proxysql + local proxy=proxysql ++ get_pvc_name_for_backup on-demand-backup-pvc-bsource ++ local backup_name=on-demand-backup-pvc-bsource +++ kubectl_bin get pxc-backup on-demand-backup-pvc-bsource -o 'jsonpath={.status.destination}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.qlW85n0KB1 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.gWqxHutOSV +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc-backup on-demand-backup-pvc-bsource -o 'jsonpath={.status.destination}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.qlW85n0KB1 +++ cat /tmp/tmp.gWqxHutOSV +++ rm /tmp/tmp.qlW85n0KB1 /tmp/tmp.gWqxHutOSV +++ return 0 ++ local destination=pvc/xb-on-demand-backup-pvc-bsource-20260330102335-42dfe747 ++ [[ -z pvc/xb-on-demand-backup-pvc-bsource-20260330102335-42dfe747 ]] ++ local pvc_name=xb-on-demand-backup-pvc-bsource-20260330102335-42dfe747 ++ echo xb-on-demand-backup-pvc-bsource-20260330102335-42dfe747 + pvc_name=xb-on-demand-backup-pvc-bsource-20260330102335-42dfe747 + dest=pvc/xb-on-demand-backup-pvc-bsource-20260330102335-42dfe747 + log 'run pxc-restore/on-demand-backup-pvc-bsource' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-03-30T10:24:22+0000]' run pxc-restore/on-demand-backup-pvc-bsource [2026-03-30T10:24:22+0000] run pxc-restore/on-demand-backup-pvc-bsource + YAML_INPUT_FILE=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/restore-on-demand-backup-pvc-bsource.yaml ++ /usr/bin/sed -e s~minio-service.#namespace~minio-service.demand-backup-28441~ ++ /usr/bin/sed -e 's##pvc/xb-on-demand-backup-pvc-bsource-20260330102335-42dfe747#g' ++ cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/restore-on-demand-backup-pvc-bsource.yaml + MODIFIED_YAML='apiVersion: pxc.percona.com/v1 kind: PerconaXtraDBClusterRestore metadata: name: on-demand-backup-pvc-bsource spec: pxcCluster: demand-backup backupSource: destination: pvc/xb-on-demand-backup-pvc-bsource-20260330102335-42dfe747 storageName: pvc resources: requests: memory: 0.5G cpu: 500m limits: memory: "1G" cpu: "2"' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + FINAL_YAML='apiVersion: pxc.percona.com/v1 kind: PerconaXtraDBClusterRestore metadata: name: on-demand-backup-pvc-bsource spec: pxcCluster: demand-backup backupSource: destination: pvc/xb-on-demand-backup-pvc-bsource-20260330102335-42dfe747 storageName: pvc resources: requests: memory: 0.5G cpu: 500m limits: memory: "1G" cpu: "2"' + kubectl_bin apply -f - + echo 'apiVersion: pxc.percona.com/v1 kind: PerconaXtraDBClusterRestore metadata: name: on-demand-backup-pvc-bsource spec: pxcCluster: demand-backup backupSource: destination: pvc/xb-on-demand-backup-pvc-bsource-20260330102335-42dfe747 storageName: pvc resources: requests: memory: 0.5G cpu: 500m limits: memory: "1G" cpu: "2"' ++ mktemp + local LAST_OUT=/tmp/tmp.Y9BCIiuF5W ++ mktemp + local LAST_ERR=/tmp/tmp.0OyYHAgn5t + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Y9BCIiuF5W perconaxtradbclusterrestore.pxc.percona.com/on-demand-backup-pvc-bsource created + cat /tmp/tmp.0OyYHAgn5t + rm /tmp/tmp.Y9BCIiuF5W /tmp/tmp.0OyYHAgn5t + return 0 + wait_backup_restore on-demand-backup-pvc-bsource + local backup_name=on-demand-backup-pvc-bsource + local target_state=Succeeded + local wait_time=720 + set +o xtrace waiting for pxc-restore/on-demand-backup-pvc-bsource to reach Succeeded state 2026-03-30T10:24:26 pxc-restore/on-demand-backup-pvc-bsource state: Starting 2026-03-30T10:24:28 pxc-restore/on-demand-backup-pvc-bsource state: Starting 2026-03-30T10:24:30 pxc-restore/on-demand-backup-pvc-bsource state: Starting 2026-03-30T10:24:33 pxc-restore/on-demand-backup-pvc-bsource state: Starting 2026-03-30T10:24:35 pxc-restore/on-demand-backup-pvc-bsource state: Starting 2026-03-30T10:24:38 pxc-restore/on-demand-backup-pvc-bsource state: Starting 2026-03-30T10:24:41 pxc-restore/on-demand-backup-pvc-bsource state: Stopping Cluster 2026-03-30T10:24:43 pxc-restore/on-demand-backup-pvc-bsource state: Stopping Cluster 2026-03-30T10:24:45 pxc-restore/on-demand-backup-pvc-bsource state: Stopping Cluster 2026-03-30T10:24:48 pxc-restore/on-demand-backup-pvc-bsource state: Stopping Cluster 2026-03-30T10:24:51 pxc-restore/on-demand-backup-pvc-bsource state: Stopping Cluster 2026-03-30T10:24:54 pxc-restore/on-demand-backup-pvc-bsource state: Stopping Cluster 2026-03-30T10:24:57 pxc-restore/on-demand-backup-pvc-bsource state: Stopping Cluster 2026-03-30T10:25:00 pxc-restore/on-demand-backup-pvc-bsource state: Stopping Cluster 2026-03-30T10:25:02 pxc-restore/on-demand-backup-pvc-bsource state: Stopping Cluster 2026-03-30T10:25:05 pxc-restore/on-demand-backup-pvc-bsource state: Stopping Cluster 2026-03-30T10:25:07 pxc-restore/on-demand-backup-pvc-bsource state: Stopping Cluster 2026-03-30T10:25:10 pxc-restore/on-demand-backup-pvc-bsource state: Stopping Cluster 2026-03-30T10:25:12 pxc-restore/on-demand-backup-pvc-bsource state: Stopping Cluster 2026-03-30T10:25:15 pxc-restore/on-demand-backup-pvc-bsource state: Stopping Cluster 2026-03-30T10:25:17 pxc-restore/on-demand-backup-pvc-bsource state: Stopping Cluster 2026-03-30T10:25:19 pxc-restore/on-demand-backup-pvc-bsource state: Stopping Cluster 2026-03-30T10:25:22 pxc-restore/on-demand-backup-pvc-bsource state: Stopping Cluster 2026-03-30T10:25:24 pxc-restore/on-demand-backup-pvc-bsource state: Stopping Cluster 2026-03-30T10:25:26 pxc-restore/on-demand-backup-pvc-bsource state: Stopping Cluster 2026-03-30T10:25:28 pxc-restore/on-demand-backup-pvc-bsource state: Stopping Cluster 2026-03-30T10:25:30 pxc-restore/on-demand-backup-pvc-bsource state: Stopping Cluster 2026-03-30T10:25:32 pxc-restore/on-demand-backup-pvc-bsource state: Stopping Cluster 2026-03-30T10:25:34 pxc-restore/on-demand-backup-pvc-bsource state: Restoring 2026-03-30T10:25:36 pxc-restore/on-demand-backup-pvc-bsource state: Restoring 2026-03-30T10:25:39 pxc-restore/on-demand-backup-pvc-bsource state: Restoring 2026-03-30T10:25:41 pxc-restore/on-demand-backup-pvc-bsource state: Restoring 2026-03-30T10:25:43 pxc-restore/on-demand-backup-pvc-bsource state: Restoring 2026-03-30T10:25:45 pxc-restore/on-demand-backup-pvc-bsource state: Restoring 2026-03-30T10:25:47 pxc-restore/on-demand-backup-pvc-bsource state: Restoring 2026-03-30T10:25:50 pxc-restore/on-demand-backup-pvc-bsource state: Restoring 2026-03-30T10:25:52 pxc-restore/on-demand-backup-pvc-bsource state: Restoring 2026-03-30T10:25:54 pxc-restore/on-demand-backup-pvc-bsource state: Restoring 2026-03-30T10:25:56 pxc-restore/on-demand-backup-pvc-bsource state: Restoring 2026-03-30T10:25:58 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:01 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:03 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:05 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:07 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:09 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:11 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:13 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:15 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:17 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:19 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:21 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:24 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:26 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:28 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:30 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:33 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:35 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:37 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:39 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:42 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:44 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:46 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:48 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:49 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:51 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:54 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:56 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:26:58 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:27:00 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:27:02 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:27:04 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:27:07 pxc-restore/on-demand-backup-pvc-bsource state: Preparing Cluster 2026-03-30T10:27:10 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:27:13 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:27:15 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:27:18 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:27:20 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:27:22 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:27:25 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:27:27 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:27:29 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:27:32 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:27:34 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:27:36 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:27:38 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:27:41 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:27:43 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:27:45 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:27:47 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:27:49 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:27:51 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:27:53 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:27:56 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:27:58 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:00 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:03 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:05 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:07 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:10 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:12 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:14 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:17 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:19 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:21 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:23 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:26 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:28 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:30 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:33 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:35 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:37 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:39 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:42 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:44 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:46 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:48 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:51 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:53 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:55 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:28:57 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:00 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:02 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:03 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:05 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:08 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:10 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:12 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:14 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:17 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:19 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:21 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:23 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:25 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:27 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:30 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:32 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:34 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:36 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:38 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:40 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:42 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:44 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:46 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:48 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:50 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:53 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:55 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:29:57 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:30:00 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:30:02 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:30:04 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:30:06 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:30:08 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:30:10 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:30:12 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:30:14 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:30:16 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:30:18 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:30:20 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:30:22 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:30:25 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:30:27 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:30:28 pxc-restore/on-demand-backup-pvc-bsource state: Starting Cluster 2026-03-30T10:30:29 pxc-restore/on-demand-backup-pvc-bsource state: Succeeded + kubectl_bin logs job/restore-job-on-demand-backup-pvc-bsource-demand-backup ++ mktemp + local LAST_OUT=/tmp/tmp.fkyakqzZlF ++ mktemp + local LAST_ERR=/tmp/tmp.WF3JEsS2bA + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl logs job/restore-job-on-demand-backup-pvc-bsource-demand-backup + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.fkyakqzZlF + LIB_PATH=/opt/percona/backup/lib/pxc + . /opt/percona/backup/lib/pxc/check-version.sh + . /opt/percona/backup/lib/pxc/vault.sh ++ set -o errexit ++ keyring_vault=/etc/mysql/vault-keyring-secret/keyring_vault.conf + SOCAT_OPTS=TCP:restore-src-on-demand-backup-pvc-bsource-demand-backup:3307,retry=30 + check_ssl + CA=/var/run/secrets/kubernetes.io/serviceaccount/ca.crt + '[' -f /var/run/secrets/kubernetes.io/serviceaccount/service-ca.crt ']' + SSL_DIR=/etc/mysql/ssl + '[' -f /etc/mysql/ssl/ca.crt ']' + CA=/etc/mysql/ssl/ca.crt + SSL_INTERNAL_DIR=/etc/mysql/ssl-internal + '[' -f /etc/mysql/ssl-internal/ca.crt ']' + CA=/etc/mysql/ssl-internal/ca.crt + KEY=/etc/mysql/ssl/tls.key + CERT=/etc/mysql/ssl/tls.crt + '[' -f /etc/mysql/ssl-internal/tls.key ']' + '[' -f /etc/mysql/ssl-internal/tls.crt ']' + KEY=/etc/mysql/ssl-internal/tls.key + CERT=/etc/mysql/ssl-internal/tls.crt + '[' -f /etc/mysql/ssl-internal/ca.crt ']' + '[' -f /etc/mysql/ssl-internal/tls.key ']' + '[' -f /etc/mysql/ssl-internal/tls.crt ']' + SOCAT_OPTS='openssl-connect:restore-src-on-demand-backup-pvc-bsource-demand-backup:3307,reuseaddr,cert=/etc/mysql/ssl-internal/tls.crt,key=/etc/mysql/ssl-internal/tls.key,cafile=/etc/mysql/ssl-internal/ca.crt,verify=1,commonname='\'''\'',retry=30,no-sni=1' + ping -c1 restore-src-on-demand-backup-pvc-bsource-demand-backup /opt/percona/backup/recovery-pvc-joiner.sh: line 40: ping: command not found + : + rm -rf /datadir/#ib_16384_0.dblwr /datadir/#ib_16384_1.dblwr /datadir/#ib_16384_10.dblwr /datadir/#ib_16384_11.dblwr /datadir/#ib_16384_12.dblwr /datadir/#ib_16384_13.dblwr /datadir/#ib_16384_14.dblwr /datadir/#ib_16384_15.dblwr /datadir/#ib_16384_2.dblwr /datadir/#ib_16384_3.dblwr /datadir/#ib_16384_4.dblwr /datadir/#ib_16384_5.dblwr /datadir/#ib_16384_6.dblwr /datadir/#ib_16384_7.dblwr /datadir/#ib_16384_8.dblwr /datadir/#ib_16384_9.dblwr /datadir/#innodb_redo /datadir/#innodb_temp /datadir/auth_plugin /datadir/auto.cnf /datadir/binlog.000009 /datadir/binlog.000010 /datadir/binlog.000011 /datadir/binlog.000012 /datadir/binlog.index /datadir/ca-key.pem /datadir/ca.pem /datadir/client-cert.pem /datadir/client-key.pem /datadir/galera.cache /datadir/get-pxc-state /datadir/grastate.dat /datadir/ib_buffer_pool /datadir/ibdata1 /datadir/innobackup.backup.log /datadir/liveness-check.sh /datadir/myApp /datadir/mysql /datadir/mysql-state-monitor /datadir/mysql-state-monitor.log /datadir/mysql.ibd /datadir/mysql.state /datadir/notify.sock /datadir/peer-list /datadir/performance_schema /datadir/pmm-prerun.sh /datadir/prepare-job-on-demand-backup-pvc-demand-backup-mbpv4.pid /datadir/prepare_restored_cluster.sh /datadir/private_key.pem /datadir/public_key.pem /datadir/pxc-configure-pxc.sh /datadir/pxc-entrypoint.sh /datadir/readiness-check.sh /datadir/server-cert.pem /datadir/server-key.pem /datadir/sst_info /datadir/sys /datadir/undo_001 /datadir/undo_002 /datadir/version_info /datadir/wsrep_cmd_notify_handler.sh /datadir/wsrep_recovery_verbose_history.log /datadir/xtrabackup-server-sidecar /datadir/xtrabackup_info ++ mktemp --directory /datadir/pxc_sst_XXXX + tmp=/datadir/pxc_sst_NiRR + socat -u 'openssl-connect:restore-src-on-demand-backup-pvc-bsource-demand-backup:3307,reuseaddr,cert=/etc/mysql/ssl-internal/tls.crt,key=/etc/mysql/ssl-internal/tls.key,cafile=/etc/mysql/ssl-internal/ca.crt,verify=1,commonname='\'''\'',retry=30,no-sni=1' stdio ++ get_xtrabackup_version +++ xtrabackup --version +++ grep '^xtrabackup version' +++ sed 's/-.*//' +++ awk '{print $3}' ++ xtrabackup_ver=8.0.35 ++ echo 8.0.35 + XTRABACKUP_VERSION=8.0.35 + check_for_version 8.0.35 8.0.0 + '[' -z 8.0.35 ']' + '[' -z 8.0.0 ']' + local local_version_str + local required_version_str ++ normalize_version 8.0.35 ++ local major=0 ++ local minor=0 ++ local patch=0 ++ [[ 8.0.35 =~ ^([0-9]+)\.([0-9]+)\.?([0-9]*)([^ ])* ]] ++ major=8 ++ minor=0 ++ patch=35 ++ printf %02d%02d%02d 8 0 35 + local_version_str=080035 ++ normalize_version 8.0.0 ++ local major=0 ++ local minor=0 ++ local patch=0 ++ [[ 8.0.0 =~ ^([0-9]+)\.([0-9]+)\.?([0-9]*)([^ ])* ]] ++ major=8 ++ minor=0 ++ patch=0 ++ printf %02d%02d%02d 8 0 0 + required_version_str=080000 + [[ 080035 < 080000 ]] + return 0 + XBSTREAM_EXTRA_ARGS=' --decompress' + socat -u 'openssl-connect:restore-src-on-demand-backup-pvc-bsource-demand-backup:3307,reuseaddr,cert=/etc/mysql/ssl-internal/tls.crt,key=/etc/mysql/ssl-internal/tls.key,cafile=/etc/mysql/ssl-internal/ca.crt,verify=1,commonname='\'''\'',retry=30,no-sni=1' stdio ++ grep -c processor /proc/cpuinfo + xbstream -x -C /datadir/pxc_sst_NiRR --parallel=4 --decompress + PXB_VAULT_PREPARE_ARGS= + PXB_VAULT_MOVEBACK_ARGS= + VAULT_CONFIG_FILE=/etc/mysql/vault-keyring-secret/keyring_vault.conf + VAULT_KEYRING_COMPONENT=/opt/percona/component_keyring_vault.cnf + [[ -f /etc/mysql/vault-keyring-secret/keyring_vault.conf ]] + set +o xtrace vault configuration not found + xtrabackup --use-memory=750000000 --prepare --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_NiRR 2026-03-30T10:25:48.532320-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized server arguments: --innodb_checksum_algorithm=crc32 --innodb_log_checksums=1 --innodb_data_file_path=ibdata1:12M:autoextend --innodb_log_file_size=50331648 --innodb_page_size=16384 --innodb_undo_directory=./ --innodb_undo_tablespaces=2 --server-id=24613582 --innodb_log_checksums=ON --innodb_redo_log_encrypt=0 --innodb_undo_log_encrypt=0 2026-03-30T10:25:48.532404-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized client arguments: --use-memory=750000000 --prepare=1 --rollback-prepared-trx=1 --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_NiRR xtrabackup version 8.0.35-34 based on MySQL server 8.0.35 Linux (x86_64) (revision id: c8a25ff9) 2026-03-30T10:25:48.532437-00:00 0 [Note] [MY-011825] [Xtrabackup] cd to /datadir/pxc_sst_NiRR/ 2026-03-30T10:25:48.532495-00:00 0 [Note] [MY-011825] [Xtrabackup] This target seems to be not prepared yet. 2026-03-30T10:25:48.541947-00:00 0 [Note] [MY-011825] [Xtrabackup] xtrabackup_logfile detected: size=8388608, start_lsn=(30926558) 2026-03-30T10:25:48.542739-00:00 0 [Note] [MY-011825] [Xtrabackup] using the following InnoDB configuration for recovery: 2026-03-30T10:25:48.542753-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_home_dir = . 2026-03-30T10:25:48.542760-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_file_path = ibdata1:12M:autoextend 2026-03-30T10:25:48.542787-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_group_home_dir = . 2026-03-30T10:25:48.542796-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_files_in_group = 1 2026-03-30T10:25:48.542803-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_file_size = 8388608 2026-03-30T10:25:48.542981-00:00 0 [Note] [MY-011825] [Xtrabackup] inititialize_service_handles suceeded 2026-03-30T10:25:48.543154-00:00 0 [Note] [MY-011825] [Xtrabackup] using the following InnoDB configuration for recovery: 2026-03-30T10:25:48.543166-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_home_dir = . 2026-03-30T10:25:48.543172-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_file_path = ibdata1:12M:autoextend 2026-03-30T10:25:48.543181-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_group_home_dir = . 2026-03-30T10:25:48.543187-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_files_in_group = 1 2026-03-30T10:25:48.543192-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_file_size = 8388608 2026-03-30T10:25:48.543203-00:00 0 [Note] [MY-011825] [Xtrabackup] Starting InnoDB instance for recovery. 2026-03-30T10:25:48.543209-00:00 0 [Note] [MY-011825] [Xtrabackup] Using 749731840 bytes for buffer pool (set by --use-memory parameter) 2026-03-30T10:25:48.543237-00:00 0 [Note] [MY-012932] [InnoDB] PUNCH HOLE support available 2026-03-30T10:25:48.543250-00:00 0 [Note] [MY-012944] [InnoDB] Uses event mutexes 2026-03-30T10:25:48.543256-00:00 0 [Note] [MY-012945] [InnoDB] GCC builtin __atomic_thread_fence() is used for memory barrier 2026-03-30T10:25:48.543262-00:00 0 [Note] [MY-012948] [InnoDB] Compressed tables use zlib 1.2.13 2026-03-30T10:25:48.543412-00:00 0 [Note] [MY-012951] [InnoDB] Using hardware accelerated crc32 and polynomial multiplication. 2026-03-30T10:25:48.543766-00:00 0 [Note] [MY-012203] [InnoDB] Directories to scan './' 2026-03-30T10:25:48.543807-00:00 0 [Note] [MY-012204] [InnoDB] Scanning './' 2026-03-30T10:25:48.553170-00:00 0 [Note] [MY-012208] [InnoDB] Completed space ID check of 8 files. 2026-03-30T10:25:48.553714-00:00 0 [Note] [MY-012955] [InnoDB] Initializing buffer pool, total size = 768.000000M, instances = 1, chunk size =128.000000M 2026-03-30T10:25:48.593106-00:00 0 [Note] [MY-012957] [InnoDB] Completed initialization of buffer pool 2026-03-30T10:25:48.598214-00:00 0 [Note] [MY-011952] [InnoDB] If the mysqld execution user is authorized, page cleaner thread priority can be changed. See the man page of setpriority(). 2026-03-30T10:25:48.626693-00:00 0 [Note] [MY-013883] [InnoDB] The latest found checkpoint is at lsn = 30926558 in redo log file ./#innodb_redo/#ib_redo0. 2026-03-30T10:25:48.626760-00:00 0 [Note] [MY-012560] [InnoDB] The log sequence number 30809585 in the system tablespace does not match the log sequence number 30926558 in the redo log files! 2026-03-30T10:25:48.626769-00:00 0 [Note] [MY-012551] [InnoDB] Database was not shutdown normally! 2026-03-30T10:25:48.626775-00:00 0 [Note] [MY-012552] [InnoDB] Starting crash recovery. 2026-03-30T10:25:48.626900-00:00 0 [Note] [MY-013086] [InnoDB] Starting to parse redo log at lsn = 30926389, whereas checkpoint_lsn = 30926558 and start_lsn = 30926336 2026-03-30T10:25:48.626912-00:00 0 [Note] [MY-012550] [InnoDB] Doing recovery: scanned up to log sequence number 30926568 2026-03-30T10:25:48.643421-00:00 0 [Note] [MY-013083] [InnoDB] Log background threads are being started... 2026-03-30T10:25:48.646524-00:00 0 [Note] [MY-012532] [InnoDB] Applying a batch of 1 redo log records ... 2026-03-30T10:25:48.646580-00:00 0 [Note] [MY-012533] [InnoDB] 100% 2026-03-30T10:25:48.646593-00:00 0 [Note] [MY-012535] [InnoDB] Apply batch completed! 2026-03-30T10:25:48.746710-00:00 0 [Note] [MY-013084] [InnoDB] Log background threads are being closed... 2026-03-30T10:25:48.748206-00:00 0 [Note] [MY-013888] [InnoDB] Upgrading redo log: 1032M, LSN=30926568. 2026-03-30T10:25:48.748219-00:00 0 [Note] [MY-012968] [InnoDB] Starting to delete and rewrite redo log files. 2026-03-30T10:25:48.748269-00:00 0 [Note] [MY-011825] [InnoDB] Removing redo log file: ./#innodb_redo/#ib_redo0 2026-03-30T10:25:48.777674-00:00 0 [Note] [MY-011825] [InnoDB] Creating redo log file at ./#innodb_redo/#ib_redo0_tmp with file_id 0 with size 33554432 bytes 2026-03-30T10:25:48.783796-00:00 0 [Note] [MY-011825] [InnoDB] Renaming redo log file from ./#innodb_redo/#ib_redo0_tmp to ./#innodb_redo/#ib_redo0 2026-03-30T10:25:48.789051-00:00 0 [Note] [MY-012893] [InnoDB] New redo log files created, LSN=30926860 2026-03-30T10:25:48.789116-00:00 0 [Note] [MY-013083] [InnoDB] Log background threads are being started... 2026-03-30T10:25:48.791874-00:00 0 [Note] [MY-013252] [InnoDB] Using undo tablespace './undo_001'. 2026-03-30T10:25:48.793220-00:00 0 [Note] [MY-013252] [InnoDB] Using undo tablespace './undo_002'. 2026-03-30T10:25:48.794836-00:00 0 [Note] [MY-012910] [InnoDB] Opened 2 existing undo tablespaces. 2026-03-30T10:25:48.794890-00:00 0 [Note] [MY-011980] [InnoDB] GTID recovery trx_no: 10328 2026-03-30T10:25:48.935838-00:00 0 [Note] [MY-013776] [InnoDB] Parallel initialization of rseg complete 2026-03-30T10:25:48.935865-00:00 0 [Note] [MY-013777] [InnoDB] Time taken to initialize rseg using 4 thread: 140976 ms. 2026-03-30T10:25:48.935925-00:00 0 [Note] [MY-012923] [InnoDB] Creating shared tablespace for temporary tables 2026-03-30T10:25:48.935986-00:00 0 [Note] [MY-012265] [InnoDB] Setting file './ibtmp1' size to 12 MB. Physically writing the file full; Please wait ... 2026-03-30T10:25:48.967689-00:00 0 [Note] [MY-012266] [InnoDB] File './ibtmp1' size is now 12 MB. 2026-03-30T10:25:48.967833-00:00 0 [Note] [MY-013627] [InnoDB] Scanning temp tablespace dir:'./#innodb_temp/' 2026-03-30T10:25:49.002014-00:00 0 [Note] [MY-013018] [InnoDB] Created 128 and tracked 128 new rollback segment(s) in the temporary tablespace. 128 are now active. 2026-03-30T10:25:49.002220-00:00 0 [Note] [MY-012976] [InnoDB] 8.0.35 started; log sequence number 30926870 2026-03-30T10:25:49.003347-00:00 0 [Warning] [MY-012091] [InnoDB] Allocated tablespace ID 1 for sys/sys_config, old maximum was 0 2026-03-30T10:25:49.008835-00:00 0 [Note] [MY-011825] [Xtrabackup] Completed loading of 6 tablespaces into cache in 0.00659087 seconds 2026-03-30T10:25:49.029173-00:00 0 [Note] [MY-011825] [Xtrabackup] Time taken to build dictionary: 0.0202961 seconds 2026-03-30T10:25:50.029749-00:00 0 [Note] [MY-011825] [Xtrabackup] Recovered WSREP position: 9c7de8c4-2c21-11f1-99b4-b27407564189:6 2026-03-30T10:25:50.029798-00:00 0 [Note] [MY-011825] [Xtrabackup] starting shutdown with innodb_fast_shutdown = 1 2026-03-30T10:25:50.029851-00:00 0 [Note] [MY-012330] [InnoDB] FTS optimize thread exiting. 2026-03-30T10:25:51.029737-00:00 0 [Note] [MY-013072] [InnoDB] Starting shutdown... 2026-03-30T10:25:51.130743-00:00 0 [Note] [MY-013084] [InnoDB] Log background threads are being closed... 2026-03-30T10:25:51.151418-00:00 0 [Note] [MY-012980] [InnoDB] Shutdown completed; log sequence number 30926870 2026-03-30T10:25:51.152891-00:00 0 [Note] [MY-011825] [Xtrabackup] completed OK! + xtrabackup --defaults-group=mysqld --datadir=/datadir --move-back --force-non-empty-directories --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_NiRR 2026-03-30T10:25:51.170161-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized server arguments: --defaults_group=mysqld --datadir=/datadir 2026-03-30T10:25:51.170214-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized client arguments: --move-back=1 --force-non-empty-directories=1 --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_NiRR xtrabackup version 8.0.35-34 based on MySQL server 8.0.35 Linux (x86_64) (revision id: c8a25ff9) 2026-03-30T10:25:51.170244-00:00 0 [Note] [MY-011825] [Xtrabackup] cd to /datadir/pxc_sst_NiRR/ 2026-03-30T10:25:51.170795-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving undo_001 to /datadir/undo_001 2026-03-30T10:25:51.170847-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file undo_001 to /datadir/undo_001 2026-03-30T10:25:51.170880-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving undo_002 to /datadir/undo_002 2026-03-30T10:25:51.170901-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file undo_002 to /datadir/undo_002 2026-03-30T10:25:51.171052-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving ibdata1 to /datadir/ibdata1 2026-03-30T10:25:51.171074-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file ibdata1 to /datadir/ibdata1 2026-03-30T10:25:51.171295-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving binlog.000019 to /datadir//binlog.000019 2026-03-30T10:25:51.171319-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file binlog.000019 to /datadir//binlog.000019 2026-03-30T10:25:51.171393-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving binlog.index to /datadir//binlog.index 2026-03-30T10:25:51.171421-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file binlog.index to /datadir//binlog.index 2026-03-30T10:25:51.171856-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./myApp/myApp.ibd to /datadir/myApp/myApp.ibd 2026-03-30T10:25:51.171897-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./myApp/myApp.ibd to /datadir/myApp/myApp.ibd 2026-03-30T10:25:51.171906-00:00 1 [Note] [MY-011825] [Xtrabackup] Creating directory ./#innodb_redo 2026-03-30T10:25:51.171918-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: creating directory ./#innodb_redo 2026-03-30T10:25:51.171949-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql.ibd to /datadir/mysql.ibd 2026-03-30T10:25:51.171981-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql.ibd to /datadir/mysql.ibd 2026-03-30T10:25:51.172015-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./xtrabackup_galera_info to /datadir/xtrabackup_galera_info 2026-03-30T10:25:51.172040-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./xtrabackup_galera_info to /datadir/xtrabackup_galera_info 2026-03-30T10:25:51.172070-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./ib_buffer_pool to /datadir/ib_buffer_pool 2026-03-30T10:25:51.172102-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./ib_buffer_pool to /datadir/ib_buffer_pool 2026-03-30T10:25:51.172131-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./ibtmp1 to /datadir/ibtmp1 2026-03-30T10:25:51.172149-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./ibtmp1 to /datadir/ibtmp1 2026-03-30T10:25:51.172212-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_cluster.ibd to /datadir/mysql/wsrep_cluster.ibd 2026-03-30T10:25:51.172236-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_cluster.ibd to /datadir/mysql/wsrep_cluster.ibd 2026-03-30T10:25:51.172256-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log_225.sdi to /datadir/mysql/general_log_225.sdi 2026-03-30T10:25:51.172271-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log_225.sdi to /datadir/mysql/general_log_225.sdi 2026-03-30T10:25:51.172289-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_streaming_log.ibd to /datadir/mysql/wsrep_streaming_log.ibd 2026-03-30T10:25:51.172303-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_streaming_log.ibd to /datadir/mysql/wsrep_streaming_log.ibd 2026-03-30T10:25:51.172320-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log.CSV to /datadir/mysql/general_log.CSV 2026-03-30T10:25:51.172334-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log.CSV to /datadir/mysql/general_log.CSV 2026-03-30T10:25:51.172349-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log.CSM to /datadir/mysql/slow_log.CSM 2026-03-30T10:25:51.172362-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log.CSM to /datadir/mysql/slow_log.CSM 2026-03-30T10:25:51.172377-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log_226.sdi to /datadir/mysql/slow_log_226.sdi 2026-03-30T10:25:51.172391-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log_226.sdi to /datadir/mysql/slow_log_226.sdi 2026-03-30T10:25:51.172406-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log.CSV to /datadir/mysql/slow_log.CSV 2026-03-30T10:25:51.172420-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log.CSV to /datadir/mysql/slow_log.CSV 2026-03-30T10:25:51.172435-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_cluster_members.ibd to /datadir/mysql/wsrep_cluster_members.ibd 2026-03-30T10:25:51.172449-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_cluster_members.ibd to /datadir/mysql/wsrep_cluster_members.ibd 2026-03-30T10:25:51.172465-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log.CSM to /datadir/mysql/general_log.CSM 2026-03-30T10:25:51.172478-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log.CSM to /datadir/mysql/general_log.CSM 2026-03-30T10:25:51.172493-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./sst_info to /datadir/sst_info 2026-03-30T10:25:51.172507-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./sst_info to /datadir/sst_info 2026-03-30T10:25:51.172556-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_host_187.sdi to /datadir/performance_schema/status_by_host_187.sdi 2026-03-30T10:25:51.172591-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_host_187.sdi to /datadir/performance_schema/status_by_host_187.sdi 2026-03-30T10:25:51.172610-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_101.sdi to /datadir/performance_schema/events_waits_sum_101.sdi 2026-03-30T10:25:51.172628-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_101.sdi to /datadir/performance_schema/events_waits_sum_101.sdi 2026-03-30T10:25:51.172644-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_125.sdi to /datadir/performance_schema/events_stages_su_125.sdi 2026-03-30T10:25:51.172661-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_125.sdi to /datadir/performance_schema/events_stages_su_125.sdi 2026-03-30T10:25:51.172677-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/metadata_locks_168.sdi to /datadir/performance_schema/metadata_locks_168.sdi 2026-03-30T10:25:51.172693-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/metadata_locks_168.sdi to /datadir/performance_schema/metadata_locks_168.sdi 2026-03-30T10:25:51.172708-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/accounts_154.sdi to /datadir/performance_schema/accounts_154.sdi 2026-03-30T10:25:51.172746-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/accounts_154.sdi to /datadir/performance_schema/accounts_154.sdi 2026-03-30T10:25:51.172767-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_136.sdi to /datadir/performance_schema/events_statement_136.sdi 2026-03-30T10:25:51.172785-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_136.sdi to /datadir/performance_schema/events_statement_136.sdi 2026-03-30T10:25:51.172818-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_grou_178.sdi to /datadir/performance_schema/replication_grou_178.sdi 2026-03-30T10:25:51.172839-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_grou_178.sdi to /datadir/performance_schema/replication_grou_178.sdi 2026-03-30T10:25:51.172856-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/rwlock_instances_110.sdi to /datadir/performance_schema/rwlock_instances_110.sdi 2026-03-30T10:25:51.172870-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/rwlock_instances_110.sdi to /datadir/performance_schema/rwlock_instances_110.sdi 2026-03-30T10:25:51.172885-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_137.sdi to /datadir/performance_schema/events_statement_137.sdi 2026-03-30T10:25:51.172901-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_137.sdi to /datadir/performance_schema/events_statement_137.sdi 2026-03-30T10:25:51.172916-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_151.sdi to /datadir/performance_schema/events_errors_su_151.sdi 2026-03-30T10:25:51.172931-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_151.sdi to /datadir/performance_schema/events_errors_su_151.sdi 2026-03-30T10:25:51.172946-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_145.sdi to /datadir/performance_schema/events_transacti_145.sdi 2026-03-30T10:25:51.172962-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_145.sdi to /datadir/performance_schema/events_transacti_145.sdi 2026-03-30T10:25:51.172977-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_cur_93.sdi to /datadir/performance_schema/events_waits_cur_93.sdi 2026-03-30T10:25:51.172992-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_cur_93.sdi to /datadir/performance_schema/events_waits_cur_93.sdi 2026-03-30T10:25:51.173007-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/variables_by_thr_192.sdi to /datadir/performance_schema/variables_by_thr_192.sdi 2026-03-30T10:25:51.173023-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/variables_by_thr_192.sdi to /datadir/performance_schema/variables_by_thr_192.sdi 2026-03-30T10:25:51.173038-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_97.sdi to /datadir/performance_schema/events_waits_sum_97.sdi 2026-03-30T10:25:51.173054-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_97.sdi to /datadir/performance_schema/events_waits_sum_97.sdi 2026-03-30T10:25:51.173069-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/prepared_stateme_184.sdi to /datadir/performance_schema/prepared_stateme_184.sdi 2026-03-30T10:25:51.173086-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/prepared_stateme_184.sdi to /datadir/performance_schema/prepared_stateme_184.sdi 2026-03-30T10:25:51.173103-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_133.sdi to /datadir/performance_schema/events_statement_133.sdi 2026-03-30T10:25:51.173118-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_133.sdi to /datadir/performance_schema/events_statement_133.sdi 2026-03-30T10:25:51.173134-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_139.sdi to /datadir/performance_schema/events_statement_139.sdi 2026-03-30T10:25:51.173149-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_139.sdi to /datadir/performance_schema/events_statement_139.sdi 2026-03-30T10:25:51.173165-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_123.sdi to /datadir/performance_schema/events_stages_su_123.sdi 2026-03-30T10:25:51.173180-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_123.sdi to /datadir/performance_schema/events_stages_su_123.sdi 2026-03-30T10:25:51.173195-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_174.sdi to /datadir/performance_schema/replication_appl_174.sdi 2026-03-30T10:25:51.173210-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_174.sdi to /datadir/performance_schema/replication_appl_174.sdi 2026-03-30T10:25:51.173225-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_variable_194.sdi to /datadir/performance_schema/session_variable_194.sdi 2026-03-30T10:25:51.173239-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_variable_194.sdi to /datadir/performance_schema/session_variable_194.sdi 2026-03-30T10:25:51.173254-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_127.sdi to /datadir/performance_schema/events_stages_su_127.sdi 2026-03-30T10:25:51.173269-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_127.sdi to /datadir/performance_schema/events_stages_su_127.sdi 2026-03-30T10:25:51.173284-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_146.sdi to /datadir/performance_schema/events_transacti_146.sdi 2026-03-30T10:25:51.173299-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_146.sdi to /datadir/performance_schema/events_transacti_146.sdi 2026-03-30T10:25:51.173315-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_166.sdi to /datadir/performance_schema/memory_summary_b_166.sdi 2026-03-30T10:25:51.173329-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_166.sdi to /datadir/performance_schema/memory_summary_b_166.sdi 2026-03-30T10:25:51.173344-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_180.sdi to /datadir/performance_schema/replication_appl_180.sdi 2026-03-30T10:25:51.173359-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_180.sdi to /datadir/performance_schema/replication_appl_180.sdi 2026-03-30T10:25:51.173375-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_147.sdi to /datadir/performance_schema/events_transacti_147.sdi 2026-03-30T10:25:51.173392-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_147.sdi to /datadir/performance_schema/events_transacti_147.sdi 2026-03-30T10:25:51.173407-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_io_waits_s_116.sdi to /datadir/performance_schema/table_io_waits_s_116.sdi 2026-03-30T10:25:51.173423-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_io_waits_s_116.sdi to /datadir/performance_schema/table_io_waits_s_116.sdi 2026-03-30T10:25:51.173438-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_asyn_181.sdi to /datadir/performance_schema/replication_asyn_181.sdi 2026-03-30T10:25:51.173454-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_asyn_181.sdi to /datadir/performance_schema/replication_asyn_181.sdi 2026-03-30T10:25:51.173469-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/log_status_183.sdi to /datadir/performance_schema/log_status_183.sdi 2026-03-30T10:25:51.173484-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/log_status_183.sdi to /datadir/performance_schema/log_status_183.sdi 2026-03-30T10:25:51.173499-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/global_variables_193.sdi to /datadir/performance_schema/global_variables_193.sdi 2026-03-30T10:25:51.173514-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/global_variables_193.sdi to /datadir/performance_schema/global_variables_193.sdi 2026-03-30T10:25:51.173529-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_summary_by__104.sdi to /datadir/performance_schema/file_summary_by__104.sdi 2026-03-30T10:25:51.173544-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_summary_by__104.sdi to /datadir/performance_schema/file_summary_by__104.sdi 2026-03-30T10:25:51.173559-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_conn_173.sdi to /datadir/performance_schema/replication_conn_173.sdi 2026-03-30T10:25:51.173574-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_conn_173.sdi to /datadir/performance_schema/replication_conn_173.sdi 2026-03-30T10:25:51.173589-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_consumers_112.sdi to /datadir/performance_schema/setup_consumers_112.sdi 2026-03-30T10:25:51.173603-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_consumers_112.sdi to /datadir/performance_schema/setup_consumers_112.sdi 2026-03-30T10:25:51.173619-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_131.sdi to /datadir/performance_schema/events_statement_131.sdi 2026-03-30T10:25:51.173634-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_131.sdi to /datadir/performance_schema/events_statement_131.sdi 2026-03-30T10:25:51.173649-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_summary_b_158.sdi to /datadir/performance_schema/socket_summary_b_158.sdi 2026-03-30T10:25:51.173665-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_summary_b_158.sdi to /datadir/performance_schema/socket_summary_b_158.sdi 2026-03-30T10:25:51.173680-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_128.sdi to /datadir/performance_schema/events_statement_128.sdi 2026-03-30T10:25:51.173695-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_128.sdi to /datadir/performance_schema/events_statement_128.sdi 2026-03-30T10:25:51.173710-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_175.sdi to /datadir/performance_schema/replication_appl_175.sdi 2026-03-30T10:25:51.173781-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_175.sdi to /datadir/performance_schema/replication_appl_175.sdi 2026-03-30T10:25:51.173815-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_grou_172.sdi to /datadir/performance_schema/replication_grou_172.sdi 2026-03-30T10:25:51.173851-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_grou_172.sdi to /datadir/performance_schema/replication_grou_172.sdi 2026-03-30T10:25:51.173883-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_124.sdi to /datadir/performance_schema/events_stages_su_124.sdi 2026-03-30T10:25:51.173908-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_124.sdi to /datadir/performance_schema/events_stages_su_124.sdi 2026-03-30T10:25:51.173929-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/keyring_keys_161.sdi to /datadir/performance_schema/keyring_keys_161.sdi 2026-03-30T10:25:51.173956-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/keyring_keys_161.sdi to /datadir/performance_schema/keyring_keys_161.sdi 2026-03-30T10:25:51.173984-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/tls_channel_stat_199.sdi to /datadir/performance_schema/tls_channel_stat_199.sdi 2026-03-30T10:25:51.174004-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/tls_channel_stat_199.sdi to /datadir/performance_schema/tls_channel_stat_199.sdi 2026-03-30T10:25:51.174029-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_user_189.sdi to /datadir/performance_schema/status_by_user_189.sdi 2026-03-30T10:25:51.174086-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_user_189.sdi to /datadir/performance_schema/status_by_user_189.sdi 2026-03-30T10:25:51.174114-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_g_162.sdi to /datadir/performance_schema/memory_summary_g_162.sdi 2026-03-30T10:25:51.174140-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_g_162.sdi to /datadir/performance_schema/memory_summary_g_162.sdi 2026-03-30T10:25:51.174170-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_instrument_113.sdi to /datadir/performance_schema/setup_instrument_113.sdi 2026-03-30T10:25:51.174220-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_instrument_113.sdi to /datadir/performance_schema/setup_instrument_113.sdi 2026-03-30T10:25:51.174280-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_149.sdi to /datadir/performance_schema/events_errors_su_149.sdi 2026-03-30T10:25:51.174305-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_149.sdi to /datadir/performance_schema/events_errors_su_149.sdi 2026-03-30T10:25:51.174331-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_98.sdi to /datadir/performance_schema/events_waits_sum_98.sdi 2026-03-30T10:25:51.174381-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_98.sdi to /datadir/performance_schema/events_waits_sum_98.sdi 2026-03-30T10:25:51.174427-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/keyring_componen_202.sdi to /datadir/performance_schema/keyring_componen_202.sdi 2026-03-30T10:25:51.174456-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/keyring_componen_202.sdi to /datadir/performance_schema/keyring_componen_202.sdi 2026-03-30T10:25:51.174476-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_140.sdi to /datadir/performance_schema/events_transacti_140.sdi 2026-03-30T10:25:51.174495-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_140.sdi to /datadir/performance_schema/events_transacti_140.sdi 2026-03-30T10:25:51.174518-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_account__160.sdi to /datadir/performance_schema/session_account__160.sdi 2026-03-30T10:25:51.174535-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_account__160.sdi to /datadir/performance_schema/session_account__160.sdi 2026-03-30T10:25:51.174551-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_100.sdi to /datadir/performance_schema/events_waits_sum_100.sdi 2026-03-30T10:25:51.174567-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_100.sdi to /datadir/performance_schema/events_waits_sum_100.sdi 2026-03-30T10:25:51.174583-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_summary_by__103.sdi to /datadir/performance_schema/file_summary_by__103.sdi 2026-03-30T10:25:51.174598-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_summary_by__103.sdi to /datadir/performance_schema/file_summary_by__103.sdi 2026-03-30T10:25:51.174614-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_his_94.sdi to /datadir/performance_schema/events_waits_his_94.sdi 2026-03-30T10:25:51.174629-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_his_94.sdi to /datadir/performance_schema/events_waits_his_94.sdi 2026-03-30T10:25:51.174645-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_hi_121.sdi to /datadir/performance_schema/events_stages_hi_121.sdi 2026-03-30T10:25:51.174660-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_hi_121.sdi to /datadir/performance_schema/events_stages_hi_121.sdi 2026-03-30T10:25:51.174676-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_126.sdi to /datadir/performance_schema/events_stages_su_126.sdi 2026-03-30T10:25:51.174691-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_126.sdi to /datadir/performance_schema/events_stages_su_126.sdi 2026-03-30T10:25:51.174707-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_165.sdi to /datadir/performance_schema/memory_summary_b_165.sdi 2026-03-30T10:25:51.174747-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_165.sdi to /datadir/performance_schema/memory_summary_b_165.sdi 2026-03-30T10:25:51.174775-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_142.sdi to /datadir/performance_schema/events_transacti_142.sdi 2026-03-30T10:25:51.174823-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_142.sdi to /datadir/performance_schema/events_transacti_142.sdi 2026-03-30T10:25:51.174857-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_141.sdi to /datadir/performance_schema/events_transacti_141.sdi 2026-03-30T10:25:51.174889-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_141.sdi to /datadir/performance_schema/events_transacti_141.sdi 2026-03-30T10:25:51.174917-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_152.sdi to /datadir/performance_schema/events_errors_su_152.sdi 2026-03-30T10:25:51.174947-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_152.sdi to /datadir/performance_schema/events_errors_su_152.sdi 2026-03-30T10:25:51.174977-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_177.sdi to /datadir/performance_schema/replication_appl_177.sdi 2026-03-30T10:25:51.175004-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_177.sdi to /datadir/performance_schema/replication_appl_177.sdi 2026-03-30T10:25:51.175033-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/cond_instances_91.sdi to /datadir/performance_schema/cond_instances_91.sdi 2026-03-30T10:25:51.175102-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/cond_instances_91.sdi to /datadir/performance_schema/cond_instances_91.sdi 2026-03-30T10:25:51.175134-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/variables_info_195.sdi to /datadir/performance_schema/variables_info_195.sdi 2026-03-30T10:25:51.175160-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/variables_info_195.sdi to /datadir/performance_schema/variables_info_195.sdi 2026-03-30T10:25:51.175193-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_176.sdi to /datadir/performance_schema/replication_appl_176.sdi 2026-03-30T10:25:51.175224-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_176.sdi to /datadir/performance_schema/replication_appl_176.sdi 2026-03-30T10:25:51.175287-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_148.sdi to /datadir/performance_schema/events_errors_su_148.sdi 2026-03-30T10:25:51.175310-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_148.sdi to /datadir/performance_schema/events_errors_su_148.sdi 2026-03-30T10:25:51.175333-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/processlist_109.sdi to /datadir/performance_schema/processlist_109.sdi 2026-03-30T10:25:51.175360-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/processlist_109.sdi to /datadir/performance_schema/processlist_109.sdi 2026-03-30T10:25:51.175388-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/global_status_190.sdi to /datadir/performance_schema/global_status_190.sdi 2026-03-30T10:25:51.175412-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/global_status_190.sdi to /datadir/performance_schema/global_status_190.sdi 2026-03-30T10:25:51.175443-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/malloc_stats_201.sdi to /datadir/performance_schema/malloc_stats_201.sdi 2026-03-30T10:25:51.175468-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/malloc_stats_201.sdi to /datadir/performance_schema/malloc_stats_201.sdi 2026-03-30T10:25:51.175498-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_132.sdi to /datadir/performance_schema/events_statement_132.sdi 2026-03-30T10:25:51.175524-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_132.sdi to /datadir/performance_schema/events_statement_132.sdi 2026-03-30T10:25:51.175550-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/host_cache_105.sdi to /datadir/performance_schema/host_cache_105.sdi 2026-03-30T10:25:51.175574-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/host_cache_105.sdi to /datadir/performance_schema/host_cache_105.sdi 2026-03-30T10:25:51.175599-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/hosts_155.sdi to /datadir/performance_schema/hosts_155.sdi 2026-03-30T10:25:51.175618-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/hosts_155.sdi to /datadir/performance_schema/hosts_155.sdi 2026-03-30T10:25:51.175642-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_handles_167.sdi to /datadir/performance_schema/table_handles_167.sdi 2026-03-30T10:25:51.175668-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_handles_167.sdi to /datadir/performance_schema/table_handles_167.sdi 2026-03-30T10:25:51.175747-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_objects_114.sdi to /datadir/performance_schema/setup_objects_114.sdi 2026-03-30T10:25:51.175778-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_objects_114.sdi to /datadir/performance_schema/setup_objects_114.sdi 2026-03-30T10:25:51.175811-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_144.sdi to /datadir/performance_schema/events_transacti_144.sdi 2026-03-30T10:25:51.175841-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_144.sdi to /datadir/performance_schema/events_transacti_144.sdi 2026-03-30T10:25:51.175874-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/data_locks_169.sdi to /datadir/performance_schema/data_locks_169.sdi 2026-03-30T10:25:51.175902-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/data_locks_169.sdi to /datadir/performance_schema/data_locks_169.sdi 2026-03-30T10:25:51.175931-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/threads_119.sdi to /datadir/performance_schema/threads_119.sdi 2026-03-30T10:25:51.175958-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/threads_119.sdi to /datadir/performance_schema/threads_119.sdi 2026-03-30T10:25:51.175988-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_150.sdi to /datadir/performance_schema/events_errors_su_150.sdi 2026-03-30T10:25:51.176017-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_150.sdi to /datadir/performance_schema/events_errors_su_150.sdi 2026-03-30T10:25:51.176046-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/objects_summary__107.sdi to /datadir/performance_schema/objects_summary__107.sdi 2026-03-30T10:25:51.176074-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/objects_summary__107.sdi to /datadir/performance_schema/objects_summary__107.sdi 2026-03-30T10:25:51.176108-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/user_defined_fun_197.sdi to /datadir/performance_schema/user_defined_fun_197.sdi 2026-03-30T10:25:51.176136-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/user_defined_fun_197.sdi to /datadir/performance_schema/user_defined_fun_197.sdi 2026-03-30T10:25:51.176167-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_conn_171.sdi to /datadir/performance_schema/replication_conn_171.sdi 2026-03-30T10:25:51.176195-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_conn_171.sdi to /datadir/performance_schema/replication_conn_171.sdi 2026-03-30T10:25:51.176225-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_cu_120.sdi to /datadir/performance_schema/events_stages_cu_120.sdi 2026-03-30T10:25:51.176252-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_cu_120.sdi to /datadir/performance_schema/events_stages_cu_120.sdi 2026-03-30T10:25:51.176282-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_134.sdi to /datadir/performance_schema/events_statement_134.sdi 2026-03-30T10:25:51.176310-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_134.sdi to /datadir/performance_schema/events_statement_134.sdi 2026-03-30T10:25:51.176342-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_135.sdi to /datadir/performance_schema/events_statement_135.sdi 2026-03-30T10:25:51.176371-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_135.sdi to /datadir/performance_schema/events_statement_135.sdi 2026-03-30T10:25:51.176401-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_138.sdi to /datadir/performance_schema/events_statement_138.sdi 2026-03-30T10:25:51.176429-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_138.sdi to /datadir/performance_schema/events_statement_138.sdi 2026-03-30T10:25:51.176458-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_lock_waits_118.sdi to /datadir/performance_schema/table_lock_waits_118.sdi 2026-03-30T10:25:51.176489-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_lock_waits_118.sdi to /datadir/performance_schema/table_lock_waits_118.sdi 2026-03-30T10:25:51.176546-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/data_lock_waits_170.sdi to /datadir/performance_schema/data_lock_waits_170.sdi 2026-03-30T10:25:51.176571-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/data_lock_waits_170.sdi to /datadir/performance_schema/data_lock_waits_170.sdi 2026-03-30T10:25:51.176598-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_hi_122.sdi to /datadir/performance_schema/events_stages_hi_122.sdi 2026-03-30T10:25:51.176624-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_hi_122.sdi to /datadir/performance_schema/events_stages_hi_122.sdi 2026-03-30T10:25:51.176651-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/performance_time_108.sdi to /datadir/performance_schema/performance_time_108.sdi 2026-03-30T10:25:51.176682-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/performance_time_108.sdi to /datadir/performance_schema/performance_time_108.sdi 2026-03-30T10:25:51.176714-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_164.sdi to /datadir/performance_schema/memory_summary_b_164.sdi 2026-03-30T10:25:51.176756-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_164.sdi to /datadir/performance_schema/memory_summary_b_164.sdi 2026-03-30T10:25:51.176785-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_summary_b_157.sdi to /datadir/performance_schema/socket_summary_b_157.sdi 2026-03-30T10:25:51.176815-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_summary_b_157.sdi to /datadir/performance_schema/socket_summary_b_157.sdi 2026-03-30T10:25:51.176843-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/malloc_stats_tot_200.sdi to /datadir/performance_schema/malloc_stats_tot_200.sdi 2026-03-30T10:25:51.176876-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/malloc_stats_tot_200.sdi to /datadir/performance_schema/malloc_stats_tot_200.sdi 2026-03-30T10:25:51.176904-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_io_waits_s_117.sdi to /datadir/performance_schema/table_io_waits_s_117.sdi 2026-03-30T10:25:51.176929-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_io_waits_s_117.sdi to /datadir/performance_schema/table_io_waits_s_117.sdi 2026-03-30T10:25:51.176959-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_his_95.sdi to /datadir/performance_schema/events_waits_his_95.sdi 2026-03-30T10:25:51.177012-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_his_95.sdi to /datadir/performance_schema/events_waits_his_95.sdi 2026-03-30T10:25:51.177036-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_threads_115.sdi to /datadir/performance_schema/setup_threads_115.sdi 2026-03-30T10:25:51.177055-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_threads_115.sdi to /datadir/performance_schema/setup_threads_115.sdi 2026-03-30T10:25:51.177112-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/users_153.sdi to /datadir/performance_schema/users_153.sdi 2026-03-30T10:25:51.177134-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/users_153.sdi to /datadir/performance_schema/users_153.sdi 2026-03-30T10:25:51.177155-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_163.sdi to /datadir/performance_schema/memory_summary_b_163.sdi 2026-03-30T10:25:51.177184-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_163.sdi to /datadir/performance_schema/memory_summary_b_163.sdi 2026-03-30T10:25:51.177214-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/binary_log_trans_198.sdi to /datadir/performance_schema/binary_log_trans_198.sdi 2026-03-30T10:25:51.177239-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/binary_log_trans_198.sdi to /datadir/performance_schema/binary_log_trans_198.sdi 2026-03-30T10:25:51.177291-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_instances_156.sdi to /datadir/performance_schema/socket_instances_156.sdi 2026-03-30T10:25:51.177324-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_instances_156.sdi to /datadir/performance_schema/socket_instances_156.sdi 2026-03-30T10:25:51.177352-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_thread_188.sdi to /datadir/performance_schema/status_by_thread_188.sdi 2026-03-30T10:25:51.177373-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_thread_188.sdi to /datadir/performance_schema/status_by_thread_188.sdi 2026-03-30T10:25:51.177395-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_accoun_186.sdi to /datadir/performance_schema/status_by_accoun_186.sdi 2026-03-30T10:25:51.177421-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_accoun_186.sdi to /datadir/performance_schema/status_by_accoun_186.sdi 2026-03-30T10:25:51.177451-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_96.sdi to /datadir/performance_schema/events_waits_sum_96.sdi 2026-03-30T10:25:51.177480-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_96.sdi to /datadir/performance_schema/events_waits_sum_96.sdi 2026-03-30T10:25:51.177536-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_99.sdi to /datadir/performance_schema/events_waits_sum_99.sdi 2026-03-30T10:25:51.177586-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_99.sdi to /datadir/performance_schema/events_waits_sum_99.sdi 2026-03-30T10:25:51.177612-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/mutex_instances_106.sdi to /datadir/performance_schema/mutex_instances_106.sdi 2026-03-30T10:25:51.177633-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/mutex_instances_106.sdi to /datadir/performance_schema/mutex_instances_106.sdi 2026-03-30T10:25:51.177662-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/user_variables_b_185.sdi to /datadir/performance_schema/user_variables_b_185.sdi 2026-03-30T10:25:51.177693-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/user_variables_b_185.sdi to /datadir/performance_schema/user_variables_b_185.sdi 2026-03-30T10:25:51.177716-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_connect__159.sdi to /datadir/performance_schema/session_connect__159.sdi 2026-03-30T10:25:51.177762-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_connect__159.sdi to /datadir/performance_schema/session_connect__159.sdi 2026-03-30T10:25:51.177792-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/error_log_92.sdi to /datadir/performance_schema/error_log_92.sdi 2026-03-30T10:25:51.177822-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/error_log_92.sdi to /datadir/performance_schema/error_log_92.sdi 2026-03-30T10:25:51.177857-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_179.sdi to /datadir/performance_schema/replication_appl_179.sdi 2026-03-30T10:25:51.177887-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_179.sdi to /datadir/performance_schema/replication_appl_179.sdi 2026-03-30T10:25:51.177919-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_status_191.sdi to /datadir/performance_schema/session_status_191.sdi 2026-03-30T10:25:51.177942-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_status_191.sdi to /datadir/performance_schema/session_status_191.sdi 2026-03-30T10:25:51.177963-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_143.sdi to /datadir/performance_schema/events_transacti_143.sdi 2026-03-30T10:25:51.177994-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_143.sdi to /datadir/performance_schema/events_transacti_143.sdi 2026-03-30T10:25:51.178026-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_actors_111.sdi to /datadir/performance_schema/setup_actors_111.sdi 2026-03-30T10:25:51.178056-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_actors_111.sdi to /datadir/performance_schema/setup_actors_111.sdi 2026-03-30T10:25:51.178090-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/persisted_variab_196.sdi to /datadir/performance_schema/persisted_variab_196.sdi 2026-03-30T10:25:51.178120-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/persisted_variab_196.sdi to /datadir/performance_schema/persisted_variab_196.sdi 2026-03-30T10:25:51.178148-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_instances_102.sdi to /datadir/performance_schema/file_instances_102.sdi 2026-03-30T10:25:51.178200-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_instances_102.sdi to /datadir/performance_schema/file_instances_102.sdi 2026-03-30T10:25:51.178235-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_asyn_182.sdi to /datadir/performance_schema/replication_asyn_182.sdi 2026-03-30T10:25:51.178271-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_asyn_182.sdi to /datadir/performance_schema/replication_asyn_182.sdi 2026-03-30T10:25:51.178307-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_129.sdi to /datadir/performance_schema/events_statement_129.sdi 2026-03-30T10:25:51.178365-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_129.sdi to /datadir/performance_schema/events_statement_129.sdi 2026-03-30T10:25:51.178395-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/pxc_cluster_view_203.sdi to /datadir/performance_schema/pxc_cluster_view_203.sdi 2026-03-30T10:25:51.178450-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/pxc_cluster_view_203.sdi to /datadir/performance_schema/pxc_cluster_view_203.sdi 2026-03-30T10:25:51.178477-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_130.sdi to /datadir/performance_schema/events_statement_130.sdi 2026-03-30T10:25:51.178504-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_130.sdi to /datadir/performance_schema/events_statement_130.sdi 2026-03-30T10:25:51.178532-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./xtrabackup_info to /datadir/xtrabackup_info 2026-03-30T10:25:51.178559-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./xtrabackup_info to /datadir/xtrabackup_info 2026-03-30T10:25:51.178644-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./sys/sys_config.ibd to /datadir/sys/sys_config.ibd 2026-03-30T10:25:51.178677-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./sys/sys_config.ibd to /datadir/sys/sys_config.ibd 2026-03-30T10:25:51.272518-00:00 0 [Note] [MY-011825] [Xtrabackup] completed OK! + cat /tmp/tmp.WF3JEsS2bA Defaulted container "xtrabackup" out of: xtrabackup, backup-init (init) + rm /tmp/tmp.fkyakqzZlF /tmp/tmp.WF3JEsS2bA + return 0 + wait_for_running demand-backup-proxysql 1 + local name=demand-backup-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod demand-backup-proxysql-0 480 + local pod=demand-backup-proxysql-0 + local max_retry=480 + local ns= ++ echo demand-backup-proxysql-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=proxysql + set +o xtrace pod/demand-backup-proxysql-0 condition met waiting for pod/demand-backup-proxysql-0 to become Ready.Ok + wait_for_running demand-backup-pxc 3 + local name=demand-backup-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod demand-backup-pxc-0 480 + local pod=demand-backup-pxc-0 + local max_retry=480 + local ns= ++ echo demand-backup-pxc-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/demand-backup-pxc-0 condition met waiting for pod/demand-backup-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod demand-backup-pxc-1 480 + local pod=demand-backup-pxc-1 + local max_retry=480 + local ns= ++ echo demand-backup-pxc-1 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/demand-backup-pxc-1 condition met waiting for pod/demand-backup-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod demand-backup-pxc-2 480 + local pod=demand-backup-pxc-2 + local max_retry=480 + local ns= ++ echo demand-backup-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/demand-backup-pxc-2 condition met waiting for pod/demand-backup-pxc-2 to become Ready.Ok ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.spec.secretsName}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.B8zsFsuWkP +++ mktemp ++ local LAST_ERR=/tmp/tmp.RWoPxOSQLR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.spec.secretsName}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.B8zsFsuWkP ++ cat /tmp/tmp.RWoPxOSQLR ++ rm /tmp/tmp.B8zsFsuWkP /tmp/tmp.RWoPxOSQLR ++ return 0 + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.0FUg7QN7VA +++ mktemp ++ local LAST_ERR=/tmp/tmp.MtMY1rUxBZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.0FUg7QN7VA ++ cat /tmp/tmp.MtMY1rUxBZ ++ rm /tmp/tmp.0FUg7QN7VA /tmp/tmp.MtMY1rUxBZ ++ return 0 + local root_pass=root_password + sleep 35 + log 'check data after pxc-restore/on-demand-backup-pvc-bsource' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-03-30T10:31:22+0000]' check data after pxc-restore/on-demand-backup-pvc-bsource [2026-03-30T10:31:22+0000] check data after pxc-restore/on-demand-backup-pvc-bsource + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-0.demand-backup-pxc -uroot -p'\''root_password'\''' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-0.demand-backup-pxc -uroot -p'\''root_password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-0.demand-backup-pxc -uroot -p'\''root_password'\''' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-0.demand-backup-pxc -uroot -p'\''root_password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xN6uLnE8o5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.n0AJSD1jDm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.xN6uLnE8o5 ++ cat /tmp/tmp.n0AJSD1jDm ++ rm /tmp/tmp.xN6uLnE8o5 /tmp/tmp.n0AJSD1jDm ++ return 0 + client_pod=pxc-client-67fc4995bb-jwx2s + wait_pod pxc-client-67fc4995bb-jwx2s + local pod=pxc-client-67fc4995bb-jwx2s + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-67fc4995bb-jwx2s ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-jwx2s condition met waiting for pod/pxc-client-67fc4995bb-jwx2s to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.14SIjS2s6f/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.14SIjS2s6f/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1.sql /tmp/tmp.14SIjS2s6f/select-1.sql + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-1.demand-backup-pxc -uroot -p'\''root_password'\''' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-1.demand-backup-pxc -uroot -p'\''root_password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-1.demand-backup-pxc -uroot -p'\''root_password'\''' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-1.demand-backup-pxc -uroot -p'\''root_password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lwKXFGUSpD +++ mktemp ++ local LAST_ERR=/tmp/tmp.jr8Cm2DMZj ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.lwKXFGUSpD ++ cat /tmp/tmp.jr8Cm2DMZj ++ rm /tmp/tmp.lwKXFGUSpD /tmp/tmp.jr8Cm2DMZj ++ return 0 + client_pod=pxc-client-67fc4995bb-jwx2s + wait_pod pxc-client-67fc4995bb-jwx2s + local pod=pxc-client-67fc4995bb-jwx2s + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-jwx2s ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-jwx2s condition met waiting for pod/pxc-client-67fc4995bb-jwx2s to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.14SIjS2s6f/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.14SIjS2s6f/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1.sql /tmp/tmp.14SIjS2s6f/select-1.sql + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-2.demand-backup-pxc -uroot -p'\''root_password'\''' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-2.demand-backup-pxc -uroot -p'\''root_password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-2.demand-backup-pxc -uroot -p'\''root_password'\''' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-2.demand-backup-pxc -uroot -p'\''root_password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.f1GFinDoTv +++ mktemp ++ local LAST_ERR=/tmp/tmp.84K28k9liR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.f1GFinDoTv ++ cat /tmp/tmp.84K28k9liR ++ rm /tmp/tmp.f1GFinDoTv /tmp/tmp.84K28k9liR ++ return 0 + client_pod=pxc-client-67fc4995bb-jwx2s + wait_pod pxc-client-67fc4995bb-jwx2s + local pod=pxc-client-67fc4995bb-jwx2s + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-jwx2s ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-jwx2s condition met waiting for pod/pxc-client-67fc4995bb-jwx2s to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.14SIjS2s6f/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.14SIjS2s6f/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/select-1.sql /tmp/tmp.14SIjS2s6f/select-1.sql + '[' on-demand-backup-pvc-bsource '!=' on-demand-backup-minio ']' + log 'copy backup' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-03-30T10:31:37+0000]' copy backup [2026-03-30T10:31:37+0000] copy backup + '[' -n '' ']' + bash /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/deploy/backup/copy-backup.sh on-demand-backup-pvc-bsource /tmp/tmp.14SIjS2s6f/backup Log: /tmp/tmp.SqQwralgCm/log pvc/xb-on-demand-backup-pvc-bsource-20260330102335-42dfe747pod/backup-access created Starting pod.......[done] Downloading started tar: Removing leading `/' from member names Downloading finished pod "backup-access" deleted from demand-backup-28441 namespace You can recover data locally with following commands: $ service mysqld stop $ rm -rf /var/lib/mysql/* $ cat /tmp/tmp.14SIjS2s6f/backup/xtrabackup.stream | xbstream --decompress -x -C /var/lib/mysql $ xtrabackup --prepare --target-dir=/var/lib/mysql $ chown -R mysql:mysql /var/lib/mysql $ service mysqld start + compare_kubectl job/restore-job-on-demand-backup-pvc-bsource-demand-backup + local resource=job/restore-job-on-demand-backup-pvc-bsource-demand-backup + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_restore-job-on-demand-backup-pvc-bsource-demand-backup.yml + local new_result=/tmp/tmp.14SIjS2s6f/job_restore-job-on-demand-backup-pvc-bsource-demand-backup.yml + desc 'compare job/restore-job-on-demand-backup-pvc-bsource-demand-backup-' + set +o xtrace ----------------------------------------------------------------------------------- compare job/restore-job-on-demand-backup-pvc-bsource-demand-backup- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_restore-job-on-demand-backup-pvc-bsource-demand-backup-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_restore-job-on-demand-backup-pvc-bsource-demand-backup-80.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.32 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.32 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_restore-job-on-demand-backup-pvc-bsource-demand-backup-k129.yml ']' + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_restore-job-on-demand-backup-pvc-bsource-demand-backup-k129.yml + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_restore-job-on-demand-backup-pvc-bsource-demand-backup-k129-oc.yml ']' + version_gt 1.29 ++ echo '1.32 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_restore-job-on-demand-backup-pvc-bsource-demand-backup-k129-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_restore-job-on-demand-backup-pvc-bsource-demand-backup-k129-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_restore-job-on-demand-backup-pvc-bsource-demand-backup-k129-aks.yml ']' + kubectl_bin get -o yaml job/restore-job-on-demand-backup-pvc-bsource-demand-backup + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.spec.template.spec.containers[].env[] | select(.name == "XTRABACKUP_ENABLED")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.metadata.annotations."kubernetes.digitalocean.com/load-balancer-id") | del(.metadata.annotations."service.beta.kubernetes.io/do-loadbalancer-type") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("demand-backup-28441", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - ++ mktemp + local LAST_OUT=/tmp/tmp.PefaNRKYGC ++ mktemp + local LAST_ERR=/tmp/tmp.PSCTUScD1R + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml job/restore-job-on-demand-backup-pvc-bsource-demand-backup + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.PefaNRKYGC + cat /tmp/tmp.PSCTUScD1R + rm /tmp/tmp.PefaNRKYGC /tmp/tmp.PSCTUScD1R + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_restore-job-on-demand-backup-pvc-bsource-demand-backup-k129.yml /tmp/tmp.14SIjS2s6f/job_restore-job-on-demand-backup-pvc-bsource-demand-backup.yml + log 'compare_kubectl: job/restore-job-on-demand-backup-pvc-bsource-demand-backup OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-03-30T10:32:32+0000]' compare_kubectl: job/restore-job-on-demand-backup-pvc-bsource-demand-backup OK [2026-03-30T10:32:32+0000] compare_kubectl: job/restore-job-on-demand-backup-pvc-bsource-demand-backup OK + minio_backup_name=on-demand-backup-minio + wrong_creds_backup_name=on-demand-backup-wrong-creds + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/on-demand-backup-wrong-creds.yml ++ mktemp + local LAST_OUT=/tmp/tmp.oq5zO7xKjy ++ mktemp + local LAST_ERR=/tmp/tmp.WezlpnnGs2 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/on-demand-backup-wrong-creds.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.oq5zO7xKjy perconaxtradbclusterbackup.pxc.percona.com/on-demand-backup-wrong-creds created + cat /tmp/tmp.WezlpnnGs2 + rm /tmp/tmp.oq5zO7xKjy /tmp/tmp.WezlpnnGs2 + return 0 + run_backup demand-backup on-demand-backup-minio + local cluster=demand-backup + local backup=on-demand-backup-minio + log 'run pxc-backup/on-demand-backup-minio' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-03-30T10:32:33+0000]' run pxc-backup/on-demand-backup-minio [2026-03-30T10:32:33+0000] run pxc-backup/on-demand-backup-minio + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/on-demand-backup-minio.yml ++ mktemp + local LAST_OUT=/tmp/tmp.sBIBW3gpcr ++ mktemp + local LAST_ERR=/tmp/tmp.rZ0qvaBn7v + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/on-demand-backup-minio.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.sBIBW3gpcr perconaxtradbclusterbackup.pxc.percona.com/on-demand-backup-minio created + cat /tmp/tmp.rZ0qvaBn7v + rm /tmp/tmp.sBIBW3gpcr /tmp/tmp.rZ0qvaBn7v + return 0 + wait_backup on-demand-backup-minio + local backup=on-demand-backup-minio + local status=Succeeded + set +o xtrace waiting for pxc-backup/on-demand-backup-minio to reach Succeeded state..............................................................................................Succeeded + compare_kubectl job/xb-on-demand-backup-minio + local resource=job/xb-on-demand-backup-minio + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_xb-on-demand-backup-minio.yml + local new_result=/tmp/tmp.14SIjS2s6f/job_xb-on-demand-backup-minio.yml + desc 'compare job/xb-on-demand-backup-minio-' + set +o xtrace ----------------------------------------------------------------------------------- compare job/xb-on-demand-backup-minio- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_xb-on-demand-backup-minio-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_xb-on-demand-backup-minio-80.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.32 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.32 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_xb-on-demand-backup-minio-k129.yml ']' + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_xb-on-demand-backup-minio-k129.yml + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_xb-on-demand-backup-minio-k129-oc.yml ']' + version_gt 1.29 ++ echo '1.32 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_xb-on-demand-backup-minio-k129-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_xb-on-demand-backup-minio-k129-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_xb-on-demand-backup-minio-k129-aks.yml ']' + kubectl_bin get -o yaml job/xb-on-demand-backup-minio ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.spec.template.spec.containers[].env[] | select(.name == "XTRABACKUP_ENABLED")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.metadata.annotations."kubernetes.digitalocean.com/load-balancer-id") | del(.metadata.annotations."service.beta.kubernetes.io/do-loadbalancer-type") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("demand-backup-28441", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.0oHMLoBdpg ++ mktemp + local LAST_ERR=/tmp/tmp.8naekoUfRV + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml job/xb-on-demand-backup-minio + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.0oHMLoBdpg + cat /tmp/tmp.8naekoUfRV + rm /tmp/tmp.0oHMLoBdpg /tmp/tmp.8naekoUfRV + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/job_xb-on-demand-backup-minio-k129.yml /tmp/tmp.14SIjS2s6f/job_xb-on-demand-backup-minio.yml + log 'compare_kubectl: job/xb-on-demand-backup-minio OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-03-30T10:35:03+0000]' compare_kubectl: job/xb-on-demand-backup-minio OK [2026-03-30T10:35:03+0000] compare_kubectl: job/xb-on-demand-backup-minio OK + sleep 90 ++ kubectl_bin get job xb-on-demand-backup-wrong-creds -o yaml ++ yq '.status.conditions[] | select(.type == "Failed").lastTransitionTime' ++ xargs -I '{}' /usr/bin/date -d '{}' +%s +++ mktemp ++ local LAST_OUT=/tmp/tmp.8lLrYnahFE +++ mktemp ++ local LAST_ERR=/tmp/tmp.pa7RqgAgHV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get job xb-on-demand-backup-wrong-creds -o yaml ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8lLrYnahFE ++ cat /tmp/tmp.pa7RqgAgHV ++ rm /tmp/tmp.8lLrYnahFE /tmp/tmp.pa7RqgAgHV ++ return 0 + wrong_backup_job_fail_timestamp=1774866856 + [[ -z 1774866856 ]] ++ yq .metadata.creationTimestamp ++ kubectl get job xb-on-demand-backup-minio -o yaml ++ xargs -I '{}' /usr/bin/date -d '{}' +%s + minio_job_creation_timestamp=1774866866 + [[ -z 1774866866 ]] + [[ 1774866856 > 1774866866 ]] ++ kubectl get pod -l percona.com/backup-job-name=xb-on-demand-backup-minio -o 'jsonpath={.items[].metadata.name}' + backup_job_name=xb-on-demand-backup-minio-lnqf7 + check_logs xb-on-demand-backup-minio-lnqf7 'xbcloud put --storage=s3 --md5 --parallel=[0-9]+ --insecure (--curl-retriable-errors=7 )?--parallel=2 --curl-retriable-errors=8' + local restore=xb-on-demand-backup-minio-lnqf7 + local 'search=xbcloud put --storage=s3 --md5 --parallel=[0-9]+ --insecure (--curl-retriable-errors=7 )?--parallel=2 --curl-retriable-errors=8' + kubectl logs xb-on-demand-backup-minio-lnqf7 + grep -E 'xbcloud put --storage=s3 --md5 --parallel=[0-9]+ --insecure (--curl-retriable-errors=7 )?--parallel=2 --curl-retriable-errors=8' Defaulted container "xtrabackup" out of: xtrabackup, backup-init (init) 2026-03-30 10:34:38.756 INFO: [SST script] + xbcloud put --storage=s3 --md5 --parallel=4 --insecure --curl-retriable-errors=7 --parallel=2 --curl-retriable-errors=8 --s3-bucket=operator-testing prefix/subfolder/demand-backup-2026-03-30-10:32:35-full.sst_info 2026-03-30 10:34:39.113 INFO: [SST script] + xbcloud put --storage=s3 --md5 --parallel=4 --insecure --curl-retriable-errors=7 --parallel=2 --curl-retriable-errors=8 --s3-bucket=operator-testing prefix/subfolder/demand-backup-2026-03-30-10:32:35-full + check_logs xb-on-demand-backup-minio-lnqf7 'xbstream -x -C /tmp --parallel=2' + local restore=xb-on-demand-backup-minio-lnqf7 + local 'search=xbstream -x -C /tmp --parallel=2' + grep -E 'xbstream -x -C /tmp --parallel=2' + kubectl logs xb-on-demand-backup-minio-lnqf7 Defaulted container "xtrabackup" out of: xtrabackup, backup-init (init) 2026-03-30 10:34:38.283 INFO: [SST script] + xbstream -x -C /tmp --parallel=2 ++ echo -n newpass ++ base64 + new_pass=bmV3cGFzcw== + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "bmV3cGFzcw==", "monitor": "bmV3cGFzcw=="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.6g7g2RHRWE ++ mktemp + local LAST_ERR=/tmp/tmp.ulUqdSHhg9 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "bmV3cGFzcw==", "monitor": "bmV3cGFzcw=="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.6g7g2RHRWE secret/my-cluster-secrets patched + cat /tmp/tmp.ulUqdSHhg9 + rm /tmp/tmp.6g7g2RHRWE /tmp/tmp.ulUqdSHhg9 + return 0 + wait_cluster_consistency demand-backup 3 2 + local cluster_name=demand-backup + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/demand-backup to be ready' waiting for pxc/demand-backup to be ready++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9KUYNrv1S3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.gnZYEKxIj7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9KUYNrv1S3 ++ cat /tmp/tmp.gnZYEKxIj7 ++ rm /tmp/tmp.9KUYNrv1S3 /tmp/tmp.gnZYEKxIj7 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NFd3had6Hf +++ mktemp ++ local LAST_ERR=/tmp/tmp.lc3Fn9N6GW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NFd3had6Hf ++ cat /tmp/tmp.lc3Fn9N6GW ++ rm /tmp/tmp.NFd3had6Hf /tmp/tmp.lc3Fn9N6GW ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LpnEoPAKma +++ mktemp ++ local LAST_ERR=/tmp/tmp.uV1SooqIzw ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.LpnEoPAKma ++ cat /tmp/tmp.uV1SooqIzw ++ rm /tmp/tmp.LpnEoPAKma /tmp/tmp.uV1SooqIzw ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.I71BW4HzGE +++ mktemp ++ local LAST_ERR=/tmp/tmp.tWur1PIkd6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.I71BW4HzGE ++ cat /tmp/tmp.tWur1PIkd6 ++ rm /tmp/tmp.I71BW4HzGE /tmp/tmp.tWur1PIkd6 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AKcc4URQdp +++ mktemp ++ local LAST_ERR=/tmp/tmp.MRooOvWld1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.AKcc4URQdp ++ cat /tmp/tmp.MRooOvWld1 ++ rm /tmp/tmp.AKcc4URQdp /tmp/tmp.MRooOvWld1 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lVrpUVFM3B +++ mktemp ++ local LAST_ERR=/tmp/tmp.5xCfdpGv0v ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.lVrpUVFM3B ++ cat /tmp/tmp.5xCfdpGv0v ++ rm /tmp/tmp.lVrpUVFM3B /tmp/tmp.5xCfdpGv0v ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.z5gqFO1Txp +++ mktemp ++ local LAST_ERR=/tmp/tmp.uIgGR5XhdA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.z5gqFO1Txp ++ cat /tmp/tmp.uIgGR5XhdA ++ rm /tmp/tmp.z5gqFO1Txp /tmp/tmp.uIgGR5XhdA ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.v0a7jWlt7F +++ mktemp ++ local LAST_ERR=/tmp/tmp.MsnpQLq1Lt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.v0a7jWlt7F ++ cat /tmp/tmp.MsnpQLq1Lt ++ rm /tmp/tmp.v0a7jWlt7F /tmp/tmp.MsnpQLq1Lt ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6ygDM4Bfh2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.K3U1X4bcSh ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6ygDM4Bfh2 ++ cat /tmp/tmp.K3U1X4bcSh ++ rm /tmp/tmp.6ygDM4Bfh2 /tmp/tmp.K3U1X4bcSh ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2tKkXMtfZn +++ mktemp ++ local LAST_ERR=/tmp/tmp.BHFdHr27gr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2tKkXMtfZn ++ cat /tmp/tmp.BHFdHr27gr ++ rm /tmp/tmp.2tKkXMtfZn /tmp/tmp.BHFdHr27gr ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bvA2v16uHZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.x8SQfcoUyF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bvA2v16uHZ ++ cat /tmp/tmp.x8SQfcoUyF ++ rm /tmp/tmp.bvA2v16uHZ /tmp/tmp.x8SQfcoUyF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WaUvV91ocS +++ mktemp ++ local LAST_ERR=/tmp/tmp.YqJa6kpo9g ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.WaUvV91ocS ++ cat /tmp/tmp.YqJa6kpo9g ++ rm /tmp/tmp.WaUvV91ocS /tmp/tmp.YqJa6kpo9g ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wBbbRqpz6K +++ mktemp ++ local LAST_ERR=/tmp/tmp.s3ovXTeCVR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.wBbbRqpz6K ++ cat /tmp/tmp.s3ovXTeCVR ++ rm /tmp/tmp.wBbbRqpz6K /tmp/tmp.s3ovXTeCVR ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.aL5A8qmPtc +++ mktemp ++ local LAST_ERR=/tmp/tmp.v13gqgQaJd ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.aL5A8qmPtc ++ cat /tmp/tmp.v13gqgQaJd ++ rm /tmp/tmp.aL5A8qmPtc /tmp/tmp.v13gqgQaJd ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.f9UyHqqFWi +++ mktemp ++ local LAST_ERR=/tmp/tmp.9gDUTccMev ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.f9UyHqqFWi ++ cat /tmp/tmp.9gDUTccMev ++ rm /tmp/tmp.f9UyHqqFWi /tmp/tmp.9gDUTccMev ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MT6bP7hw9a +++ mktemp ++ local LAST_ERR=/tmp/tmp.IBdSqJNnSY ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MT6bP7hw9a ++ cat /tmp/tmp.IBdSqJNnSY ++ rm /tmp/tmp.MT6bP7hw9a /tmp/tmp.IBdSqJNnSY ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.F7JvrZd2ac +++ mktemp ++ local LAST_ERR=/tmp/tmp.lqRNdybyOP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.F7JvrZd2ac ++ cat /tmp/tmp.lqRNdybyOP ++ rm /tmp/tmp.F7JvrZd2ac /tmp/tmp.lqRNdybyOP ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BJLI2N4iiH +++ mktemp ++ local LAST_ERR=/tmp/tmp.edaJKc4vQa ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.BJLI2N4iiH ++ cat /tmp/tmp.edaJKc4vQa ++ rm /tmp/tmp.BJLI2N4iiH /tmp/tmp.edaJKc4vQa ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.faJMr0QX38 +++ mktemp ++ local LAST_ERR=/tmp/tmp.DfV9RYHgGP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.faJMr0QX38 ++ cat /tmp/tmp.DfV9RYHgGP ++ rm /tmp/tmp.faJMr0QX38 /tmp/tmp.DfV9RYHgGP ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RlKwIblZdE +++ mktemp ++ local LAST_ERR=/tmp/tmp.Dnfa1W9Fu9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.RlKwIblZdE ++ cat /tmp/tmp.Dnfa1W9Fu9 ++ rm /tmp/tmp.RlKwIblZdE /tmp/tmp.Dnfa1W9Fu9 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6jE3svd7Yj +++ mktemp ++ local LAST_ERR=/tmp/tmp.5oF3kQrPRH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6jE3svd7Yj ++ cat /tmp/tmp.5oF3kQrPRH ++ rm /tmp/tmp.6jE3svd7Yj /tmp/tmp.5oF3kQrPRH ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fqzCCO5SDu +++ mktemp ++ local LAST_ERR=/tmp/tmp.9zfApJfLJh ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.fqzCCO5SDu ++ cat /tmp/tmp.9zfApJfLJh ++ rm /tmp/tmp.fqzCCO5SDu /tmp/tmp.9zfApJfLJh ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.a8py2Cjt2n +++ mktemp ++ local LAST_ERR=/tmp/tmp.cG9LFbf2me ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.a8py2Cjt2n ++ cat /tmp/tmp.cG9LFbf2me ++ rm /tmp/tmp.a8py2Cjt2n /tmp/tmp.cG9LFbf2me ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hDRp1ZWl5t +++ mktemp ++ local LAST_ERR=/tmp/tmp.lygCRXSUqQ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.hDRp1ZWl5t ++ cat /tmp/tmp.lygCRXSUqQ ++ rm /tmp/tmp.hDRp1ZWl5t /tmp/tmp.lygCRXSUqQ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.v2WCunQx9O +++ mktemp ++ local LAST_ERR=/tmp/tmp.qYbswucvCA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.v2WCunQx9O ++ cat /tmp/tmp.qYbswucvCA ++ rm /tmp/tmp.v2WCunQx9O /tmp/tmp.qYbswucvCA ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 24 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3vt896m5Dx +++ mktemp ++ local LAST_ERR=/tmp/tmp.e51Z034C2r ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.3vt896m5Dx ++ cat /tmp/tmp.e51Z034C2r ++ rm /tmp/tmp.3vt896m5Dx /tmp/tmp.e51Z034C2r ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 25 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EUxgo8IV5A +++ mktemp ++ local LAST_ERR=/tmp/tmp.RFRjXObuQX ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.EUxgo8IV5A ++ cat /tmp/tmp.RFRjXObuQX ++ rm /tmp/tmp.EUxgo8IV5A /tmp/tmp.RFRjXObuQX ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bjLjZT0Amp +++ mktemp ++ local LAST_ERR=/tmp/tmp.rwj4MIQNlb ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bjLjZT0Amp ++ cat /tmp/tmp.rwj4MIQNlb ++ rm /tmp/tmp.bjLjZT0Amp /tmp/tmp.rwj4MIQNlb ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine demand-backup +++ local cluster_name=demand-backup ++++ get_proxy demand-backup ++++ local target_cluster=demand-backup +++++ kubectl_bin get pxc demand-backup -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.PfSEbvAb9u ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.UUkz1YYpbo +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc demand-backup -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.PfSEbvAb9u +++++ cat /tmp/tmp.UUkz1YYpbo +++++ rm /tmp/tmp.PfSEbvAb9u /tmp/tmp.UUkz1YYpbo +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc demand-backup -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.PuPa7kvaS5 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.zadE7jDyBW +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc demand-backup -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.PuPa7kvaS5 +++++ cat /tmp/tmp.zadE7jDyBW +++++ rm /tmp/tmp.PuPa7kvaS5 /tmp/tmp.zadE7jDyBW +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo demand-backup-proxysql ++++ return +++ local cluster_proxy=demand-backup-proxysql +++ echo proxysql ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9Oas6IwfqD +++ mktemp ++ local LAST_ERR=/tmp/tmp.fN1qxkcfMW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9Oas6IwfqD ++ cat /tmp/tmp.fN1qxkcfMW ++ rm /tmp/tmp.9Oas6IwfqD /tmp/tmp.fN1qxkcfMW ++ return 0 + [[ 2 == \2 ]] + echo + compare_kubectl secret/demand-backup-mysql-init + local resource=secret/demand-backup-mysql-init + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/secret_demand-backup-mysql-init.yml + local new_result=/tmp/tmp.14SIjS2s6f/secret_demand-backup-mysql-init.yml + desc 'compare secret/demand-backup-mysql-init-' + set +o xtrace ----------------------------------------------------------------------------------- compare secret/demand-backup-mysql-init- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/secret_demand-backup-mysql-init-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/secret_demand-backup-mysql-init-80.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + version_gt 1.33 ++ bc -l ++ echo '1.32 >= 1.33' + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ bc -l ++ echo '1.32 >= 1.29' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/secret_demand-backup-mysql-init-k129.yml ']' + version_gt 1.27 ++ echo '1.32 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/secret_demand-backup-mysql-init-k127.yml ']' + version_gt 1.24 ++ bc -l ++ echo '1.32 >= 1.24' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/secret_demand-backup-mysql-init-k124.yml ']' + version_gt 1.22 ++ bc -l ++ echo '1.32 >= 1.22' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/secret_demand-backup-mysql-init-k122.yml ']' + version_gt 1.21 ++ echo '1.32 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/secret_demand-backup-mysql-init-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/secret_demand-backup-mysql-init-oc.yml ']' + version_gt 1.29 ++ bc -l ++ echo '1.32 >= 1.29' + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/secret_demand-backup-mysql-init-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/secret_demand-backup-mysql-init-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/secret_demand-backup-mysql-init-aks.yml ']' + kubectl_bin get -o yaml secret/demand-backup-mysql-init + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.spec.template.spec.containers[].env[] | select(.name == "XTRABACKUP_ENABLED")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.metadata.annotations."kubernetes.digitalocean.com/load-balancer-id") | del(.metadata.annotations."service.beta.kubernetes.io/do-loadbalancer-type") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("demand-backup-28441", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - ++ mktemp + local LAST_OUT=/tmp/tmp.kz03l3KB1v ++ mktemp + local LAST_ERR=/tmp/tmp.ZGkeqPcWQl + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml secret/demand-backup-mysql-init + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.kz03l3KB1v + cat /tmp/tmp.ZGkeqPcWQl + rm /tmp/tmp.kz03l3KB1v /tmp/tmp.ZGkeqPcWQl + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/compare/secret_demand-backup-mysql-init.yml /tmp/tmp.14SIjS2s6f/secret_demand-backup-mysql-init.yml + log 'compare_kubectl: secret/demand-backup-mysql-init OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-03-30T10:39:44+0000]' compare_kubectl: secret/demand-backup-mysql-init OK [2026-03-30T10:39:44+0000] compare_kubectl: secret/demand-backup-mysql-init OK + run_recovery_check demand-backup on-demand-backup-minio + local cluster=demand-backup + local backup=on-demand-backup-minio ++ get_proxy_engine demand-backup ++ local cluster_name=demand-backup +++ get_proxy demand-backup +++ local target_cluster=demand-backup ++++ kubectl_bin get pxc demand-backup -o 'jsonpath={.spec.haproxy.enabled}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.k5RATdsV6H +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.KRRUVVgc1W ++++ local exit_status=0 +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ set +e ++++ kubectl get pxc demand-backup -o 'jsonpath={.spec.haproxy.enabled}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 ']' ++++ break ++++ cat /tmp/tmp.k5RATdsV6H ++++ cat /tmp/tmp.KRRUVVgc1W ++++ rm /tmp/tmp.k5RATdsV6H /tmp/tmp.KRRUVVgc1W ++++ return 0 +++ [[ '' == \t\r\u\e ]] ++++ kubectl_bin get pxc demand-backup -o 'jsonpath={.spec.proxysql.enabled}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.b8PpMtmA5F +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.QXxPQjjexY ++++ local exit_status=0 +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ set +e ++++ kubectl get pxc demand-backup -o 'jsonpath={.spec.proxysql.enabled}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 ']' ++++ break ++++ cat /tmp/tmp.b8PpMtmA5F ++++ cat /tmp/tmp.QXxPQjjexY ++++ rm /tmp/tmp.b8PpMtmA5F /tmp/tmp.QXxPQjjexY ++++ return 0 +++ [[ true == \t\r\u\e ]] +++ echo demand-backup-proxysql +++ return ++ local cluster_proxy=demand-backup-proxysql ++ echo proxysql + local proxy=proxysql ++ get_pvc_name_for_backup on-demand-backup-minio ++ local backup_name=on-demand-backup-minio +++ kubectl_bin get pxc-backup on-demand-backup-minio -o 'jsonpath={.status.destination}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.N0pCc6ZJZb ++++ mktemp +++ local LAST_ERR=/tmp/tmp.0NebBRMgod +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc-backup on-demand-backup-minio -o 'jsonpath={.status.destination}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.N0pCc6ZJZb +++ cat /tmp/tmp.0NebBRMgod +++ rm /tmp/tmp.N0pCc6ZJZb /tmp/tmp.0NebBRMgod +++ return 0 ++ local destination=s3://operator-testing/prefix/demand-backup-2026-03-30-10:32:35-full ++ [[ -z s3://operator-testing/prefix/demand-backup-2026-03-30-10:32:35-full ]] ++ local pvc_name=s3://operator-testing/prefix/demand-backup-2026-03-30-10:32:35-full ++ echo s3://operator-testing/prefix/demand-backup-2026-03-30-10:32:35-full + pvc_name=s3://operator-testing/prefix/demand-backup-2026-03-30-10:32:35-full + dest=pvc/s3://operator-testing/prefix/demand-backup-2026-03-30-10:32:35-full + log 'run pxc-restore/on-demand-backup-minio' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-03-30T10:39:48+0000]' run pxc-restore/on-demand-backup-minio [2026-03-30T10:39:48+0000] run pxc-restore/on-demand-backup-minio + YAML_INPUT_FILE=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/restore-on-demand-backup-minio.yaml ++ cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2382/e2e-tests/demand-backup/conf/restore-on-demand-backup-minio.yaml ++ /usr/bin/sed -e s~minio-service.#namespace~minio-service.demand-backup-28441~ ++ /usr/bin/sed -e 's##pvc/s3://operator-testing/prefix/demand-backup-2026-03-30-10:32:35-full#g' + MODIFIED_YAML='apiVersion: pxc.percona.com/v1 kind: PerconaXtraDBClusterRestore metadata: name: on-demand-backup-minio spec: pxcCluster: demand-backup backupName: on-demand-backup-minio containerOptions: args: xbstream: - '\''--parallel=3'\'' xbcloud: - '\''--parallel=3'\'' xtrabackup: - '\''--parallel=3'\'' resources: requests: memory: 0.5G cpu: 500m limits: memory: "2G" cpu: "1"' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + FINAL_YAML='apiVersion: pxc.percona.com/v1 kind: PerconaXtraDBClusterRestore metadata: name: on-demand-backup-minio spec: pxcCluster: demand-backup backupName: on-demand-backup-minio containerOptions: args: xbstream: - '\''--parallel=3'\'' xbcloud: - '\''--parallel=3'\'' xtrabackup: - '\''--parallel=3'\'' resources: requests: memory: 0.5G cpu: 500m limits: memory: "2G" cpu: "1"' + kubectl_bin apply -f - + echo 'apiVersion: pxc.percona.com/v1 kind: PerconaXtraDBClusterRestore metadata: name: on-demand-backup-minio spec: pxcCluster: demand-backup backupName: on-demand-backup-minio containerOptions: args: xbstream: - '\''--parallel=3'\'' xbcloud: - '\''--parallel=3'\'' xtrabackup: - '\''--parallel=3'\'' resources: requests: memory: 0.5G cpu: 500m limits: memory: "2G" cpu: "1"' ++ mktemp + local LAST_OUT=/tmp/tmp.ihzAawMkHI ++ mktemp + local LAST_ERR=/tmp/tmp.cm6X6osTn9 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ihzAawMkHI perconaxtradbclusterrestore.pxc.percona.com/on-demand-backup-minio created + cat /tmp/tmp.cm6X6osTn9 + rm /tmp/tmp.ihzAawMkHI /tmp/tmp.cm6X6osTn9 + return 0 + wait_backup_restore on-demand-backup-minio + local backup_name=on-demand-backup-minio + local target_state=Succeeded + local wait_time=720 + set +o xtrace waiting for pxc-restore/on-demand-backup-minio to reach Succeeded state apiVersion: pxc.percona.com/v1 kind: PerconaXtraDBClusterRestore metadata: annotations: kubectl.kubernetes.io/last-applied-configuration: | {"apiVersion":"pxc.percona.com/v1","kind":"PerconaXtraDBClusterRestore","metadata":{"annotations":{},"name":"on-demand-backup-minio","namespace":"demand-backup-28441"},"spec":{"backupName":"on-demand-backup-minio","containerOptions":{"args":{"xbcloud":["--parallel=3"],"xbstream":["--parallel=3"],"xtrabackup":["--parallel=3"]}},"pxcCluster":"demand-backup","resources":{"limits":{"cpu":"1","memory":"2G"},"requests":{"cpu":"500m","memory":"0.5G"}}}} creationTimestamp: "2026-03-30T10:39:51Z" generation: 1 name: on-demand-backup-minio namespace: demand-backup-28441 resourceVersion: "1774867191365103018" uid: 5c851edd-3b5d-4b37-add9-a40374ab24ec spec: backupName: on-demand-backup-minio containerOptions: args: xbcloud: - --parallel=3 xbstream: - --parallel=3 xtrabackup: - --parallel=3 pxcCluster: demand-backup resources: limits: cpu: "1" memory: 2G requests: cpu: 500m memory: 0.5G status: comments: 'failed to validate restore job: failed to validate backup existence: backup not found' state: Failed unsafeFlags: {} Restore object pxc-restore/on-demand-backup-minio is in Failed state. something went wrong with operator or kubernetes cluster