Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/logs/demand-backup-8-0.log grep: warning: stray \ before - Warning: version difference between client (1.34) and server (1.31) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.34) and server (1.31) exceeds the supported minor version skew of +/-1 + main + create_infra demand-backup-32192 + local ns=demand-backup-32192 + '[' -n pxc-operator ']' + kubectl get pxc --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch pxc -n demand-backup-19345 demand-backup --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/demand-backup patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.eBZjJ86x03 ++ mktemp + local LAST_ERR=/tmp/tmp.kmak5MuKUi + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.eBZjJ86x03 perconaxtradbcluster.pxc.percona.com "demand-backup" deleted from demand-backup-19345 namespace + cat /tmp/tmp.kmak5MuKUi + rm /tmp/tmp.eBZjJ86x03 /tmp/tmp.kmak5MuKUi + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.D3SuRHtxRB ++ mktemp + local LAST_ERR=/tmp/tmp.bMeHTLQbCk + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.D3SuRHtxRB perconaxtradbclusterbackup.pxc.percona.com "on-demand-backup-pvc" deleted from demand-backup-19345 namespace + cat /tmp/tmp.bMeHTLQbCk + rm /tmp/tmp.D3SuRHtxRB /tmp/tmp.bMeHTLQbCk + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.hoRrrY6RGW ++ mktemp + local LAST_ERR=/tmp/tmp.hiLnPAfVll + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.hoRrrY6RGW perconaxtradbclusterrestore.pxc.percona.com "on-demand-backup-pvc" deleted from demand-backup-19345 namespace + cat /tmp/tmp.hiLnPAfVll + rm /tmp/tmp.hoRrrY6RGW /tmp/tmp.hiLnPAfVll + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + egrep -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + awk '{print$1}' + '[' -n '' ']' + desc 'cleaned up old namespaces pxc-operator' + xargs kubectl delete ns + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pxc-operator ++ mktemp ++ mktemp egrep: warning: egrep is obsolescent; using grep -E + local LAST_OUT=/tmp/tmp.voJwXbFb4T + local LAST_OUT=/tmp/tmp.YAvotWaEuC ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.hiHgHfyIkd + local exit_status=0 ++ seq 0 2 + local LAST_ERR=/tmp/tmp.9KrJUkHwuj + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete namespace pxc-operator + for i in $(seq 0 2) + set +e + kubectl get ns + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.voJwXbFb4T + cat /tmp/tmp.9KrJUkHwuj + rm /tmp/tmp.voJwXbFb4T /tmp/tmp.9KrJUkHwuj + return 0 namespace "demand-backup-19345" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.YAvotWaEuC namespace "pxc-operator" deleted + cat /tmp/tmp.hiHgHfyIkd + rm /tmp/tmp.YAvotWaEuC /tmp/tmp.hiHgHfyIkd + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + echo -n 'waiting for namespace/pxc-operator to be deleted' waiting for namespace/pxc-operator to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pxc-operator" not found + desc 'create namespace pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.SBX53x1p8U ++ mktemp + local LAST_ERR=/tmp/tmp.zzohi2gMwe + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.SBX53x1p8U namespace/pxc-operator created + cat /tmp/tmp.zzohi2gMwe + rm /tmp/tmp.SBX53x1p8U /tmp/tmp.zzohi2gMwe + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.Itm7OWY08d +++ mktemp ++ local LAST_ERR=/tmp/tmp.SQklSRaPGy ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Itm7OWY08d ++ cat /tmp/tmp.SQklSRaPGy ++ rm /tmp/tmp.Itm7OWY08d /tmp/tmp.SQklSRaPGy ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2200-89830e6d-1-cluster7 --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.9RPnuwuAXp ++ mktemp + local LAST_ERR=/tmp/tmp.HWjJK4Po95 + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2200-89830e6d-1-cluster7 --namespace=pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.9RPnuwuAXp Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2200-89830e6d-1-cluster7" modified. + cat /tmp/tmp.HWjJK4Po95 + rm /tmp/tmp.9RPnuwuAXp /tmp/tmp.HWjJK4Po95 + return 0 + deploy_operator + desc 'start PXC operator' + set +o xtrace ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.RBb621Don9 ++ mktemp + local LAST_ERR=/tmp/tmp.lJb6ztpevZ + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.RBb621Don9 customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied + cat /tmp/tmp.lJb6ztpevZ + rm /tmp/tmp.RBb621Don9 /tmp/tmp.lJb6ztpevZ + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: pxc-operator^' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.jH8hTDUvk2 ++ mktemp + local LAST_ERR=/tmp/tmp.xkerETzl1U + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.jH8hTDUvk2 clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.xkerETzl1U + rm /tmp/tmp.jH8hTDUvk2 /tmp/tmp.xkerETzl1U + return 0 + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/deploy/cw-operator.yaml + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-2200-89830e6d^' + sed -e 's^failureThreshold: .*^failureThreshold: 10^' + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "DISABLE_TELEMETRY").value) = "true"' - + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "LOG_LEVEL").value) = "VERBOSE"' - + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.ZDpHCVcy7Y ++ mktemp + local LAST_ERR=/tmp/tmp.ToKRyjbDox + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ZDpHCVcy7Y deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.ToKRyjbDox + rm /tmp/tmp.ZDpHCVcy7Y /tmp/tmp.ToKRyjbDox + return 0 + sleep 10 + kubectl_bin wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s ++ mktemp + local LAST_OUT=/tmp/tmp.eYPQchyieP ++ mktemp + local LAST_ERR=/tmp/tmp.nDHhJsN7ri + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.eYPQchyieP pod/percona-xtradb-cluster-operator-59745b97cd-s5wt4 condition met + cat /tmp/tmp.nDHhJsN7ri + rm /tmp/tmp.eYPQchyieP /tmp/tmp.nDHhJsN7ri + return 0 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.LwrrQgVqP3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.cqcBhPHa6c ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.LwrrQgVqP3 ++ cat /tmp/tmp.cqcBhPHa6c ++ rm /tmp/tmp.LwrrQgVqP3 /tmp/tmp.cqcBhPHa6c ++ return 0 + wait_pod percona-xtradb-cluster-operator-59745b97cd-s5wt4 480 pxc-operator + local pod=percona-xtradb-cluster-operator-59745b97cd-s5wt4 + local max_retry=480 + local ns=pxc-operator ++ echo percona-xtradb-cluster-operator-59745b97cd-s5wt4 ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-59745b97cd-s5wt4 condition met waiting for pod/percona-xtradb-cluster-operator-59745b97cd-s5wt4 to become Ready.Ok + sleep 3 + create_namespace demand-backup-32192 + local namespace=demand-backup-32192 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ sed s/NAMESPACE// ++ awk '-F ' '{print $2}' + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + egrep -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + awk '{print$1}' + '[' -n '' ']' + desc 'cleaned up old namespaces demand-backup-32192' + set +o xtrace ++ mktemp + xargs kubectl delete ns ----------------------------------------------------------------------------------- cleaned up old namespaces demand-backup-32192 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace demand-backup-32192 ++ mktemp + local LAST_OUT=/tmp/tmp.98svsEDbAs egrep: warning: egrep is obsolescent; using grep -E ++ mktemp + local LAST_OUT=/tmp/tmp.fIua5cI3Mn ++ mktemp + local LAST_ERR=/tmp/tmp.PNBjJ9xItI + local exit_status=0 + local LAST_ERR=/tmp/tmp.feEluYXE0n + local exit_status=0 ++ seq 0 2 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get ns + for i in $(seq 0 2) + set +e + kubectl delete namespace demand-backup-32192 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in $(seq 0 2) + set +e + kubectl delete namespace demand-backup-32192 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.98svsEDbAs + cat /tmp/tmp.PNBjJ9xItI + rm /tmp/tmp.98svsEDbAs /tmp/tmp.PNBjJ9xItI + return 0 error: resource(s) were provided, but no name was specified + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in $(seq 0 2) + set +e + kubectl delete namespace demand-backup-32192 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.fIua5cI3Mn + cat /tmp/tmp.feEluYXE0n Error from server (NotFound): namespaces "demand-backup-32192" not found + rm /tmp/tmp.fIua5cI3Mn /tmp/tmp.feEluYXE0n + return 1 + : + wait_for_delete namespace/demand-backup-32192 + local res=namespace/demand-backup-32192 + echo -n 'waiting for namespace/demand-backup-32192 to be deleted' waiting for namespace/demand-backup-32192 to be deleted+ set +o xtrace Error from server (NotFound): namespaces "demand-backup-32192" not found + desc 'create namespace demand-backup-32192' + set +o xtrace ----------------------------------------------------------------------------------- create namespace demand-backup-32192 ----------------------------------------------------------------------------------- + kubectl_bin create namespace demand-backup-32192 ++ mktemp + local LAST_OUT=/tmp/tmp.6kk1MpFuBx ++ mktemp + local LAST_ERR=/tmp/tmp.zAN6MQ0Z7x + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace demand-backup-32192 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.6kk1MpFuBx namespace/demand-backup-32192 created + cat /tmp/tmp.zAN6MQ0Z7x + rm /tmp/tmp.6kk1MpFuBx /tmp/tmp.zAN6MQ0Z7x + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.aJyPuziq67 +++ mktemp ++ local LAST_ERR=/tmp/tmp.RB3NlmWFLa ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.aJyPuziq67 ++ cat /tmp/tmp.RB3NlmWFLa ++ rm /tmp/tmp.aJyPuziq67 /tmp/tmp.RB3NlmWFLa ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2200-89830e6d-1-cluster7 --namespace=demand-backup-32192 ++ mktemp + local LAST_OUT=/tmp/tmp.p5VZObVeNw ++ mktemp + local LAST_ERR=/tmp/tmp.4ETAxv6J50 + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2200-89830e6d-1-cluster7 --namespace=demand-backup-32192 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.p5VZObVeNw Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2200-89830e6d-1-cluster7" modified. + cat /tmp/tmp.4ETAxv6J50 + rm /tmp/tmp.p5VZObVeNw /tmp/tmp.4ETAxv6J50 + return 0 + apply_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.Z5bGdzcTVu ++ mktemp + local LAST_ERR=/tmp/tmp.hnr8fAcCzi + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Z5bGdzcTVu secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.hnr8fAcCzi + rm /tmp/tmp.Z5bGdzcTVu /tmp/tmp.hnr8fAcCzi + return 0 + cluster=demand-backup + spinup_pxc demand-backup /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/conf/demand-backup.yml + local cluster=demand-backup + local config=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/conf/demand-backup.yml + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/conf/secrets.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/conf/client.yml + local port=3306 + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.hGueilJxYo ++ mktemp + local LAST_ERR=/tmp/tmp.35Wvspd2NH + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/conf/secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.hGueilJxYo secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created + cat /tmp/tmp.35Wvspd2NH + rm /tmp/tmp.hGueilJxYo /tmp/tmp.35Wvspd2NH + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/conf/client.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/conf/client.yml + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/conf/client.yml ++ mktemp + /usr/sbin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/sbin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/sbin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/sbin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2200-89830e6d#' + /usr/sbin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/sbin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + local LAST_OUT=/tmp/tmp.YM1FUeYn9s ++ mktemp + /usr/sbin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/sbin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/sbin/sed -e s~minio-service.#namespace~minio-service.demand-backup-32192~ + /usr/sbin/sed -e 's#apply:.*#apply: Never#' + /usr/sbin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + local LAST_ERR=/tmp/tmp.q4QMOjCsP1 + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.YM1FUeYn9s deployment.apps/pxc-client created + cat /tmp/tmp.q4QMOjCsP1 + rm /tmp/tmp.YM1FUeYn9s /tmp/tmp.q4QMOjCsP1 + return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/conf/demand-backup.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/conf/demand-backup.yml + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/conf/demand-backup.yml + /usr/sbin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/sbin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' ++ mktemp + /usr/sbin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/sbin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/sbin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2200-89830e6d#' + /usr/sbin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + local LAST_OUT=/tmp/tmp.j0D9p1nTlk ++ mktemp + /usr/sbin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/sbin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/sbin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + local LAST_ERR=/tmp/tmp.BG3sqpbrcz + /usr/sbin/sed -e s~minio-service.#namespace~minio-service.demand-backup-32192~ + local exit_status=0 + /usr/sbin/sed -e 's#apply:.*#apply: Never#' ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.j0D9p1nTlk perconaxtradbcluster.pxc.percona.com/demand-backup created + cat /tmp/tmp.BG3sqpbrcz + rm /tmp/tmp.j0D9p1nTlk /tmp/tmp.BG3sqpbrcz + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- ++ get_proxy demand-backup ++ local target_cluster=demand-backup +++ kubectl_bin get pxc demand-backup -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Ulr08S4Avn ++++ mktemp +++ local LAST_ERR=/tmp/tmp.i4i2ISYB9d +++ local exit_status=0 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get pxc demand-backup -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.Ulr08S4Avn +++ cat /tmp/tmp.i4i2ISYB9d +++ rm /tmp/tmp.Ulr08S4Avn /tmp/tmp.i4i2ISYB9d +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc demand-backup -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.3FZIuoK7zl ++++ mktemp +++ local LAST_ERR=/tmp/tmp.NCsM0xORso +++ local exit_status=0 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get pxc demand-backup -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.3FZIuoK7zl +++ cat /tmp/tmp.NCsM0xORso +++ rm /tmp/tmp.3FZIuoK7zl /tmp/tmp.NCsM0xORso +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo demand-backup-proxysql ++ return + local proxy=demand-backup-proxysql + kubectl_bin wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n demand-backup-32192 ++ mktemp + local LAST_OUT=/tmp/tmp.bZwAeOPRpc ++ mktemp + local LAST_ERR=/tmp/tmp.8OmsNUwJRD + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n demand-backup-32192 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in $(seq 0 2) + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n demand-backup-32192 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in $(seq 0 2) + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n demand-backup-32192 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.bZwAeOPRpc + cat /tmp/tmp.8OmsNUwJRD error: no matching resources found + rm /tmp/tmp.bZwAeOPRpc /tmp/tmp.8OmsNUwJRD + return 1 + true + wait_for_running demand-backup-proxysql 1 + local name=demand-backup-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in $(seq 0 $last_pod) + wait_pod demand-backup-proxysql-0 480 + local pod=demand-backup-proxysql-0 + local max_retry=480 + local ns= ++ echo demand-backup-proxysql-0 ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container=proxysql + set +o xtrace pod/demand-backup-proxysql-0 condition met waiting for pod/demand-backup-proxysql-0 to become Ready.Ok + wait_for_running demand-backup-pxc 3 + local name=demand-backup-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in $(seq 0 $last_pod) + wait_pod demand-backup-pxc-0 480 + local pod=demand-backup-pxc-0 + local max_retry=480 + local ns= ++ echo demand-backup-pxc-0 ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container=pxc + set +o xtrace pod/demand-backup-pxc-0 condition met waiting for pod/demand-backup-pxc-0 to become Ready.Ok + for i in $(seq 0 $last_pod) + wait_pod demand-backup-pxc-1 480 + local pod=demand-backup-pxc-1 + local max_retry=480 + local ns= ++ echo demand-backup-pxc-1 ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container=pxc + set +o xtrace pod/demand-backup-pxc-1 condition met waiting for pod/demand-backup-pxc-1 to become Ready.Ok + for i in $(seq 0 $last_pod) + wait_pod demand-backup-pxc-2 480 + local pod=demand-backup-pxc-2 + local max_retry=480 + local ns= ++ echo demand-backup-pxc-2 ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container=pxc + set +o xtrace pod/demand-backup-pxc-2 condition met waiting for pod/demand-backup-pxc-2 to become Ready.Ok + sleep 10 ++ kubectl get pxc demand-backup -o 'jsonpath={.spec.secretsName}' + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.oNpiBNFox2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.WuzsP3kQK0 ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.oNpiBNFox2 ++ cat /tmp/tmp.WuzsP3kQK0 ++ rm /tmp/tmp.oNpiBNFox2 /tmp/tmp.WuzsP3kQK0 ++ return 0 + local root_pass=root_password + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' '-h demand-backup-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' + local 'uri=-h demand-backup-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8ENWWMuEe1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.7t3kml17fk ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8ENWWMuEe1 ++ cat /tmp/tmp.7t3kml17fk ++ rm /tmp/tmp.8ENWWMuEe1 /tmp/tmp.7t3kml17fk ++ return 0 + client_pod=pxc-client-59944c5bbf-5pzc2 + wait_pod pxc-client-59944c5bbf-5pzc2 + local pod=pxc-client-59944c5bbf-5pzc2 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-5pzc2 ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container= + set +o xtrace pod/pxc-client-59944c5bbf-5pzc2 condition met waiting for pod/pxc-client-59944c5bbf-5pzc2 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h demand-backup-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h demand-backup-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oQNteTIJtb +++ mktemp ++ local LAST_ERR=/tmp/tmp.Rz9kBZpp83 ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.oQNteTIJtb ++ cat /tmp/tmp.Rz9kBZpp83 ++ rm /tmp/tmp.oQNteTIJtb /tmp/tmp.Rz9kBZpp83 ++ return 0 + client_pod=pxc-client-59944c5bbf-5pzc2 + wait_pod pxc-client-59944c5bbf-5pzc2 + local pod=pxc-client-59944c5bbf-5pzc2 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-5pzc2 ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container= + set +o xtrace pod/pxc-client-59944c5bbf-5pzc2 condition met waiting for pod/pxc-client-59944c5bbf-5pzc2 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in $(seq 0 $((size - 1))) + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-0.demand-backup-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-0.demand-backup-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-0.demand-backup-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-0.demand-backup-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8aTlwWrqEZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.ngExtgVOVm ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8aTlwWrqEZ ++ cat /tmp/tmp.ngExtgVOVm ++ rm /tmp/tmp.8aTlwWrqEZ /tmp/tmp.ngExtgVOVm ++ return 0 + client_pod=pxc-client-59944c5bbf-5pzc2 + wait_pod pxc-client-59944c5bbf-5pzc2 + local pod=pxc-client-59944c5bbf-5pzc2 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-5pzc2 ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container= + set +o xtrace pod/pxc-client-59944c5bbf-5pzc2 condition met waiting for pod/pxc-client-59944c5bbf-5pzc2 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.ZiSlkZxGqo/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/compare/select-1.sql /tmp/tmp.ZiSlkZxGqo/select-1.sql + for i in $(seq 0 $((size - 1))) + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-1.demand-backup-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-1.demand-backup-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-1.demand-backup-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-1.demand-backup-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XcbbWDK8LN +++ mktemp ++ local LAST_ERR=/tmp/tmp.wVr3fCdugm ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XcbbWDK8LN ++ cat /tmp/tmp.wVr3fCdugm ++ rm /tmp/tmp.XcbbWDK8LN /tmp/tmp.wVr3fCdugm ++ return 0 + client_pod=pxc-client-59944c5bbf-5pzc2 + wait_pod pxc-client-59944c5bbf-5pzc2 + local pod=pxc-client-59944c5bbf-5pzc2 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-5pzc2 ++ egrep '^(pxc|proxysql)$' ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' egrep: warning: egrep is obsolescent; using grep -E + local container= + set +o xtrace pod/pxc-client-59944c5bbf-5pzc2 condition met waiting for pod/pxc-client-59944c5bbf-5pzc2 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.ZiSlkZxGqo/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/compare/select-1.sql /tmp/tmp.ZiSlkZxGqo/select-1.sql + for i in $(seq 0 $((size - 1))) + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-2.demand-backup-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-2.demand-backup-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-2.demand-backup-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-2.demand-backup-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BbFIzi4Rsu +++ mktemp ++ local LAST_ERR=/tmp/tmp.9MiKDke37Y ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.BbFIzi4Rsu ++ cat /tmp/tmp.9MiKDke37Y ++ rm /tmp/tmp.BbFIzi4Rsu /tmp/tmp.9MiKDke37Y ++ return 0 + client_pod=pxc-client-59944c5bbf-5pzc2 + wait_pod pxc-client-59944c5bbf-5pzc2 + local pod=pxc-client-59944c5bbf-5pzc2 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-5pzc2 ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container= + set +o xtrace pod/pxc-client-59944c5bbf-5pzc2 condition met waiting for pod/pxc-client-59944c5bbf-5pzc2 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.ZiSlkZxGqo/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/compare/select-1.sql /tmp/tmp.ZiSlkZxGqo/select-1.sql ++ is_keyring_plugin_in_use demand-backup ++ local cluster=demand-backup ++ kubectl_bin exec -it demand-backup-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ egrep -o 'early-plugin-load=keyring_\w+.so' +++ mktemp egrep: warning: egrep is obsolescent; using grep -E ++ local LAST_OUT=/tmp/tmp.txVyvwlIpz +++ mktemp ++ local LAST_ERR=/tmp/tmp.GDIBC2epzQ ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl exec -it demand-backup-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.txVyvwlIpz ++ cat /tmp/tmp.GDIBC2epzQ Unable to use a TTY - input is not a terminal or the right kind of file ++ rm /tmp/tmp.txVyvwlIpz /tmp/tmp.GDIBC2epzQ ++ return 0 + '[' '' ']' + start_minio demand-backup-ssl + deploy_helm demand-backup-32192 + helm repo add hashicorp https://helm.releases.hashicorp.com "hashicorp" already exists with the same configuration, skipping + helm repo add minio https://charts.min.io/ "minio" already exists with the same configuration, skipping + helm repo update Hang tight while we grab the latest from your chart repositories... ...Successfully got an update from the "minio" chart repository ...Successfully got an update from the "chaos-mesh" chart repository ...Successfully got an update from the "hashicorp" chart repository Update Complete. ⎈Happy Helming!⎈ + local cert_secret=demand-backup-ssl + local endpoint=http://minio-service:9000 + minio_args=('--version' '5.4.0' '--set' 'replicas=1' '--set' 'mode=standalone' '--set' 'resources.requests.memory=256Mi' '--set' 'rootUser=rootuser' '--set' 'rootPassword=rootpass123' '--set' 'users[0].accessKey=some-access-key' '--set' 'users[0].secretKey=some-secret-key' '--set' 'users[0].policy=consoleAdmin' '--set' 'service.type=ClusterIP' '--set' 'configPathmc=/tmp/' '--set' 'securityContext.enabled=false' '--set' 'persistence.size=2G') + local minio_args + [[ -n demand-backup-ssl ]] + endpoint=https://minio-service:9000 + minio_args+=(--set tls.enabled=true --set tls.certSecret="$cert_secret" --set tls.publicCrt=tls.crt --set tls.privateKey=tls.key) + desc 'install Minio' + set +o xtrace ----------------------------------------------------------------------------------- install Minio ----------------------------------------------------------------------------------- + helm uninstall minio-service Error: uninstall: Release not loaded: minio-service: release: not found + : + retry 10 60 helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G --set tls.enabled=true --set tls.certSecret=demand-backup-ssl --set tls.publicCrt=tls.crt --set tls.privateKey=tls.key minio/minio + local max=10 + local delay=60 + shift 2 + local n=1 + helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G --set tls.enabled=true --set tls.certSecret=demand-backup-ssl --set tls.publicCrt=tls.crt --set tls.privateKey=tls.key minio/minio NAME: minio-service LAST DEPLOYED: Wed Oct 1 01:57:00 2025 NAMESPACE: demand-backup-32192 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: MinIO can be accessed via port 9000 on the following DNS name from within your cluster: minio-service.demand-backup-32192.cluster.local To access MinIO from localhost, run the below commands: 1. export POD_NAME=$(kubectl get pods --namespace demand-backup-32192 -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}") 2. kubectl port-forward $POD_NAME 9000 --namespace demand-backup-32192 Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/ You can now access MinIO server on http://localhost:9000. Follow the below steps to connect to MinIO server with mc client: 1. Download the MinIO mc client - https://min.io/docs/minio/linux/reference/minio-mc.html#quickstart 2. export MC_HOST_minio-service-local=http://$(kubectl get secret --namespace demand-backup-32192 minio-service -o jsonpath="{.data.rootUser}" | base64 --decode):$(kubectl get secret --namespace demand-backup-32192 minio-service -o jsonpath="{.data.rootPassword}" | base64 --decode)@localhost:9000 3. mc ls minio-service-local + sleep 30 ++ kubectl_bin get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JdY8p18EQz +++ mktemp ++ local LAST_ERR=/tmp/tmp.zRRDNwvTHp ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.JdY8p18EQz ++ cat /tmp/tmp.zRRDNwvTHp ++ rm /tmp/tmp.JdY8p18EQz /tmp/tmp.zRRDNwvTHp ++ return 0 + MINIO_POD=minio-service-f46fd9c97-tdfwd + wait_pod minio-service-f46fd9c97-tdfwd + local pod=minio-service-f46fd9c97-tdfwd + local max_retry=480 + local ns= ++ echo minio-service-f46fd9c97-tdfwd ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container= + set +o xtrace pod/minio-service-f46fd9c97-tdfwd condition met waiting for pod/minio-service-f46fd9c97-tdfwd to become Ready.Ok + kubectl_bin run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- /usr/bin/env AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url https://minio-service:9000 --no-verify-ssl s3 mb s3://operator-testing ++ mktemp + local LAST_OUT=/tmp/tmp.A35QNYbNYL ++ mktemp + local LAST_ERR=/tmp/tmp.XbynIOl20W + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- /usr/bin/env AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url https://minio-service:9000 --no-verify-ssl s3 mb s3://operator-testing + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.A35QNYbNYL /usr/lib/python2.7/site-packages/botocore/vendored/requests/packages/urllib3/connectionpool.py:768: InsecureRequestWarning: Unverified HTTPS request is being made. Adding certificate verification is strongly advised. See: https://urllib3.readthedocs.org/en/latest/security.html InsecureRequestWarning) make_bucket: operator-testing pod "aws-cli" deleted from demand-backup-32192 namespace + cat /tmp/tmp.XbynIOl20W + rm /tmp/tmp.A35QNYbNYL /tmp/tmp.XbynIOl20W + return 0 + run_backup demand-backup on-demand-backup-pvc + local cluster=demand-backup + local backup=on-demand-backup-pvc + log 'run pxc-backup/on-demand-backup-pvc' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-10-01T01:58:09+0000]' run pxc-backup/on-demand-backup-pvc [2025-10-01T01:58:09+0000] run pxc-backup/on-demand-backup-pvc + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/conf/on-demand-backup-pvc.yml ++ mktemp + local LAST_OUT=/tmp/tmp.tZKAZoy0oS ++ mktemp + local LAST_ERR=/tmp/tmp.MRBHNgcmm9 + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/conf/on-demand-backup-pvc.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.tZKAZoy0oS perconaxtradbclusterbackup.pxc.percona.com/on-demand-backup-pvc created + cat /tmp/tmp.MRBHNgcmm9 + rm /tmp/tmp.tZKAZoy0oS /tmp/tmp.MRBHNgcmm9 + return 0 + wait_backup on-demand-backup-pvc + local backup=on-demand-backup-pvc + local status=Succeeded + set +o xtrace waiting for pxc-backup/on-demand-backup-pvc to reach Succeeded state........................Succeeded + run_recovery_check demand-backup on-demand-backup-pvc + local cluster=demand-backup + local backup=on-demand-backup-pvc ++ get_proxy_engine demand-backup ++ local cluster_name=demand-backup +++ get_proxy demand-backup +++ local target_cluster=demand-backup ++++ kubectl_bin get pxc demand-backup -o 'jsonpath={.spec.haproxy.enabled}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.v8mF4JJJCv +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.lRxJbEYD7J ++++ local exit_status=0 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get pxc demand-backup -o 'jsonpath={.spec.haproxy.enabled}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 ']' ++++ break ++++ cat /tmp/tmp.v8mF4JJJCv ++++ cat /tmp/tmp.lRxJbEYD7J ++++ rm /tmp/tmp.v8mF4JJJCv /tmp/tmp.lRxJbEYD7J ++++ return 0 +++ [[ '' == \t\r\u\e ]] ++++ kubectl_bin get pxc demand-backup -o 'jsonpath={.spec.proxysql.enabled}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.IS9FtXpycU +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.pWlXGvaj4E ++++ local exit_status=0 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get pxc demand-backup -o 'jsonpath={.spec.proxysql.enabled}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 ']' ++++ break ++++ cat /tmp/tmp.IS9FtXpycU ++++ cat /tmp/tmp.pWlXGvaj4E ++++ rm /tmp/tmp.IS9FtXpycU /tmp/tmp.pWlXGvaj4E ++++ return 0 +++ [[ true == \t\r\u\e ]] +++ echo demand-backup-proxysql +++ return ++ local cluster_proxy=demand-backup-proxysql ++ echo proxysql + local proxy=proxysql + log 'run pxc-restore/on-demand-backup-pvc' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-10-01T01:58:54+0000]' run pxc-restore/on-demand-backup-pvc [2025-10-01T01:58:54+0000] run pxc-restore/on-demand-backup-pvc + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/conf/restore-on-demand-backup-pvc.yaml + /usr/sbin/sed -e s~minio-service.#namespace~minio-service.demand-backup-32192~ + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.WI1EiZfmPh ++ mktemp + local LAST_ERR=/tmp/tmp.SWnEsHX7qk + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.WI1EiZfmPh perconaxtradbclusterrestore.pxc.percona.com/on-demand-backup-pvc created + cat /tmp/tmp.SWnEsHX7qk + rm /tmp/tmp.WI1EiZfmPh /tmp/tmp.SWnEsHX7qk + return 0 + wait_backup_restore on-demand-backup-pvc + local backup_name=on-demand-backup-pvc + local target_state=Succeeded + local wait_time=720 + set +o xtrace waiting for pxc-restore/on-demand-backup-pvc to reach Succeeded state 2025-10-01T01:58:57 pxc-restore/on-demand-backup-pvc state: Starting 2025-10-01T01:59:00 pxc-restore/on-demand-backup-pvc state: Starting 2025-10-01T01:59:02 pxc-restore/on-demand-backup-pvc state: Starting 2025-10-01T01:59:04 pxc-restore/on-demand-backup-pvc state: Starting 2025-10-01T01:59:06 pxc-restore/on-demand-backup-pvc state: Starting 2025-10-01T01:59:08 pxc-restore/on-demand-backup-pvc state: Starting 2025-10-01T01:59:10 pxc-restore/on-demand-backup-pvc state: Starting 2025-10-01T01:59:13 pxc-restore/on-demand-backup-pvc state: Starting 2025-10-01T01:59:15 pxc-restore/on-demand-backup-pvc state: Starting 2025-10-01T01:59:17 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T01:59:19 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T01:59:21 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T01:59:23 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T01:59:26 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T01:59:28 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T01:59:30 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T01:59:32 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T01:59:34 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T01:59:36 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T01:59:39 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T01:59:41 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T01:59:43 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T01:59:45 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T01:59:47 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T01:59:49 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T01:59:52 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T01:59:54 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T01:59:56 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T01:59:58 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T02:00:01 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T02:00:03 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T02:00:05 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T02:00:07 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T02:00:09 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T02:00:12 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T02:00:14 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T02:00:16 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-01T02:00:18 pxc-restore/on-demand-backup-pvc state: Restoring 2025-10-01T02:00:20 pxc-restore/on-demand-backup-pvc state: Restoring 2025-10-01T02:00:23 pxc-restore/on-demand-backup-pvc state: Restoring 2025-10-01T02:00:25 pxc-restore/on-demand-backup-pvc state: Restoring 2025-10-01T02:00:27 pxc-restore/on-demand-backup-pvc state: Restoring 2025-10-01T02:00:29 pxc-restore/on-demand-backup-pvc state: Restoring 2025-10-01T02:00:31 pxc-restore/on-demand-backup-pvc state: Restoring 2025-10-01T02:00:34 pxc-restore/on-demand-backup-pvc state: Restoring 2025-10-01T02:00:36 pxc-restore/on-demand-backup-pvc state: Restoring 2025-10-01T02:00:38 pxc-restore/on-demand-backup-pvc state: Restoring 2025-10-01T02:00:40 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:00:42 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:00:45 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:00:47 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:00:49 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:00:51 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:00:53 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:00:56 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:00:58 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:00 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:02 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:04 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:07 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:09 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:11 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:13 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:15 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:17 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:20 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:22 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:24 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:26 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:28 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:31 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:33 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:35 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:37 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:39 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:41 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:44 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:46 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:48 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:50 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:52 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:55 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:57 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:01:59 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:02:01 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:02:03 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-01T02:02:06 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:02:08 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:02:10 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:02:12 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:02:15 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:02:17 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:02:19 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:02:21 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:02:23 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:02:26 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:02:28 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:02:30 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:02:32 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:02:34 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:02:36 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:02:38 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:02:41 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:02:43 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:02:45 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:02:47 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:02:49 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:02:51 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:02:53 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:02:56 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:02:58 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:00 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:02 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:04 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:06 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:09 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:11 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:13 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:15 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:17 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:20 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:21 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:23 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:25 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:28 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:30 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:32 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:34 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:36 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:38 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:41 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:43 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:45 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:47 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:49 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:51 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:54 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:56 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:03:58 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:00 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:02 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:04 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:07 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:09 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:11 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:13 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:15 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:17 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:19 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:22 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:24 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:26 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:28 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:30 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:32 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:35 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:37 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:39 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:41 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:43 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:46 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:48 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:50 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:52 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:55 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:57 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:04:59 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:05:01 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:05:03 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:05:06 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:05:08 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:05:10 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:05:12 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:05:15 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:05:17 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:05:19 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:05:21 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:05:23 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:05:26 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:05:28 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:05:30 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-01T02:05:32 pxc-restore/on-demand-backup-pvc state: Succeeded + kubectl_bin logs job/restore-job-on-demand-backup-pvc-demand-backup ++ mktemp + local LAST_OUT=/tmp/tmp.K2fVtpKacw ++ mktemp + local LAST_ERR=/tmp/tmp.mjCT1refbt + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl logs job/restore-job-on-demand-backup-pvc-demand-backup + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.K2fVtpKacw + LIB_PATH=/opt/percona/backup/lib/pxc + . /opt/percona/backup/lib/pxc/check-version.sh + . /opt/percona/backup/lib/pxc/vault.sh ++ set -o errexit ++ keyring_vault=/etc/mysql/vault-keyring-secret/keyring_vault.conf + SOCAT_OPTS=TCP:restore-src-on-demand-backup-pvc-demand-backup:3307,retry=30 + check_ssl + CA=/var/run/secrets/kubernetes.io/serviceaccount/ca.crt + '[' -f /var/run/secrets/kubernetes.io/serviceaccount/service-ca.crt ']' + SSL_DIR=/etc/mysql/ssl + '[' -f /etc/mysql/ssl/ca.crt ']' + CA=/etc/mysql/ssl/ca.crt + SSL_INTERNAL_DIR=/etc/mysql/ssl-internal + '[' -f /etc/mysql/ssl-internal/ca.crt ']' + CA=/etc/mysql/ssl-internal/ca.crt + KEY=/etc/mysql/ssl/tls.key + CERT=/etc/mysql/ssl/tls.crt + '[' -f /etc/mysql/ssl-internal/tls.key ']' + '[' -f /etc/mysql/ssl-internal/tls.crt ']' + KEY=/etc/mysql/ssl-internal/tls.key + CERT=/etc/mysql/ssl-internal/tls.crt + '[' -f /etc/mysql/ssl-internal/ca.crt ']' + '[' -f /etc/mysql/ssl-internal/tls.key ']' + '[' -f /etc/mysql/ssl-internal/tls.crt ']' + SOCAT_OPTS='openssl-connect:restore-src-on-demand-backup-pvc-demand-backup:3307,reuseaddr,cert=/etc/mysql/ssl-internal/tls.crt,key=/etc/mysql/ssl-internal/tls.key,cafile=/etc/mysql/ssl-internal/ca.crt,verify=1,commonname='\'''\'',retry=30,no-sni=1' + ping -c1 restore-src-on-demand-backup-pvc-demand-backup /opt/percona/backup/recovery-pvc-joiner.sh: line 40: ping: command not found + : + rm -rf /datadir/#ib_16384_0.dblwr /datadir/#ib_16384_1.dblwr /datadir/#ib_16384_10.dblwr /datadir/#ib_16384_11.dblwr /datadir/#ib_16384_12.dblwr /datadir/#ib_16384_13.dblwr /datadir/#ib_16384_14.dblwr /datadir/#ib_16384_15.dblwr /datadir/#ib_16384_2.dblwr /datadir/#ib_16384_3.dblwr /datadir/#ib_16384_4.dblwr /datadir/#ib_16384_5.dblwr /datadir/#ib_16384_6.dblwr /datadir/#ib_16384_7.dblwr /datadir/#ib_16384_8.dblwr /datadir/#ib_16384_9.dblwr /datadir/#innodb_redo /datadir/#innodb_temp /datadir/auth_plugin /datadir/auto.cnf /datadir/binlog.000001 /datadir/binlog.000002 /datadir/binlog.000003 /datadir/binlog.index /datadir/galera.cache /datadir/get-pxc-state /datadir/grastate.dat /datadir/ib_buffer_pool /datadir/ibdata1 /datadir/innobackup.backup.log /datadir/liveness-check.sh /datadir/myApp /datadir/mysql /datadir/mysql-state-monitor /datadir/mysql-state-monitor.log /datadir/mysql.ibd /datadir/mysql.state /datadir/notify.sock /datadir/peer-list /datadir/performance_schema /datadir/pmm-prerun.sh /datadir/private_key.pem /datadir/public_key.pem /datadir/pxc-configure-pxc.sh /datadir/pxc-entrypoint.sh /datadir/readiness-check.sh /datadir/sys /datadir/undo_001 /datadir/undo_002 /datadir/version_info /datadir/wsrep_cmd_notify_handler.sh ++ mktemp --directory /datadir/pxc_sst_XXXX + tmp=/datadir/pxc_sst_e2M7 + socat -u 'openssl-connect:restore-src-on-demand-backup-pvc-demand-backup:3307,reuseaddr,cert=/etc/mysql/ssl-internal/tls.crt,key=/etc/mysql/ssl-internal/tls.key,cafile=/etc/mysql/ssl-internal/ca.crt,verify=1,commonname='\'''\'',retry=30,no-sni=1' stdio ++ parse_ini mysql-version /datadir/pxc_sst_e2M7/sst_info ++ local key=mysql-version ++ local file_path=/datadir/pxc_sst_e2M7/sst_info ++ awk -F '=[ ]*' '/mysql-version[ ]*=/ {print $2}' /datadir/pxc_sst_e2M7/sst_info + MYSQL_VERSION=8.0.42-33.1 + check_for_version 8.0.42-33.1 8.0.0 + '[' -z 8.0.42-33.1 ']' + '[' -z 8.0.0 ']' + local local_version_str + local required_version_str ++ normalize_version 8.0.42-33.1 ++ local major=0 ++ local minor=0 ++ local patch=0 ++ [[ 8.0.42-33.1 =~ ^([0-9]+)\.([0-9]+)\.?([0-9]*)([^ ])* ]] ++ major=8 ++ minor=0 ++ patch=42 ++ printf %02d%02d%02d 8 0 42 + local_version_str=080042 ++ normalize_version 8.0.0 ++ local major=0 ++ local minor=0 ++ local patch=0 ++ [[ 8.0.0 =~ ^([0-9]+)\.([0-9]+)\.?([0-9]*)([^ ])* ]] ++ major=8 ++ minor=0 ++ patch=0 ++ printf %02d%02d%02d 8 0 0 + required_version_str=080000 + [[ 080042 < 080000 ]] + return 0 + XBSTREAM_EXTRA_ARGS=' --decompress' + socat -u 'openssl-connect:restore-src-on-demand-backup-pvc-demand-backup:3307,reuseaddr,cert=/etc/mysql/ssl-internal/tls.crt,key=/etc/mysql/ssl-internal/tls.key,cafile=/etc/mysql/ssl-internal/ca.crt,verify=1,commonname='\'''\'',retry=30,no-sni=1' stdio ++ grep -c processor /proc/cpuinfo + xbstream -x -C /datadir/pxc_sst_e2M7 --parallel=4 --decompress + set +o xtrace vault configuration not found + xtrabackup --use-memory=2GB --prepare --binlog-info=ON --rollback-prepared-trx --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_e2M7 2025-10-01T02:00:32.020074-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized server arguments: --innodb_checksum_algorithm=crc32 --innodb_log_checksums=1 --innodb_data_file_path=ibdata1:12M:autoextend --innodb_log_file_size=50331648 --innodb_page_size=16384 --innodb_undo_directory=./ --innodb_undo_tablespaces=2 --server-id=21832152 --innodb_log_checksums=ON --innodb_redo_log_encrypt=0 --innodb_undo_log_encrypt=0 2025-10-01T02:00:32.020161-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized client arguments: --use-memory=2GB --prepare=1 --rollback-prepared-trx=1 --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_e2M7 xtrabackup version 8.0.35-33 based on MySQL server 8.0.35 Linux (x86_64) (revision id: a982afdd) 2025-10-01T02:00:32.020195-00:00 0 [Note] [MY-011825] [Xtrabackup] cd to /datadir/pxc_sst_e2M7/ 2025-10-01T02:00:32.020258-00:00 0 [Note] [MY-011825] [Xtrabackup] This target seems to be not prepared yet. 2025-10-01T02:00:32.029560-00:00 0 [Note] [MY-011825] [Xtrabackup] xtrabackup_logfile detected: size=8388608, start_lsn=(30285593) 2025-10-01T02:00:32.030319-00:00 0 [Note] [MY-011825] [Xtrabackup] using the following InnoDB configuration for recovery: 2025-10-01T02:00:32.030332-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_home_dir = . 2025-10-01T02:00:32.030339-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_file_path = ibdata1:12M:autoextend 2025-10-01T02:00:32.030364-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_group_home_dir = . 2025-10-01T02:00:32.030395-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_files_in_group = 1 2025-10-01T02:00:32.030405-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_file_size = 8388608 2025-10-01T02:00:32.030584-00:00 0 [Note] [MY-011825] [Xtrabackup] inititialize_service_handles suceeded 2025-10-01T02:00:32.030747-00:00 0 [Note] [MY-011825] [Xtrabackup] using the following InnoDB configuration for recovery: 2025-10-01T02:00:32.030758-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_home_dir = . 2025-10-01T02:00:32.030764-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_file_path = ibdata1:12M:autoextend 2025-10-01T02:00:32.030773-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_group_home_dir = . 2025-10-01T02:00:32.030778-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_files_in_group = 1 2025-10-01T02:00:32.030784-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_file_size = 8388608 2025-10-01T02:00:32.030794-00:00 0 [Note] [MY-011825] [Xtrabackup] Starting InnoDB instance for recovery. 2025-10-01T02:00:32.030803-00:00 0 [Note] [MY-011825] [Xtrabackup] Using 2147483648 bytes for buffer pool (set by --use-memory parameter) 2025-10-01T02:00:32.030830-00:00 0 [Note] [MY-012932] [InnoDB] PUNCH HOLE support available 2025-10-01T02:00:32.030844-00:00 0 [Note] [MY-012944] [InnoDB] Uses event mutexes 2025-10-01T02:00:32.030850-00:00 0 [Note] [MY-012945] [InnoDB] GCC builtin __atomic_thread_fence() is used for memory barrier 2025-10-01T02:00:32.030856-00:00 0 [Note] [MY-012948] [InnoDB] Compressed tables use zlib 1.2.13 2025-10-01T02:00:32.031010-00:00 0 [Note] [MY-012951] [InnoDB] Using hardware accelerated crc32 and polynomial multiplication. 2025-10-01T02:00:32.031342-00:00 0 [Note] [MY-012203] [InnoDB] Directories to scan './' 2025-10-01T02:00:32.031382-00:00 0 [Note] [MY-012204] [InnoDB] Scanning './' 2025-10-01T02:00:32.040629-00:00 0 [Note] [MY-012208] [InnoDB] Completed space ID check of 8 files. 2025-10-01T02:00:32.041178-00:00 0 [Note] [MY-012955] [InnoDB] Initializing buffer pool, total size = 2.000000G, instances = 1, chunk size =128.000000M 2025-10-01T02:00:32.142157-00:00 0 [Note] [MY-012957] [InnoDB] Completed initialization of buffer pool 2025-10-01T02:00:32.150785-00:00 0 [Note] [MY-011952] [InnoDB] If the mysqld execution user is authorized, page cleaner thread priority can be changed. See the man page of setpriority(). 2025-10-01T02:00:32.179216-00:00 0 [Note] [MY-013883] [InnoDB] The latest found checkpoint is at lsn = 30285593 in redo log file ./#innodb_redo/#ib_redo0. 2025-10-01T02:00:32.179255-00:00 0 [Note] [MY-012560] [InnoDB] The log sequence number 30166951 in the system tablespace does not match the log sequence number 30285593 in the redo log files! 2025-10-01T02:00:32.179260-00:00 0 [Note] [MY-012551] [InnoDB] Database was not shutdown normally! 2025-10-01T02:00:32.179265-00:00 0 [Note] [MY-012552] [InnoDB] Starting crash recovery. 2025-10-01T02:00:32.179395-00:00 0 [Note] [MY-013086] [InnoDB] Starting to parse redo log at lsn = 30285424, whereas checkpoint_lsn = 30285593 and start_lsn = 30285312 2025-10-01T02:00:32.179406-00:00 0 [Note] [MY-012550] [InnoDB] Doing recovery: scanned up to log sequence number 30285603 2025-10-01T02:00:32.199353-00:00 0 [Note] [MY-013083] [InnoDB] Log background threads are being started... 2025-10-01T02:00:32.200974-00:00 0 [Note] [MY-012532] [InnoDB] Applying a batch of 1 redo log records ... 2025-10-01T02:00:32.201023-00:00 0 [Note] [MY-012533] [InnoDB] 100% 2025-10-01T02:00:32.201064-00:00 0 [Note] [MY-012535] [InnoDB] Apply batch completed! 2025-10-01T02:00:32.304399-00:00 0 [Note] [MY-013084] [InnoDB] Log background threads are being closed... 2025-10-01T02:00:32.305862-00:00 0 [Note] [MY-013888] [InnoDB] Upgrading redo log: 1032M, LSN=30285603. 2025-10-01T02:00:32.305884-00:00 0 [Note] [MY-012968] [InnoDB] Starting to delete and rewrite redo log files. 2025-10-01T02:00:32.305935-00:00 0 [Note] [MY-011825] [InnoDB] Removing redo log file: ./#innodb_redo/#ib_redo0 2025-10-01T02:00:32.334890-00:00 0 [Note] [MY-011825] [InnoDB] Creating redo log file at ./#innodb_redo/#ib_redo0_tmp with file_id 0 with size 33554432 bytes 2025-10-01T02:00:32.340767-00:00 0 [Note] [MY-011825] [InnoDB] Renaming redo log file from ./#innodb_redo/#ib_redo0_tmp to ./#innodb_redo/#ib_redo0 2025-10-01T02:00:32.346124-00:00 0 [Note] [MY-012893] [InnoDB] New redo log files created, LSN=30285836 2025-10-01T02:00:32.346179-00:00 0 [Note] [MY-013083] [InnoDB] Log background threads are being started... 2025-10-01T02:00:32.346526-00:00 0 [Note] [MY-013252] [InnoDB] Using undo tablespace './undo_001'. 2025-10-01T02:00:32.347779-00:00 0 [Note] [MY-013252] [InnoDB] Using undo tablespace './undo_002'. 2025-10-01T02:00:32.349388-00:00 0 [Note] [MY-012910] [InnoDB] Opened 2 existing undo tablespaces. 2025-10-01T02:00:32.349439-00:00 0 [Note] [MY-011980] [InnoDB] GTID recovery trx_no: 5200 2025-10-01T02:00:32.488550-00:00 0 [Note] [MY-013776] [InnoDB] Parallel initialization of rseg complete 2025-10-01T02:00:32.488564-00:00 0 [Note] [MY-013777] [InnoDB] Time taken to initialize rseg using 4 thread: 139130 ms. 2025-10-01T02:00:32.488609-00:00 0 [Note] [MY-012923] [InnoDB] Creating shared tablespace for temporary tables 2025-10-01T02:00:32.488653-00:00 0 [Note] [MY-012265] [InnoDB] Setting file './ibtmp1' size to 12 MB. Physically writing the file full; Please wait ... 2025-10-01T02:00:32.519397-00:00 0 [Note] [MY-012266] [InnoDB] File './ibtmp1' size is now 12 MB. 2025-10-01T02:00:32.519505-00:00 0 [Note] [MY-013627] [InnoDB] Scanning temp tablespace dir:'./#innodb_temp/' 2025-10-01T02:00:32.553849-00:00 0 [Note] [MY-013018] [InnoDB] Created 128 and tracked 128 new rollback segment(s) in the temporary tablespace. 128 are now active. 2025-10-01T02:00:32.553996-00:00 0 [Note] [MY-012976] [InnoDB] 8.0.35 started; log sequence number 30285846 2025-10-01T02:00:32.555054-00:00 0 [Warning] [MY-012091] [InnoDB] Allocated tablespace ID 1 for sys/sys_config, old maximum was 0 2025-10-01T02:00:32.560449-00:00 0 [Note] [MY-011825] [Xtrabackup] Completed loading of 6 tablespaces into cache in 0.00643142 seconds 2025-10-01T02:00:32.580271-00:00 0 [Note] [MY-011825] [Xtrabackup] Time taken to build dictionary: 0.0198038 seconds 2025-10-01T02:00:33.580668-00:00 0 [Note] [MY-011825] [Xtrabackup] Recovered WSREP position: 4305a2a1-9e69-11f0-ba30-cff55bd2012b:39 2025-10-01T02:00:33.580710-00:00 0 [Note] [MY-011825] [Xtrabackup] starting shutdown with innodb_fast_shutdown = 1 2025-10-01T02:00:33.580757-00:00 0 [Note] [MY-012330] [InnoDB] FTS optimize thread exiting. 2025-10-01T02:00:34.580664-00:00 0 [Note] [MY-013072] [InnoDB] Starting shutdown... 2025-10-01T02:00:34.681569-00:00 0 [Note] [MY-013084] [InnoDB] Log background threads are being closed... 2025-10-01T02:00:34.713017-00:00 0 [Note] [MY-012980] [InnoDB] Shutdown completed; log sequence number 30285846 2025-10-01T02:00:34.715774-00:00 0 [Note] [MY-011825] [Xtrabackup] completed OK! + xtrabackup --defaults-group=mysqld --datadir=/datadir --move-back --binlog-info=ON --force-non-empty-directories --keyring-vault-config=/etc/mysql/vault-keyring-secret/keyring_vault.conf --early-plugin-load=keyring_vault.so --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_e2M7 2025-10-01T02:00:34.738018-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized server arguments: --defaults_group=mysqld --datadir=/datadir 2025-10-01T02:00:34.738085-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized client arguments: --move-back=1 --force-non-empty-directories=1 --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_e2M7 xtrabackup version 8.0.35-33 based on MySQL server 8.0.35 Linux (x86_64) (revision id: a982afdd) 2025-10-01T02:00:34.738118-00:00 0 [Note] [MY-011825] [Xtrabackup] cd to /datadir/pxc_sst_e2M7/ 2025-10-01T02:00:34.738578-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving undo_001 to /datadir/undo_001 2025-10-01T02:00:34.738612-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file undo_001 to /datadir/undo_001 2025-10-01T02:00:34.738643-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving undo_002 to /datadir/undo_002 2025-10-01T02:00:34.738670-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file undo_002 to /datadir/undo_002 2025-10-01T02:00:34.738830-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving ibdata1 to /datadir/ibdata1 2025-10-01T02:00:34.738853-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file ibdata1 to /datadir/ibdata1 2025-10-01T02:00:34.739084-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving binlog.000009 to /datadir//binlog.000009 2025-10-01T02:00:34.739109-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file binlog.000009 to /datadir//binlog.000009 2025-10-01T02:00:34.739186-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving binlog.index to /datadir//binlog.index 2025-10-01T02:00:34.739213-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file binlog.index to /datadir//binlog.index 2025-10-01T02:00:34.739554-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_135.sdi to /datadir/performance_schema/events_statement_135.sdi 2025-10-01T02:00:34.739598-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_135.sdi to /datadir/performance_schema/events_statement_135.sdi 2025-10-01T02:00:34.739634-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_150.sdi to /datadir/performance_schema/events_errors_su_150.sdi 2025-10-01T02:00:34.739663-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_150.sdi to /datadir/performance_schema/events_errors_su_150.sdi 2025-10-01T02:00:34.739709-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_132.sdi to /datadir/performance_schema/events_statement_132.sdi 2025-10-01T02:00:34.739737-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_132.sdi to /datadir/performance_schema/events_statement_132.sdi 2025-10-01T02:00:34.739768-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_126.sdi to /datadir/performance_schema/events_stages_su_126.sdi 2025-10-01T02:00:34.739818-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_126.sdi to /datadir/performance_schema/events_stages_su_126.sdi 2025-10-01T02:00:34.739851-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_131.sdi to /datadir/performance_schema/events_statement_131.sdi 2025-10-01T02:00:34.739890-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_131.sdi to /datadir/performance_schema/events_statement_131.sdi 2025-10-01T02:00:34.739918-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_164.sdi to /datadir/performance_schema/memory_summary_b_164.sdi 2025-10-01T02:00:34.739948-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_164.sdi to /datadir/performance_schema/memory_summary_b_164.sdi 2025-10-01T02:00:34.739993-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_137.sdi to /datadir/performance_schema/events_statement_137.sdi 2025-10-01T02:00:34.740024-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_137.sdi to /datadir/performance_schema/events_statement_137.sdi 2025-10-01T02:00:34.740057-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_176.sdi to /datadir/performance_schema/replication_appl_176.sdi 2025-10-01T02:00:34.740087-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_176.sdi to /datadir/performance_schema/replication_appl_176.sdi 2025-10-01T02:00:34.740141-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_connect__159.sdi to /datadir/performance_schema/session_connect__159.sdi 2025-10-01T02:00:34.740171-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_connect__159.sdi to /datadir/performance_schema/session_connect__159.sdi 2025-10-01T02:00:34.740200-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_97.sdi to /datadir/performance_schema/events_waits_sum_97.sdi 2025-10-01T02:00:34.740229-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_97.sdi to /datadir/performance_schema/events_waits_sum_97.sdi 2025-10-01T02:00:34.740254-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_101.sdi to /datadir/performance_schema/events_waits_sum_101.sdi 2025-10-01T02:00:34.740286-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_101.sdi to /datadir/performance_schema/events_waits_sum_101.sdi 2025-10-01T02:00:34.740314-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_summary_by__103.sdi to /datadir/performance_schema/file_summary_by__103.sdi 2025-10-01T02:00:34.740353-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_summary_by__103.sdi to /datadir/performance_schema/file_summary_by__103.sdi 2025-10-01T02:00:34.740385-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_163.sdi to /datadir/performance_schema/memory_summary_b_163.sdi 2025-10-01T02:00:34.740412-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_163.sdi to /datadir/performance_schema/memory_summary_b_163.sdi 2025-10-01T02:00:34.740437-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/global_variables_193.sdi to /datadir/performance_schema/global_variables_193.sdi 2025-10-01T02:00:34.740462-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/global_variables_193.sdi to /datadir/performance_schema/global_variables_193.sdi 2025-10-01T02:00:34.740511-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/metadata_locks_168.sdi to /datadir/performance_schema/metadata_locks_168.sdi 2025-10-01T02:00:34.740537-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/metadata_locks_168.sdi to /datadir/performance_schema/metadata_locks_168.sdi 2025-10-01T02:00:34.740580-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_his_94.sdi to /datadir/performance_schema/events_waits_his_94.sdi 2025-10-01T02:00:34.740608-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_his_94.sdi to /datadir/performance_schema/events_waits_his_94.sdi 2025-10-01T02:00:34.740638-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/log_status_183.sdi to /datadir/performance_schema/log_status_183.sdi 2025-10-01T02:00:34.740663-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/log_status_183.sdi to /datadir/performance_schema/log_status_183.sdi 2025-10-01T02:00:34.740693-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_147.sdi to /datadir/performance_schema/events_transacti_147.sdi 2025-10-01T02:00:34.740718-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_147.sdi to /datadir/performance_schema/events_transacti_147.sdi 2025-10-01T02:00:34.740753-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_125.sdi to /datadir/performance_schema/events_stages_su_125.sdi 2025-10-01T02:00:34.740781-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_125.sdi to /datadir/performance_schema/events_stages_su_125.sdi 2025-10-01T02:00:34.740808-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_133.sdi to /datadir/performance_schema/events_statement_133.sdi 2025-10-01T02:00:34.740836-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_133.sdi to /datadir/performance_schema/events_statement_133.sdi 2025-10-01T02:00:34.740863-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_96.sdi to /datadir/performance_schema/events_waits_sum_96.sdi 2025-10-01T02:00:34.740905-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_96.sdi to /datadir/performance_schema/events_waits_sum_96.sdi 2025-10-01T02:00:34.740932-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/malloc_stats_tot_200.sdi to /datadir/performance_schema/malloc_stats_tot_200.sdi 2025-10-01T02:00:34.740957-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/malloc_stats_tot_200.sdi to /datadir/performance_schema/malloc_stats_tot_200.sdi 2025-10-01T02:00:34.740984-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/keyring_keys_161.sdi to /datadir/performance_schema/keyring_keys_161.sdi 2025-10-01T02:00:34.741005-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/keyring_keys_161.sdi to /datadir/performance_schema/keyring_keys_161.sdi 2025-10-01T02:00:34.741033-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_146.sdi to /datadir/performance_schema/events_transacti_146.sdi 2025-10-01T02:00:34.741063-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_146.sdi to /datadir/performance_schema/events_transacti_146.sdi 2025-10-01T02:00:34.741093-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_127.sdi to /datadir/performance_schema/events_stages_su_127.sdi 2025-10-01T02:00:34.741120-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_127.sdi to /datadir/performance_schema/events_stages_su_127.sdi 2025-10-01T02:00:34.741175-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_asyn_181.sdi to /datadir/performance_schema/replication_asyn_181.sdi 2025-10-01T02:00:34.741203-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_asyn_181.sdi to /datadir/performance_schema/replication_asyn_181.sdi 2025-10-01T02:00:34.741235-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_variable_194.sdi to /datadir/performance_schema/session_variable_194.sdi 2025-10-01T02:00:34.741265-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_variable_194.sdi to /datadir/performance_schema/session_variable_194.sdi 2025-10-01T02:00:34.741296-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/variables_by_thr_192.sdi to /datadir/performance_schema/variables_by_thr_192.sdi 2025-10-01T02:00:34.741323-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/variables_by_thr_192.sdi to /datadir/performance_schema/variables_by_thr_192.sdi 2025-10-01T02:00:34.741348-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/processlist_109.sdi to /datadir/performance_schema/processlist_109.sdi 2025-10-01T02:00:34.741368-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/processlist_109.sdi to /datadir/performance_schema/processlist_109.sdi 2025-10-01T02:00:34.741393-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_179.sdi to /datadir/performance_schema/replication_appl_179.sdi 2025-10-01T02:00:34.741421-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_179.sdi to /datadir/performance_schema/replication_appl_179.sdi 2025-10-01T02:00:34.741449-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/global_status_190.sdi to /datadir/performance_schema/global_status_190.sdi 2025-10-01T02:00:34.741472-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/global_status_190.sdi to /datadir/performance_schema/global_status_190.sdi 2025-10-01T02:00:34.741502-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_threads_115.sdi to /datadir/performance_schema/setup_threads_115.sdi 2025-10-01T02:00:34.741523-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_threads_115.sdi to /datadir/performance_schema/setup_threads_115.sdi 2025-10-01T02:00:34.741541-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_130.sdi to /datadir/performance_schema/events_statement_130.sdi 2025-10-01T02:00:34.741558-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_130.sdi to /datadir/performance_schema/events_statement_130.sdi 2025-10-01T02:00:34.741587-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/persisted_variab_196.sdi to /datadir/performance_schema/persisted_variab_196.sdi 2025-10-01T02:00:34.741615-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/persisted_variab_196.sdi to /datadir/performance_schema/persisted_variab_196.sdi 2025-10-01T02:00:34.741646-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/performance_time_108.sdi to /datadir/performance_schema/performance_time_108.sdi 2025-10-01T02:00:34.741671-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/performance_time_108.sdi to /datadir/performance_schema/performance_time_108.sdi 2025-10-01T02:00:34.741702-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_conn_173.sdi to /datadir/performance_schema/replication_conn_173.sdi 2025-10-01T02:00:34.741731-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_conn_173.sdi to /datadir/performance_schema/replication_conn_173.sdi 2025-10-01T02:00:34.741764-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_g_162.sdi to /datadir/performance_schema/memory_summary_g_162.sdi 2025-10-01T02:00:34.741789-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_g_162.sdi to /datadir/performance_schema/memory_summary_g_162.sdi 2025-10-01T02:00:34.741818-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_actors_111.sdi to /datadir/performance_schema/setup_actors_111.sdi 2025-10-01T02:00:34.741843-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_actors_111.sdi to /datadir/performance_schema/setup_actors_111.sdi 2025-10-01T02:00:34.741882-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_98.sdi to /datadir/performance_schema/events_waits_sum_98.sdi 2025-10-01T02:00:34.741908-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_98.sdi to /datadir/performance_schema/events_waits_sum_98.sdi 2025-10-01T02:00:34.741942-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_consumers_112.sdi to /datadir/performance_schema/setup_consumers_112.sdi 2025-10-01T02:00:34.741967-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_consumers_112.sdi to /datadir/performance_schema/setup_consumers_112.sdi 2025-10-01T02:00:34.741994-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_io_waits_s_117.sdi to /datadir/performance_schema/table_io_waits_s_117.sdi 2025-10-01T02:00:34.742020-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_io_waits_s_117.sdi to /datadir/performance_schema/table_io_waits_s_117.sdi 2025-10-01T02:00:34.742048-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_124.sdi to /datadir/performance_schema/events_stages_su_124.sdi 2025-10-01T02:00:34.742076-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_124.sdi to /datadir/performance_schema/events_stages_su_124.sdi 2025-10-01T02:00:34.742103-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/error_log_92.sdi to /datadir/performance_schema/error_log_92.sdi 2025-10-01T02:00:34.742127-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/error_log_92.sdi to /datadir/performance_schema/error_log_92.sdi 2025-10-01T02:00:34.742156-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_status_191.sdi to /datadir/performance_schema/session_status_191.sdi 2025-10-01T02:00:34.742183-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_status_191.sdi to /datadir/performance_schema/session_status_191.sdi 2025-10-01T02:00:34.742211-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_instances_156.sdi to /datadir/performance_schema/socket_instances_156.sdi 2025-10-01T02:00:34.742239-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_instances_156.sdi to /datadir/performance_schema/socket_instances_156.sdi 2025-10-01T02:00:34.742269-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/pxc_cluster_view_203.sdi to /datadir/performance_schema/pxc_cluster_view_203.sdi 2025-10-01T02:00:34.742300-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/pxc_cluster_view_203.sdi to /datadir/performance_schema/pxc_cluster_view_203.sdi 2025-10-01T02:00:34.742320-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/rwlock_instances_110.sdi to /datadir/performance_schema/rwlock_instances_110.sdi 2025-10-01T02:00:34.742348-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/rwlock_instances_110.sdi to /datadir/performance_schema/rwlock_instances_110.sdi 2025-10-01T02:00:34.742371-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_cur_93.sdi to /datadir/performance_schema/events_waits_cur_93.sdi 2025-10-01T02:00:34.742394-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_cur_93.sdi to /datadir/performance_schema/events_waits_cur_93.sdi 2025-10-01T02:00:34.742420-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_io_waits_s_116.sdi to /datadir/performance_schema/table_io_waits_s_116.sdi 2025-10-01T02:00:34.742446-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_io_waits_s_116.sdi to /datadir/performance_schema/table_io_waits_s_116.sdi 2025-10-01T02:00:34.742478-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_136.sdi to /datadir/performance_schema/events_statement_136.sdi 2025-10-01T02:00:34.742509-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_136.sdi to /datadir/performance_schema/events_statement_136.sdi 2025-10-01T02:00:34.742538-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/user_variables_b_185.sdi to /datadir/performance_schema/user_variables_b_185.sdi 2025-10-01T02:00:34.742566-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/user_variables_b_185.sdi to /datadir/performance_schema/user_variables_b_185.sdi 2025-10-01T02:00:34.742596-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_177.sdi to /datadir/performance_schema/replication_appl_177.sdi 2025-10-01T02:00:34.742626-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_177.sdi to /datadir/performance_schema/replication_appl_177.sdi 2025-10-01T02:00:34.742653-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_account__160.sdi to /datadir/performance_schema/session_account__160.sdi 2025-10-01T02:00:34.742697-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_account__160.sdi to /datadir/performance_schema/session_account__160.sdi 2025-10-01T02:00:34.742742-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_141.sdi to /datadir/performance_schema/events_transacti_141.sdi 2025-10-01T02:00:34.742771-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_141.sdi to /datadir/performance_schema/events_transacti_141.sdi 2025-10-01T02:00:34.742799-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_144.sdi to /datadir/performance_schema/events_transacti_144.sdi 2025-10-01T02:00:34.742824-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_144.sdi to /datadir/performance_schema/events_transacti_144.sdi 2025-10-01T02:00:34.742849-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_hi_121.sdi to /datadir/performance_schema/events_stages_hi_121.sdi 2025-10-01T02:00:34.742891-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_hi_121.sdi to /datadir/performance_schema/events_stages_hi_121.sdi 2025-10-01T02:00:34.742924-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/malloc_stats_201.sdi to /datadir/performance_schema/malloc_stats_201.sdi 2025-10-01T02:00:34.742953-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/malloc_stats_201.sdi to /datadir/performance_schema/malloc_stats_201.sdi 2025-10-01T02:00:34.742985-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_grou_172.sdi to /datadir/performance_schema/replication_grou_172.sdi 2025-10-01T02:00:34.743016-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_grou_172.sdi to /datadir/performance_schema/replication_grou_172.sdi 2025-10-01T02:00:34.743046-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_139.sdi to /datadir/performance_schema/events_statement_139.sdi 2025-10-01T02:00:34.743075-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_139.sdi to /datadir/performance_schema/events_statement_139.sdi 2025-10-01T02:00:34.743100-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/users_153.sdi to /datadir/performance_schema/users_153.sdi 2025-10-01T02:00:34.743125-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/users_153.sdi to /datadir/performance_schema/users_153.sdi 2025-10-01T02:00:34.743154-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_his_95.sdi to /datadir/performance_schema/events_waits_his_95.sdi 2025-10-01T02:00:34.743182-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_his_95.sdi to /datadir/performance_schema/events_waits_his_95.sdi 2025-10-01T02:00:34.743211-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/threads_119.sdi to /datadir/performance_schema/threads_119.sdi 2025-10-01T02:00:34.743235-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/threads_119.sdi to /datadir/performance_schema/threads_119.sdi 2025-10-01T02:00:34.743265-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_123.sdi to /datadir/performance_schema/events_stages_su_123.sdi 2025-10-01T02:00:34.743291-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_123.sdi to /datadir/performance_schema/events_stages_su_123.sdi 2025-10-01T02:00:34.743319-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_objects_114.sdi to /datadir/performance_schema/setup_objects_114.sdi 2025-10-01T02:00:34.743362-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_objects_114.sdi to /datadir/performance_schema/setup_objects_114.sdi 2025-10-01T02:00:34.743393-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_145.sdi to /datadir/performance_schema/events_transacti_145.sdi 2025-10-01T02:00:34.743420-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_145.sdi to /datadir/performance_schema/events_transacti_145.sdi 2025-10-01T02:00:34.743448-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/data_lock_waits_170.sdi to /datadir/performance_schema/data_lock_waits_170.sdi 2025-10-01T02:00:34.743466-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/data_lock_waits_170.sdi to /datadir/performance_schema/data_lock_waits_170.sdi 2025-10-01T02:00:34.743498-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_host_187.sdi to /datadir/performance_schema/status_by_host_187.sdi 2025-10-01T02:00:34.743524-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_host_187.sdi to /datadir/performance_schema/status_by_host_187.sdi 2025-10-01T02:00:34.743571-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/binary_log_trans_198.sdi to /datadir/performance_schema/binary_log_trans_198.sdi 2025-10-01T02:00:34.743599-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/binary_log_trans_198.sdi to /datadir/performance_schema/binary_log_trans_198.sdi 2025-10-01T02:00:34.743625-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_151.sdi to /datadir/performance_schema/events_errors_su_151.sdi 2025-10-01T02:00:34.743652-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_151.sdi to /datadir/performance_schema/events_errors_su_151.sdi 2025-10-01T02:00:34.743681-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_lock_waits_118.sdi to /datadir/performance_schema/table_lock_waits_118.sdi 2025-10-01T02:00:34.743706-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_lock_waits_118.sdi to /datadir/performance_schema/table_lock_waits_118.sdi 2025-10-01T02:00:34.743734-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_summary_by__104.sdi to /datadir/performance_schema/file_summary_by__104.sdi 2025-10-01T02:00:34.743755-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_summary_by__104.sdi to /datadir/performance_schema/file_summary_by__104.sdi 2025-10-01T02:00:34.743779-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/keyring_componen_202.sdi to /datadir/performance_schema/keyring_componen_202.sdi 2025-10-01T02:00:34.743803-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/keyring_componen_202.sdi to /datadir/performance_schema/keyring_componen_202.sdi 2025-10-01T02:00:34.743833-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/variables_info_195.sdi to /datadir/performance_schema/variables_info_195.sdi 2025-10-01T02:00:34.743859-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/variables_info_195.sdi to /datadir/performance_schema/variables_info_195.sdi 2025-10-01T02:00:34.743903-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_accoun_186.sdi to /datadir/performance_schema/status_by_accoun_186.sdi 2025-10-01T02:00:34.743927-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_accoun_186.sdi to /datadir/performance_schema/status_by_accoun_186.sdi 2025-10-01T02:00:34.743958-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_165.sdi to /datadir/performance_schema/memory_summary_b_165.sdi 2025-10-01T02:00:34.743988-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_165.sdi to /datadir/performance_schema/memory_summary_b_165.sdi 2025-10-01T02:00:34.744017-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_conn_171.sdi to /datadir/performance_schema/replication_conn_171.sdi 2025-10-01T02:00:34.744045-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_conn_171.sdi to /datadir/performance_schema/replication_conn_171.sdi 2025-10-01T02:00:34.744074-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_grou_178.sdi to /datadir/performance_schema/replication_grou_178.sdi 2025-10-01T02:00:34.744102-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_grou_178.sdi to /datadir/performance_schema/replication_grou_178.sdi 2025-10-01T02:00:34.744131-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_142.sdi to /datadir/performance_schema/events_transacti_142.sdi 2025-10-01T02:00:34.744155-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_142.sdi to /datadir/performance_schema/events_transacti_142.sdi 2025-10-01T02:00:34.744182-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_hi_122.sdi to /datadir/performance_schema/events_stages_hi_122.sdi 2025-10-01T02:00:34.744210-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_hi_122.sdi to /datadir/performance_schema/events_stages_hi_122.sdi 2025-10-01T02:00:34.744239-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_149.sdi to /datadir/performance_schema/events_errors_su_149.sdi 2025-10-01T02:00:34.744267-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_149.sdi to /datadir/performance_schema/events_errors_su_149.sdi 2025-10-01T02:00:34.744293-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_thread_188.sdi to /datadir/performance_schema/status_by_thread_188.sdi 2025-10-01T02:00:34.744322-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_thread_188.sdi to /datadir/performance_schema/status_by_thread_188.sdi 2025-10-01T02:00:34.744355-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_143.sdi to /datadir/performance_schema/events_transacti_143.sdi 2025-10-01T02:00:34.744380-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_143.sdi to /datadir/performance_schema/events_transacti_143.sdi 2025-10-01T02:00:34.744411-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_166.sdi to /datadir/performance_schema/memory_summary_b_166.sdi 2025-10-01T02:00:34.744441-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_166.sdi to /datadir/performance_schema/memory_summary_b_166.sdi 2025-10-01T02:00:34.744471-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/mutex_instances_106.sdi to /datadir/performance_schema/mutex_instances_106.sdi 2025-10-01T02:00:34.744502-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/mutex_instances_106.sdi to /datadir/performance_schema/mutex_instances_106.sdi 2025-10-01T02:00:34.744527-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/prepared_stateme_184.sdi to /datadir/performance_schema/prepared_stateme_184.sdi 2025-10-01T02:00:34.744554-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/prepared_stateme_184.sdi to /datadir/performance_schema/prepared_stateme_184.sdi 2025-10-01T02:00:34.744584-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_174.sdi to /datadir/performance_schema/replication_appl_174.sdi 2025-10-01T02:00:34.744611-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_174.sdi to /datadir/performance_schema/replication_appl_174.sdi 2025-10-01T02:00:34.744636-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_175.sdi to /datadir/performance_schema/replication_appl_175.sdi 2025-10-01T02:00:34.744663-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_175.sdi to /datadir/performance_schema/replication_appl_175.sdi 2025-10-01T02:00:34.744691-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_instances_102.sdi to /datadir/performance_schema/file_instances_102.sdi 2025-10-01T02:00:34.744717-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_instances_102.sdi to /datadir/performance_schema/file_instances_102.sdi 2025-10-01T02:00:34.744768-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/hosts_155.sdi to /datadir/performance_schema/hosts_155.sdi 2025-10-01T02:00:34.744795-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/hosts_155.sdi to /datadir/performance_schema/hosts_155.sdi 2025-10-01T02:00:34.744824-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_handles_167.sdi to /datadir/performance_schema/table_handles_167.sdi 2025-10-01T02:00:34.744849-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_handles_167.sdi to /datadir/performance_schema/table_handles_167.sdi 2025-10-01T02:00:34.744882-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_99.sdi to /datadir/performance_schema/events_waits_sum_99.sdi 2025-10-01T02:00:34.744935-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_99.sdi to /datadir/performance_schema/events_waits_sum_99.sdi 2025-10-01T02:00:34.744964-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_129.sdi to /datadir/performance_schema/events_statement_129.sdi 2025-10-01T02:00:34.744994-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_129.sdi to /datadir/performance_schema/events_statement_129.sdi 2025-10-01T02:00:34.745021-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/cond_instances_91.sdi to /datadir/performance_schema/cond_instances_91.sdi 2025-10-01T02:00:34.745049-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/cond_instances_91.sdi to /datadir/performance_schema/cond_instances_91.sdi 2025-10-01T02:00:34.745082-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_148.sdi to /datadir/performance_schema/events_errors_su_148.sdi 2025-10-01T02:00:34.745112-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_148.sdi to /datadir/performance_schema/events_errors_su_148.sdi 2025-10-01T02:00:34.745141-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_152.sdi to /datadir/performance_schema/events_errors_su_152.sdi 2025-10-01T02:00:34.745171-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_152.sdi to /datadir/performance_schema/events_errors_su_152.sdi 2025-10-01T02:00:34.745197-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_summary_b_157.sdi to /datadir/performance_schema/socket_summary_b_157.sdi 2025-10-01T02:00:34.745223-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_summary_b_157.sdi to /datadir/performance_schema/socket_summary_b_157.sdi 2025-10-01T02:00:34.745255-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/data_locks_169.sdi to /datadir/performance_schema/data_locks_169.sdi 2025-10-01T02:00:34.745284-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/data_locks_169.sdi to /datadir/performance_schema/data_locks_169.sdi 2025-10-01T02:00:34.745314-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_summary_b_158.sdi to /datadir/performance_schema/socket_summary_b_158.sdi 2025-10-01T02:00:34.745343-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_summary_b_158.sdi to /datadir/performance_schema/socket_summary_b_158.sdi 2025-10-01T02:00:34.745372-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/host_cache_105.sdi to /datadir/performance_schema/host_cache_105.sdi 2025-10-01T02:00:34.745398-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/host_cache_105.sdi to /datadir/performance_schema/host_cache_105.sdi 2025-10-01T02:00:34.745425-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_cu_120.sdi to /datadir/performance_schema/events_stages_cu_120.sdi 2025-10-01T02:00:34.745453-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_cu_120.sdi to /datadir/performance_schema/events_stages_cu_120.sdi 2025-10-01T02:00:34.745484-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_134.sdi to /datadir/performance_schema/events_statement_134.sdi 2025-10-01T02:00:34.745513-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_134.sdi to /datadir/performance_schema/events_statement_134.sdi 2025-10-01T02:00:34.745540-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_128.sdi to /datadir/performance_schema/events_statement_128.sdi 2025-10-01T02:00:34.745571-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_128.sdi to /datadir/performance_schema/events_statement_128.sdi 2025-10-01T02:00:34.745598-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_instrument_113.sdi to /datadir/performance_schema/setup_instrument_113.sdi 2025-10-01T02:00:34.745625-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_instrument_113.sdi to /datadir/performance_schema/setup_instrument_113.sdi 2025-10-01T02:00:34.745673-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/tls_channel_stat_199.sdi to /datadir/performance_schema/tls_channel_stat_199.sdi 2025-10-01T02:00:34.745703-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/tls_channel_stat_199.sdi to /datadir/performance_schema/tls_channel_stat_199.sdi 2025-10-01T02:00:34.745729-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/accounts_154.sdi to /datadir/performance_schema/accounts_154.sdi 2025-10-01T02:00:34.745751-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/accounts_154.sdi to /datadir/performance_schema/accounts_154.sdi 2025-10-01T02:00:34.745774-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_asyn_182.sdi to /datadir/performance_schema/replication_asyn_182.sdi 2025-10-01T02:00:34.745798-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_asyn_182.sdi to /datadir/performance_schema/replication_asyn_182.sdi 2025-10-01T02:00:34.745825-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_140.sdi to /datadir/performance_schema/events_transacti_140.sdi 2025-10-01T02:00:34.745855-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_140.sdi to /datadir/performance_schema/events_transacti_140.sdi 2025-10-01T02:00:34.745896-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_180.sdi to /datadir/performance_schema/replication_appl_180.sdi 2025-10-01T02:00:34.745924-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_180.sdi to /datadir/performance_schema/replication_appl_180.sdi 2025-10-01T02:00:34.745955-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/user_defined_fun_197.sdi to /datadir/performance_schema/user_defined_fun_197.sdi 2025-10-01T02:00:34.745984-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/user_defined_fun_197.sdi to /datadir/performance_schema/user_defined_fun_197.sdi 2025-10-01T02:00:34.746013-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_100.sdi to /datadir/performance_schema/events_waits_sum_100.sdi 2025-10-01T02:00:34.746042-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_100.sdi to /datadir/performance_schema/events_waits_sum_100.sdi 2025-10-01T02:00:34.746069-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_user_189.sdi to /datadir/performance_schema/status_by_user_189.sdi 2025-10-01T02:00:34.746085-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_user_189.sdi to /datadir/performance_schema/status_by_user_189.sdi 2025-10-01T02:00:34.746110-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/objects_summary__107.sdi to /datadir/performance_schema/objects_summary__107.sdi 2025-10-01T02:00:34.746163-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/objects_summary__107.sdi to /datadir/performance_schema/objects_summary__107.sdi 2025-10-01T02:00:34.746195-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_138.sdi to /datadir/performance_schema/events_statement_138.sdi 2025-10-01T02:00:34.746225-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_138.sdi to /datadir/performance_schema/events_statement_138.sdi 2025-10-01T02:00:34.746256-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./xtrabackup_galera_info to /datadir/xtrabackup_galera_info 2025-10-01T02:00:34.746282-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./xtrabackup_galera_info to /datadir/xtrabackup_galera_info 2025-10-01T02:00:34.746354-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log.CSV to /datadir/mysql/slow_log.CSV 2025-10-01T02:00:34.746380-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log.CSV to /datadir/mysql/slow_log.CSV 2025-10-01T02:00:34.746409-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_cluster.ibd to /datadir/mysql/wsrep_cluster.ibd 2025-10-01T02:00:34.746433-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_cluster.ibd to /datadir/mysql/wsrep_cluster.ibd 2025-10-01T02:00:34.746462-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log.CSM to /datadir/mysql/general_log.CSM 2025-10-01T02:00:34.746489-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log.CSM to /datadir/mysql/general_log.CSM 2025-10-01T02:00:34.746515-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_cluster_members.ibd to /datadir/mysql/wsrep_cluster_members.ibd 2025-10-01T02:00:34.746540-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_cluster_members.ibd to /datadir/mysql/wsrep_cluster_members.ibd 2025-10-01T02:00:34.746572-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log_226.sdi to /datadir/mysql/slow_log_226.sdi 2025-10-01T02:00:34.746612-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log_226.sdi to /datadir/mysql/slow_log_226.sdi 2025-10-01T02:00:34.746638-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log.CSV to /datadir/mysql/general_log.CSV 2025-10-01T02:00:34.746665-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log.CSV to /datadir/mysql/general_log.CSV 2025-10-01T02:00:34.746696-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log_225.sdi to /datadir/mysql/general_log_225.sdi 2025-10-01T02:00:34.746720-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log_225.sdi to /datadir/mysql/general_log_225.sdi 2025-10-01T02:00:34.746749-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_streaming_log.ibd to /datadir/mysql/wsrep_streaming_log.ibd 2025-10-01T02:00:34.746774-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_streaming_log.ibd to /datadir/mysql/wsrep_streaming_log.ibd 2025-10-01T02:00:34.746802-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log.CSM to /datadir/mysql/slow_log.CSM 2025-10-01T02:00:34.746825-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log.CSM to /datadir/mysql/slow_log.CSM 2025-10-01T02:00:34.746852-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./xtrabackup_info to /datadir/xtrabackup_info 2025-10-01T02:00:34.746889-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./xtrabackup_info to /datadir/xtrabackup_info 2025-10-01T02:00:34.746919-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./ibtmp1 to /datadir/ibtmp1 2025-10-01T02:00:34.746940-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./ibtmp1 to /datadir/ibtmp1 2025-10-01T02:00:34.746967-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./ib_buffer_pool to /datadir/ib_buffer_pool 2025-10-01T02:00:34.746990-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./ib_buffer_pool to /datadir/ib_buffer_pool 2025-10-01T02:00:34.747015-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql.ibd to /datadir/mysql.ibd 2025-10-01T02:00:34.747039-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql.ibd to /datadir/mysql.ibd 2025-10-01T02:00:34.747047-00:00 1 [Note] [MY-011825] [Xtrabackup] Creating directory ./#innodb_redo 2025-10-01T02:00:34.747058-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: creating directory ./#innodb_redo 2025-10-01T02:00:34.747083-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./sst_info to /datadir/sst_info 2025-10-01T02:00:34.747106-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./sst_info to /datadir/sst_info 2025-10-01T02:00:34.747172-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./sys/sys_config.ibd to /datadir/sys/sys_config.ibd 2025-10-01T02:00:34.747201-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./sys/sys_config.ibd to /datadir/sys/sys_config.ibd 2025-10-01T02:00:34.747261-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./myApp/myApp.ibd to /datadir/myApp/myApp.ibd 2025-10-01T02:00:34.747289-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./myApp/myApp.ibd to /datadir/myApp/myApp.ibd 2025-10-01T02:00:34.840339-00:00 0 [Note] [MY-011825] [Xtrabackup] completed OK! + cat /tmp/tmp.mjCT1refbt Defaulted container "xtrabackup" out of: xtrabackup, backup-init (init) + rm /tmp/tmp.K2fVtpKacw /tmp/tmp.mjCT1refbt + return 0 + wait_for_running demand-backup-proxysql 1 + local name=demand-backup-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in $(seq 0 $last_pod) + wait_pod demand-backup-proxysql-0 480 + local pod=demand-backup-proxysql-0 + local max_retry=480 + local ns= ++ echo demand-backup-proxysql-0 ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container=proxysql + set +o xtrace pod/demand-backup-proxysql-0 condition met waiting for pod/demand-backup-proxysql-0 to become Ready.Ok + wait_for_running demand-backup-pxc 3 + local name=demand-backup-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in $(seq 0 $last_pod) + wait_pod demand-backup-pxc-0 480 + local pod=demand-backup-pxc-0 + local max_retry=480 + local ns= ++ echo demand-backup-pxc-0 ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container=pxc + set +o xtrace pod/demand-backup-pxc-0 condition met waiting for pod/demand-backup-pxc-0 to become Ready.Ok + for i in $(seq 0 $last_pod) + wait_pod demand-backup-pxc-1 480 + local pod=demand-backup-pxc-1 + local max_retry=480 + local ns= ++ echo demand-backup-pxc-1 ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container=pxc + set +o xtrace pod/demand-backup-pxc-1 condition met waiting for pod/demand-backup-pxc-1 to become Ready.Ok + for i in $(seq 0 $last_pod) + wait_pod demand-backup-pxc-2 480 + local pod=demand-backup-pxc-2 + local max_retry=480 + local ns= ++ echo demand-backup-pxc-2 ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container=pxc + set +o xtrace pod/demand-backup-pxc-2 condition met waiting for pod/demand-backup-pxc-2 to become Ready.Ok ++ kubectl_bin get pxc demand-backup -o 'jsonpath={.spec.secretsName}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Si6f1RPH2p +++ mktemp ++ local LAST_ERR=/tmp/tmp.nU0MMSS4Yv ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pxc demand-backup -o 'jsonpath={.spec.secretsName}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Si6f1RPH2p ++ cat /tmp/tmp.nU0MMSS4Yv ++ rm /tmp/tmp.Si6f1RPH2p /tmp/tmp.nU0MMSS4Yv ++ return 0 + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.GXXV6daV73 +++ mktemp ++ local LAST_ERR=/tmp/tmp.X7TtiUH3eW ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GXXV6daV73 ++ cat /tmp/tmp.X7TtiUH3eW ++ rm /tmp/tmp.GXXV6daV73 /tmp/tmp.X7TtiUH3eW ++ return 0 + local root_pass=root_password + sleep 35 + log 'check data after pxc-restore/on-demand-backup-pvc' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-10-01T02:06:30+0000]' check data after pxc-restore/on-demand-backup-pvc [2025-10-01T02:06:30+0000] check data after pxc-restore/on-demand-backup-pvc + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-0.demand-backup-pxc -uroot -p'\''root_password'\''' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-0.demand-backup-pxc -uroot -p'\''root_password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-0.demand-backup-pxc -uroot -p'\''root_password'\''' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-0.demand-backup-pxc -uroot -p'\''root_password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9KqnVI1ETV +++ mktemp ++ local LAST_ERR=/tmp/tmp.oZoWfzCxpU ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9KqnVI1ETV ++ cat /tmp/tmp.oZoWfzCxpU ++ rm /tmp/tmp.9KqnVI1ETV /tmp/tmp.oZoWfzCxpU ++ return 0 + client_pod=pxc-client-59944c5bbf-5pzc2 + wait_pod pxc-client-59944c5bbf-5pzc2 + local pod=pxc-client-59944c5bbf-5pzc2 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-5pzc2 ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container= + set +o xtrace pod/pxc-client-59944c5bbf-5pzc2 condition met waiting for pod/pxc-client-59944c5bbf-5pzc2 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.ZiSlkZxGqo/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/compare/select-1.sql /tmp/tmp.ZiSlkZxGqo/select-1.sql + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-1.demand-backup-pxc -uroot -p'\''root_password'\''' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-1.demand-backup-pxc -uroot -p'\''root_password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-1.demand-backup-pxc -uroot -p'\''root_password'\''' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-1.demand-backup-pxc -uroot -p'\''root_password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TyBZ1gNe7H +++ mktemp ++ local LAST_ERR=/tmp/tmp.FrkiHLBbuV ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.TyBZ1gNe7H ++ cat /tmp/tmp.FrkiHLBbuV ++ rm /tmp/tmp.TyBZ1gNe7H /tmp/tmp.FrkiHLBbuV ++ return 0 + client_pod=pxc-client-59944c5bbf-5pzc2 + wait_pod pxc-client-59944c5bbf-5pzc2 + local pod=pxc-client-59944c5bbf-5pzc2 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-5pzc2 ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container= + set +o xtrace pod/pxc-client-59944c5bbf-5pzc2 condition met waiting for pod/pxc-client-59944c5bbf-5pzc2 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.ZiSlkZxGqo/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/compare/select-1.sql /tmp/tmp.ZiSlkZxGqo/select-1.sql + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-2.demand-backup-pxc -uroot -p'\''root_password'\''' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-2.demand-backup-pxc -uroot -p'\''root_password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h demand-backup-pxc-2.demand-backup-pxc -uroot -p'\''root_password'\''' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h demand-backup-pxc-2.demand-backup-pxc -uroot -p'\''root_password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.L9lI2nPckH +++ mktemp ++ local LAST_ERR=/tmp/tmp.GpOTGL07Tx ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.L9lI2nPckH ++ cat /tmp/tmp.GpOTGL07Tx ++ rm /tmp/tmp.L9lI2nPckH /tmp/tmp.GpOTGL07Tx ++ return 0 + client_pod=pxc-client-59944c5bbf-5pzc2 + wait_pod pxc-client-59944c5bbf-5pzc2 + local pod=pxc-client-59944c5bbf-5pzc2 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-5pzc2 ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container= + set +o xtrace pod/pxc-client-59944c5bbf-5pzc2 condition met waiting for pod/pxc-client-59944c5bbf-5pzc2 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.ZiSlkZxGqo/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/e2e-tests/demand-backup/compare/select-1.sql /tmp/tmp.ZiSlkZxGqo/select-1.sql + '[' on-demand-backup-pvc '!=' on-demand-backup-minio ']' + log 'copy backup' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-10-01T02:06:47+0000]' copy backup [2025-10-01T02:06:47+0000] copy backup + '[' -n '' ']' + bash /mnt/jenkins/workspace/cloud-pxc-operator_PR-2200/deploy/backup/copy-backup.sh on-demand-backup-pvc /tmp/tmp.ZiSlkZxGqo/backup which: no xbcloud in (/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) No xtrabackup binaries found, please install them: https://www.percona.com/downloads/Percona-XtraBackup-LATEST https://formulae.brew.sh/formula/percona-xtrabackup