Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/logs/restore-to-encrypted-cluster-8-0.log grep: warning: stray \ before - Warning: version difference between client (1.34) and server (1.31) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.34) and server (1.31) exceeds the supported minor version skew of +/-1 + main + create_infra restore-to-encrypted-cluster-20222 + local ns=restore-to-encrypted-cluster-20222 + '[' -n pxc-operator ']' + kubectl get pxc --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch pxc -n restore-to-encrypted-cluster-13410 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.BZB1HJ8Q3m ++ mktemp + local LAST_ERR=/tmp/tmp.xzKkDYXdF1 + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.BZB1HJ8Q3m perconaxtradbcluster.pxc.percona.com "some-name" deleted from restore-to-encrypted-cluster-13410 namespace + cat /tmp/tmp.xzKkDYXdF1 + rm /tmp/tmp.BZB1HJ8Q3m /tmp/tmp.xzKkDYXdF1 + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.ttUh36PJWs ++ mktemp + local LAST_ERR=/tmp/tmp.XcDuFPiHHX + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ttUh36PJWs perconaxtradbclusterbackup.pxc.percona.com "on-demand-backup-aws-s3" deleted from restore-to-encrypted-cluster-13410 namespace perconaxtradbclusterbackup.pxc.percona.com "on-demand-backup-pvc" deleted from restore-to-encrypted-cluster-13410 namespace + cat /tmp/tmp.XcDuFPiHHX + rm /tmp/tmp.ttUh36PJWs /tmp/tmp.XcDuFPiHHX + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.Mvt3iiH9i7 ++ mktemp + local LAST_ERR=/tmp/tmp.42JSvm9L82 + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Mvt3iiH9i7 perconaxtradbclusterrestore.pxc.percona.com "on-demand-backup-pvc" deleted from restore-to-encrypted-cluster-13410 namespace + cat /tmp/tmp.42JSvm9L82 + rm /tmp/tmp.Mvt3iiH9i7 /tmp/tmp.42JSvm9L82 + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ sed s/NAMESPACE// ++ awk '-F ' '{print $2}' + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep validate-auth + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + egrep -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' ++ mktemp + awk '{print$1}' + '[' -n '' ']' + desc 'cleaned up old namespaces pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- + xargs kubectl delete ns + kubectl_bin delete namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.P5UExvrxlB egrep: warning: egrep is obsolescent; using grep -E ++ mktemp + local LAST_OUT=/tmp/tmp.jpcnO5XEDU ++ mktemp + local LAST_ERR=/tmp/tmp.1OSmqGeJyG + local exit_status=0 ++ seq 0 2 + local LAST_ERR=/tmp/tmp.XcTUeAnBuN + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get ns + for i in $(seq 0 2) + set +e + kubectl delete namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.P5UExvrxlB + cat /tmp/tmp.1OSmqGeJyG + rm /tmp/tmp.P5UExvrxlB /tmp/tmp.1OSmqGeJyG + return 0 namespace "restore-to-encrypted-cluster-13410" deleted namespace "vault-service-1-30354" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.jpcnO5XEDU namespace "pxc-operator" deleted + cat /tmp/tmp.XcTUeAnBuN + rm /tmp/tmp.jpcnO5XEDU /tmp/tmp.XcTUeAnBuN + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + echo -n 'waiting for namespace/pxc-operator to be deleted' waiting for namespace/pxc-operator to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pxc-operator" not found + desc 'create namespace pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.YqDKsI9thv ++ mktemp + local LAST_ERR=/tmp/tmp.Y8xUOYGtXq + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.YqDKsI9thv namespace/pxc-operator created + cat /tmp/tmp.Y8xUOYGtXq + rm /tmp/tmp.YqDKsI9thv /tmp/tmp.Y8xUOYGtXq + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.ndz1UmsS82 +++ mktemp ++ local LAST_ERR=/tmp/tmp.j1R0Gom4h8 ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ndz1UmsS82 ++ cat /tmp/tmp.j1R0Gom4h8 ++ rm /tmp/tmp.ndz1UmsS82 /tmp/tmp.j1R0Gom4h8 ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2203-07b4356f-1-cluster8 --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.Z7BUhXHBNK ++ mktemp + local LAST_ERR=/tmp/tmp.fsjeVcKSK1 + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2203-07b4356f-1-cluster8 --namespace=pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Z7BUhXHBNK Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2203-07b4356f-1-cluster8" modified. + cat /tmp/tmp.fsjeVcKSK1 + rm /tmp/tmp.Z7BUhXHBNK /tmp/tmp.fsjeVcKSK1 + return 0 + deploy_operator + desc 'start PXC operator' + set +o xtrace ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.SJmgvYEW9V ++ mktemp + local LAST_ERR=/tmp/tmp.LxXvMAFey1 + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.SJmgvYEW9V customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied + cat /tmp/tmp.LxXvMAFey1 + rm /tmp/tmp.SJmgvYEW9V /tmp/tmp.LxXvMAFey1 + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: pxc-operator^' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.a0KWK8Uu0u ++ mktemp + local LAST_ERR=/tmp/tmp.K5gYLf7oIG + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.a0KWK8Uu0u clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.K5gYLf7oIG + rm /tmp/tmp.a0KWK8Uu0u /tmp/tmp.K5gYLf7oIG + return 0 + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/deploy/cw-operator.yaml + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-2203-07b4356f^' + sed -e 's^failureThreshold: .*^failureThreshold: 10^' + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "DISABLE_TELEMETRY").value) = "true"' - + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "LOG_LEVEL").value) = "VERBOSE"' - + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.NCbX4htpZ5 ++ mktemp + local LAST_ERR=/tmp/tmp.s7pj1hIWer + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.NCbX4htpZ5 deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.s7pj1hIWer + rm /tmp/tmp.NCbX4htpZ5 /tmp/tmp.s7pj1hIWer + return 0 + sleep 10 + kubectl_bin wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s ++ mktemp + local LAST_OUT=/tmp/tmp.53hPB2OAUw ++ mktemp + local LAST_ERR=/tmp/tmp.np1GZwWaxg + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.53hPB2OAUw pod/percona-xtradb-cluster-operator-7db859f455-px77q condition met + cat /tmp/tmp.np1GZwWaxg + rm /tmp/tmp.53hPB2OAUw /tmp/tmp.np1GZwWaxg + return 0 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ grep -c percona-xtradb-cluster-operator +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.mjL3sE6SLR +++ mktemp ++ local LAST_ERR=/tmp/tmp.1qpZRlthFU ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.mjL3sE6SLR ++ cat /tmp/tmp.1qpZRlthFU ++ rm /tmp/tmp.mjL3sE6SLR /tmp/tmp.1qpZRlthFU ++ return 0 + wait_pod percona-xtradb-cluster-operator-7db859f455-px77q 480 pxc-operator + local pod=percona-xtradb-cluster-operator-7db859f455-px77q + local max_retry=480 + local ns=pxc-operator ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo percona-xtradb-cluster-operator-7db859f455-px77q ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-7db859f455-px77q condition met waiting for pod/percona-xtradb-cluster-operator-7db859f455-px77q to become Ready.Ok + sleep 3 + create_namespace restore-to-encrypted-cluster-20222 + local namespace=restore-to-encrypted-cluster-20222 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + egrep -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + awk '{print$1}' + '[' -n '' ']' + desc 'cleaned up old namespaces restore-to-encrypted-cluster-20222' + set +o xtrace + xargs kubectl delete ns ----------------------------------------------------------------------------------- cleaned up old namespaces restore-to-encrypted-cluster-20222 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace restore-to-encrypted-cluster-20222 ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.ZI1R3sp4KI egrep: warning: egrep is obsolescent; using grep -E ++ mktemp + local LAST_OUT=/tmp/tmp.UKRYragzcI + local LAST_ERR=/tmp/tmp.J32h5Ex9sM + local exit_status=0 ++ seq 0 2 ++ mktemp + for i in $(seq 0 2) + set +e + kubectl get ns + local LAST_ERR=/tmp/tmp.BvNucKjvy1 + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete namespace restore-to-encrypted-cluster-20222 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ZI1R3sp4KI + cat /tmp/tmp.J32h5Ex9sM + rm /tmp/tmp.ZI1R3sp4KI /tmp/tmp.J32h5Ex9sM + return 0 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in $(seq 0 2) + set +e + kubectl delete namespace restore-to-encrypted-cluster-20222 error: resource(s) were provided, but no name was specified + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in $(seq 0 2) + set +e + kubectl delete namespace restore-to-encrypted-cluster-20222 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.UKRYragzcI + cat /tmp/tmp.BvNucKjvy1 Error from server (NotFound): namespaces "restore-to-encrypted-cluster-20222" not found + rm /tmp/tmp.UKRYragzcI /tmp/tmp.BvNucKjvy1 + return 1 + : + wait_for_delete namespace/restore-to-encrypted-cluster-20222 + local res=namespace/restore-to-encrypted-cluster-20222 + echo -n 'waiting for namespace/restore-to-encrypted-cluster-20222 to be deleted' waiting for namespace/restore-to-encrypted-cluster-20222 to be deleted+ set +o xtrace Error from server (NotFound): namespaces "restore-to-encrypted-cluster-20222" not found + desc 'create namespace restore-to-encrypted-cluster-20222' + set +o xtrace ----------------------------------------------------------------------------------- create namespace restore-to-encrypted-cluster-20222 ----------------------------------------------------------------------------------- + kubectl_bin create namespace restore-to-encrypted-cluster-20222 ++ mktemp + local LAST_OUT=/tmp/tmp.3LvPLRfP81 ++ mktemp + local LAST_ERR=/tmp/tmp.eXJWuqvSfe + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace restore-to-encrypted-cluster-20222 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.3LvPLRfP81 namespace/restore-to-encrypted-cluster-20222 created + cat /tmp/tmp.eXJWuqvSfe + rm /tmp/tmp.3LvPLRfP81 /tmp/tmp.eXJWuqvSfe + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.DA7z765ciO +++ mktemp ++ local LAST_ERR=/tmp/tmp.sAVLs6m4cT ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DA7z765ciO ++ cat /tmp/tmp.sAVLs6m4cT ++ rm /tmp/tmp.DA7z765ciO /tmp/tmp.sAVLs6m4cT ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2203-07b4356f-1-cluster8 --namespace=restore-to-encrypted-cluster-20222 ++ mktemp + local LAST_OUT=/tmp/tmp.qd0zKEbhOS ++ mktemp + local LAST_ERR=/tmp/tmp.7NhDO5gLME + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2203-07b4356f-1-cluster8 --namespace=restore-to-encrypted-cluster-20222 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.qd0zKEbhOS Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2203-07b4356f-1-cluster8" modified. + cat /tmp/tmp.7NhDO5gLME + rm /tmp/tmp.qd0zKEbhOS /tmp/tmp.7NhDO5gLME + return 0 + apply_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.3pEFV4xDH6 ++ mktemp + local LAST_ERR=/tmp/tmp.OudAv2D6T8 + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.3pEFV4xDH6 secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.OudAv2D6T8 + rm /tmp/tmp.3pEFV4xDH6 /tmp/tmp.OudAv2D6T8 + return 0 + cluster=some-name + spinup_pxc some-name /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/conf/some-name.yml + local cluster=some-name + local config=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/conf/some-name.yml + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/conf/secrets.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/conf/client.yml + local port=3306 + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.tD6SOCvbgb ++ mktemp + local LAST_ERR=/tmp/tmp.kwvQ7Qhh7p + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/conf/secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.tD6SOCvbgb secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created + cat /tmp/tmp.kwvQ7Qhh7p + rm /tmp/tmp.tD6SOCvbgb /tmp/tmp.kwvQ7Qhh7p + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/conf/client.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/conf/client.yml + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/conf/client.yml + /usr/sbin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' ++ mktemp + /usr/sbin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/sbin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/sbin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2203-07b4356f#' + /usr/sbin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + local LAST_OUT=/tmp/tmp.5YIv2pFPtL + /usr/sbin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/sbin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/sbin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/sbin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/sbin/sed -e s~minio-service.#namespace~minio-service.restore-to-encrypted-cluster-20222~ ++ mktemp + /usr/sbin/sed -e 's#apply:.*#apply: Never#' + local LAST_ERR=/tmp/tmp.DRdomrXIca + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.5YIv2pFPtL deployment.apps/pxc-client created + cat /tmp/tmp.DRdomrXIca + rm /tmp/tmp.5YIv2pFPtL /tmp/tmp.DRdomrXIca + return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/conf/some-name.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/conf/some-name.yml + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/conf/some-name.yml + /usr/sbin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' ++ mktemp + /usr/sbin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/sbin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/sbin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2203-07b4356f#' + /usr/sbin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/sbin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + local LAST_OUT=/tmp/tmp.e1ynISHklG + /usr/sbin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' ++ mktemp + /usr/sbin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/sbin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/sbin/sed -e s~minio-service.#namespace~minio-service.restore-to-encrypted-cluster-20222~ + local LAST_ERR=/tmp/tmp.c0xJ2hD8qK + local exit_status=0 ++ seq 0 2 + /usr/sbin/sed -e 's#apply:.*#apply: Never#' + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.e1ynISHklG perconaxtradbcluster.pxc.percona.com/some-name created + cat /tmp/tmp.c0xJ2hD8qK + rm /tmp/tmp.e1ynISHklG /tmp/tmp.c0xJ2hD8qK + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- ++ get_proxy some-name ++ local target_cluster=some-name +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.2sYZAJQk7z ++++ mktemp +++ local LAST_ERR=/tmp/tmp.R6UAB0WqQK +++ local exit_status=0 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.2sYZAJQk7z +++ cat /tmp/tmp.R6UAB0WqQK +++ rm /tmp/tmp.2sYZAJQk7z /tmp/tmp.R6UAB0WqQK +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.ARoRQQTD8K ++++ mktemp +++ local LAST_ERR=/tmp/tmp.PfGViUleKo +++ local exit_status=0 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.ARoRQQTD8K +++ cat /tmp/tmp.PfGViUleKo +++ rm /tmp/tmp.ARoRQQTD8K /tmp/tmp.PfGViUleKo +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo some-name-proxysql ++ return + local proxy=some-name-proxysql + kubectl_bin wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n restore-to-encrypted-cluster-20222 ++ mktemp + local LAST_OUT=/tmp/tmp.RVaKUpjb4z ++ mktemp + local LAST_ERR=/tmp/tmp.jJPpWivD8g + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n restore-to-encrypted-cluster-20222 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in $(seq 0 2) + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n restore-to-encrypted-cluster-20222 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in $(seq 0 2) + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n restore-to-encrypted-cluster-20222 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.RVaKUpjb4z + cat /tmp/tmp.jJPpWivD8g error: no matching resources found + rm /tmp/tmp.RVaKUpjb4z /tmp/tmp.jJPpWivD8g + return 1 + true + wait_for_running some-name-proxysql 1 + local name=some-name-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in $(seq 0 $last_pod) + wait_pod some-name-proxysql-0 480 + local pod=some-name-proxysql-0 + local max_retry=480 + local ns= ++ echo some-name-proxysql-0 ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container=proxysql + set +o xtrace pod/some-name-proxysql-0 condition met waiting for pod/some-name-proxysql-0 to become Ready.Ok + wait_for_running some-name-pxc 3 + local name=some-name-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in $(seq 0 $last_pod) + wait_pod some-name-pxc-0 480 + local pod=some-name-pxc-0 + local max_retry=480 + local ns= ++ echo some-name-pxc-0 ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container=pxc + set +o xtrace pod/some-name-pxc-0 condition met waiting for pod/some-name-pxc-0 to become Ready.Ok + for i in $(seq 0 $last_pod) + wait_pod some-name-pxc-1 480 + local pod=some-name-pxc-1 + local max_retry=480 + local ns= ++ echo some-name-pxc-1 ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container=pxc + set +o xtrace pod/some-name-pxc-1 condition met waiting for pod/some-name-pxc-1 to become Ready.Ok + for i in $(seq 0 $last_pod) + wait_pod some-name-pxc-2 480 + local pod=some-name-pxc-2 + local max_retry=480 + local ns= ++ echo some-name-pxc-2 ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container=pxc + set +o xtrace pod/some-name-pxc-2 condition met waiting for pod/some-name-pxc-2 to become Ready.Ok + sleep 10 ++ kubectl get pxc some-name -o 'jsonpath={.spec.secretsName}' + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.lkOeatynxU +++ mktemp ++ local LAST_ERR=/tmp/tmp.PelFaPBmig ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.lkOeatynxU ++ cat /tmp/tmp.PelFaPBmig ++ rm /tmp/tmp.lkOeatynxU /tmp/tmp.PelFaPBmig ++ return 0 + local root_pass=root_password + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' '-h some-name-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' + local 'uri=-h some-name-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Rbyu6jDSMj +++ mktemp ++ local LAST_ERR=/tmp/tmp.RpD7Bv1yMe ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Rbyu6jDSMj ++ cat /tmp/tmp.RpD7Bv1yMe ++ rm /tmp/tmp.Rbyu6jDSMj /tmp/tmp.RpD7Bv1yMe ++ return 0 + client_pod=pxc-client-59944c5bbf-jrsmf + wait_pod pxc-client-59944c5bbf-jrsmf + local pod=pxc-client-59944c5bbf-jrsmf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-jrsmf ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container= + set +o xtrace pod/pxc-client-59944c5bbf-jrsmf condition met waiting for pod/pxc-client-59944c5bbf-jrsmf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h some-name-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h some-name-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.v1EAHIh4mG +++ mktemp ++ local LAST_ERR=/tmp/tmp.uScNxrFpbE ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.v1EAHIh4mG ++ cat /tmp/tmp.uScNxrFpbE ++ rm /tmp/tmp.v1EAHIh4mG /tmp/tmp.uScNxrFpbE ++ return 0 + client_pod=pxc-client-59944c5bbf-jrsmf + wait_pod pxc-client-59944c5bbf-jrsmf + local pod=pxc-client-59944c5bbf-jrsmf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-jrsmf ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container= + set +o xtrace pod/pxc-client-59944c5bbf-jrsmf condition met waiting for pod/pxc-client-59944c5bbf-jrsmf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in $(seq 0 $((size - 1))) + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/restore-to-encrypted-cluster/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.s61Dp3zdEn +++ mktemp ++ local LAST_ERR=/tmp/tmp.mhWh7PWKHr ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.s61Dp3zdEn ++ cat /tmp/tmp.mhWh7PWKHr ++ rm /tmp/tmp.s61Dp3zdEn /tmp/tmp.mhWh7PWKHr ++ return 0 + client_pod=pxc-client-59944c5bbf-jrsmf + wait_pod pxc-client-59944c5bbf-jrsmf + local pod=pxc-client-59944c5bbf-jrsmf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-jrsmf ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container= + set +o xtrace pod/pxc-client-59944c5bbf-jrsmf condition met waiting for pod/pxc-client-59944c5bbf-jrsmf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.oBtCviH3md/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql /tmp/tmp.oBtCviH3md/select-1.sql + for i in $(seq 0 $((size - 1))) + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/restore-to-encrypted-cluster/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.st2H5WqWRW +++ mktemp ++ local LAST_ERR=/tmp/tmp.v4pTQ536rQ ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.st2H5WqWRW ++ cat /tmp/tmp.v4pTQ536rQ ++ rm /tmp/tmp.st2H5WqWRW /tmp/tmp.v4pTQ536rQ ++ return 0 + client_pod=pxc-client-59944c5bbf-jrsmf + wait_pod pxc-client-59944c5bbf-jrsmf + local pod=pxc-client-59944c5bbf-jrsmf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-jrsmf ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container= + set +o xtrace pod/pxc-client-59944c5bbf-jrsmf condition met waiting for pod/pxc-client-59944c5bbf-jrsmf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.oBtCviH3md/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql /tmp/tmp.oBtCviH3md/select-1.sql + for i in $(seq 0 $((size - 1))) + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/restore-to-encrypted-cluster/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1Na5Iof7iH +++ mktemp ++ local LAST_ERR=/tmp/tmp.tROx7f1G8d ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1Na5Iof7iH ++ cat /tmp/tmp.tROx7f1G8d ++ rm /tmp/tmp.1Na5Iof7iH /tmp/tmp.tROx7f1G8d ++ return 0 + client_pod=pxc-client-59944c5bbf-jrsmf + wait_pod pxc-client-59944c5bbf-jrsmf + local pod=pxc-client-59944c5bbf-jrsmf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-jrsmf ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container= + set +o xtrace pod/pxc-client-59944c5bbf-jrsmf condition met waiting for pod/pxc-client-59944c5bbf-jrsmf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.oBtCviH3md/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql /tmp/tmp.oBtCviH3md/select-1.sql ++ is_keyring_plugin_in_use some-name ++ local cluster=some-name ++ kubectl_bin exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ egrep -o 'early-plugin-load=keyring_\w+.so' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2CBGk63mwg +++ mktemp egrep: warning: egrep is obsolescent; using grep -E ++ local LAST_ERR=/tmp/tmp.Qrw9at2pNY ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2CBGk63mwg ++ cat /tmp/tmp.Qrw9at2pNY Unable to use a TTY - input is not a terminal or the right kind of file ++ rm /tmp/tmp.2CBGk63mwg /tmp/tmp.Qrw9at2pNY ++ return 0 + '[' '' ']' + keyring_plugin_must_not_be_in_use some-name + local cluster=some-name + is_keyring_plugin_in_use some-name + local cluster=some-name + kubectl_bin exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' + egrep -o 'early-plugin-load=keyring_\w+.so' ++ mktemp egrep: warning: egrep is obsolescent; using grep -E + local LAST_OUT=/tmp/tmp.JrAof626Un ++ mktemp + local LAST_ERR=/tmp/tmp.Vl6D2fWcoF + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.JrAof626Un + cat /tmp/tmp.Vl6D2fWcoF Unable to use a TTY - input is not a terminal or the right kind of file + rm /tmp/tmp.JrAof626Un /tmp/tmp.Vl6D2fWcoF + return 0 + table_must_not_be_encrypted some-name myApp + local cluster=some-name + local table=myApp + is_table_encrypted some-name myApp + local cluster=some-name + local table=myApp + run_mysql 'SELECT CREATE_OPTIONS FROM INFORMATION_SCHEMA.TABLES WHERE TABLE_NAME=\"myApp\";' '-h some-name-proxysql -uroot -proot_password' + local 'command=SELECT CREATE_OPTIONS FROM INFORMATION_SCHEMA.TABLES WHERE TABLE_NAME=\"myApp\";' + egrep -o 'ENCRYPTION=('\''Y'\''|"Y")' + local 'uri=-h some-name-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp egrep: warning: egrep is obsolescent; using grep -E ++ local LAST_OUT=/tmp/tmp.DCzX1UL2t5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.S9A9hzJeXm ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DCzX1UL2t5 ++ cat /tmp/tmp.S9A9hzJeXm ++ rm /tmp/tmp.DCzX1UL2t5 /tmp/tmp.S9A9hzJeXm ++ return 0 + client_pod=pxc-client-59944c5bbf-jrsmf + wait_pod pxc-client-59944c5bbf-jrsmf + local pod=pxc-client-59944c5bbf-jrsmf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-jrsmf ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container= + set +o xtrace pod/pxc-client-59944c5bbf-jrsmf condition met waiting for pod/pxc-client-59944c5bbf-jrsmf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_backup some-name on-demand-backup-pvc + local cluster=some-name + local backup=on-demand-backup-pvc + log 'run pxc-backup/on-demand-backup-pvc' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-10-02T02:32:54+0000]' run pxc-backup/on-demand-backup-pvc [2025-10-02T02:32:54+0000] run pxc-backup/on-demand-backup-pvc + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/restore-to-encrypted-cluster/conf/on-demand-backup-pvc.yml ++ mktemp + local LAST_OUT=/tmp/tmp.HPaL0xUOBx ++ mktemp + local LAST_ERR=/tmp/tmp.LlPJ3VxLpW + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/restore-to-encrypted-cluster/conf/on-demand-backup-pvc.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.HPaL0xUOBx perconaxtradbclusterbackup.pxc.percona.com/on-demand-backup-pvc created + cat /tmp/tmp.LlPJ3VxLpW + rm /tmp/tmp.HPaL0xUOBx /tmp/tmp.LlPJ3VxLpW + return 0 + wait_backup on-demand-backup-pvc + local backup=on-demand-backup-pvc + local status=Succeeded + set +o xtrace waiting for pxc-backup/on-demand-backup-pvc to reach Succeeded state..........................Succeeded + '[' -z '' ']' + run_backup some-name on-demand-backup-aws-s3 + local cluster=some-name + local backup=on-demand-backup-aws-s3 + log 'run pxc-backup/on-demand-backup-aws-s3' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-10-02T02:33:37+0000]' run pxc-backup/on-demand-backup-aws-s3 [2025-10-02T02:33:37+0000] run pxc-backup/on-demand-backup-aws-s3 + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/restore-to-encrypted-cluster/conf/on-demand-backup-aws-s3.yml ++ mktemp + local LAST_OUT=/tmp/tmp.sOX746i9bG ++ mktemp + local LAST_ERR=/tmp/tmp.PVWguyRWTd + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/restore-to-encrypted-cluster/conf/on-demand-backup-aws-s3.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.sOX746i9bG perconaxtradbclusterbackup.pxc.percona.com/on-demand-backup-aws-s3 created + cat /tmp/tmp.PVWguyRWTd + rm /tmp/tmp.sOX746i9bG /tmp/tmp.PVWguyRWTd + return 0 + wait_backup on-demand-backup-aws-s3 + local backup=on-demand-backup-aws-s3 + local status=Succeeded + set +o xtrace waiting for pxc-backup/on-demand-backup-aws-s3 to reach Succeeded state............................Succeeded + vault1=vault-service-1-14483 + start_vault vault-service-1-14483 + name=vault-service-1-14483 + protocol=http + local platform=kubernetes + [[ -n '' ]] + create_namespace vault-service-1-14483 skip_clean + local namespace=vault-service-1-14483 + local skip_clean_namespace=skip_clean + [[ 1 == 1 ]] + [[ -z skip_clean ]] + '[' -n '' ']' + desc 'cleaned up old namespaces vault-service-1-14483' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces vault-service-1-14483 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace vault-service-1-14483 ++ mktemp + local LAST_OUT=/tmp/tmp.AngAFTFkTr ++ mktemp + local LAST_ERR=/tmp/tmp.MAYBloPVmG + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete namespace vault-service-1-14483 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in $(seq 0 2) + set +e + kubectl delete namespace vault-service-1-14483 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in $(seq 0 2) + set +e + kubectl delete namespace vault-service-1-14483 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.AngAFTFkTr + cat /tmp/tmp.MAYBloPVmG Error from server (NotFound): namespaces "vault-service-1-14483" not found + rm /tmp/tmp.AngAFTFkTr /tmp/tmp.MAYBloPVmG + return 1 + : + wait_for_delete namespace/vault-service-1-14483 + local res=namespace/vault-service-1-14483 + echo -n 'waiting for namespace/vault-service-1-14483 to be deleted' waiting for namespace/vault-service-1-14483 to be deleted+ set +o xtrace Error from server (NotFound): namespaces "vault-service-1-14483" not found + desc 'create namespace vault-service-1-14483' + set +o xtrace ----------------------------------------------------------------------------------- create namespace vault-service-1-14483 ----------------------------------------------------------------------------------- + kubectl_bin create namespace vault-service-1-14483 ++ mktemp + local LAST_OUT=/tmp/tmp.3pPaqXdJAL ++ mktemp + local LAST_ERR=/tmp/tmp.wEKC2m2bcn + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace vault-service-1-14483 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.3pPaqXdJAL namespace/vault-service-1-14483 created + cat /tmp/tmp.wEKC2m2bcn + rm /tmp/tmp.3pPaqXdJAL /tmp/tmp.wEKC2m2bcn + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.56oGtQmbwT +++ mktemp ++ local LAST_ERR=/tmp/tmp.xnLrMgmgtt ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.56oGtQmbwT ++ cat /tmp/tmp.xnLrMgmgtt ++ rm /tmp/tmp.56oGtQmbwT /tmp/tmp.xnLrMgmgtt ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2203-07b4356f-1-cluster8 --namespace=vault-service-1-14483 ++ mktemp + local LAST_OUT=/tmp/tmp.TIyzkNQpwD ++ mktemp + local LAST_ERR=/tmp/tmp.Rq65rd3VXQ + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2203-07b4356f-1-cluster8 --namespace=vault-service-1-14483 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.TIyzkNQpwD Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2203-07b4356f-1-cluster8" modified. + cat /tmp/tmp.Rq65rd3VXQ + rm /tmp/tmp.TIyzkNQpwD /tmp/tmp.Rq65rd3VXQ + return 0 + deploy_helm vault-service-1-14483 + helm repo add hashicorp https://helm.releases.hashicorp.com "hashicorp" already exists with the same configuration, skipping + helm repo add minio https://charts.min.io/ "minio" already exists with the same configuration, skipping + helm repo update Hang tight while we grab the latest from your chart repositories... ...Successfully got an update from the "chaos-mesh" chart repository ...Successfully got an update from the "minio" chart repository ...Successfully got an update from the "percona" chart repository ...Successfully got an update from the "hashicorp" chart repository Update Complete. ⎈Happy Helming!⎈ + helm uninstall vault-service-1-14483 Error: uninstall: Release not loaded: vault-service-1-14483: release: not found + : + desc 'install Vault vault-service-1-14483' + set +o xtrace ----------------------------------------------------------------------------------- install Vault vault-service-1-14483 ----------------------------------------------------------------------------------- + '[' http == https ']' + helm install vault-service-1-14483 hashicorp/vault --disable-openapi-validation --version 0.30.0 --namespace vault-service-1-14483 --set dataStorage.enabled=false --set global.platform=kubernetes NAME: vault-service-1-14483 LAST DEPLOYED: Thu Oct 2 02:34:25 2025 NAMESPACE: vault-service-1-14483 STATUS: deployed REVISION: 1 NOTES: Thank you for installing HashiCorp Vault! Now that you have deployed Vault, you should look over the docs on using Vault with Kubernetes available here: https://developer.hashicorp.com/vault/docs Your release is named vault-service-1-14483. To learn more about the release, try: $ helm status vault-service-1-14483 $ helm get manifest vault-service-1-14483 + [[ -n '' ]] + set +o xtrace pod/vault-service-1-14483-0......{"running":{"startedAt":"2025-10-02T02:34:41Z"}} + kubectl_bin exec -it vault-service-1-14483-0 -- vault operator init -tls-skip-verify -key-shares=1 -key-threshold=1 -format=json ++ mktemp + local LAST_OUT=/tmp/tmp.zVA9QITZ9V ++ mktemp + local LAST_ERR=/tmp/tmp.WJiRXqC2C2 + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec -it vault-service-1-14483-0 -- vault operator init -tls-skip-verify -key-shares=1 -key-threshold=1 -format=json + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.zVA9QITZ9V + cat /tmp/tmp.WJiRXqC2C2 Unable to use a TTY - input is not a terminal or the right kind of file + rm /tmp/tmp.zVA9QITZ9V /tmp/tmp.WJiRXqC2C2 + return 0 ++ jq -r '.unseal_keys_b64[]' + unsealKey=bjYOI3pTAwKFDdp8B77ZFM7W/5DrH29jyChyD+CgpV8= ++ jq -r .root_token + token=hvs.rx8IbSwkyogFtCaLg4PvLP7j + sleep 10 + kubectl_bin exec -it vault-service-1-14483-0 -- vault operator unseal -tls-skip-verify bjYOI3pTAwKFDdp8B77ZFM7W/5DrH29jyChyD+CgpV8= ++ mktemp + local LAST_OUT=/tmp/tmp.c3ZkSu2n8Y ++ mktemp + local LAST_ERR=/tmp/tmp.XZo4qT62AD + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec -it vault-service-1-14483-0 -- vault operator unseal -tls-skip-verify bjYOI3pTAwKFDdp8B77ZFM7W/5DrH29jyChyD+CgpV8= + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.c3ZkSu2n8Y Key Value --- ----- Seal Type shamir Initialized true Sealed false Total Shares 1 Threshold 1 Version 1.19.0 Build Date 2025-03-04T12:36:40Z Storage Type file Cluster Name vault-cluster-42bff641 Cluster ID 730906f9-3537-65a9-ad3d-9ddeba31ed96 HA Enabled false + cat /tmp/tmp.XZo4qT62AD Unable to use a TTY - input is not a terminal or the right kind of file + rm /tmp/tmp.c3ZkSu2n8Y /tmp/tmp.XZo4qT62AD + return 0 + kubectl_bin exec -it vault-service-1-14483-0 -- sh -c 'export VAULT_TOKEN=hvs.rx8IbSwkyogFtCaLg4PvLP7j && export VAULT_LOG_LEVEL=trace && vault secrets enable --version=1 -tls-skip-verify -path=secret kv && vault audit enable file file_path=/vault/vault-audit.log' ++ mktemp + local LAST_OUT=/tmp/tmp.Uf5ULWAwqE ++ mktemp + local LAST_ERR=/tmp/tmp.OPI1vhd2kw + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec -it vault-service-1-14483-0 -- sh -c 'export VAULT_TOKEN=hvs.rx8IbSwkyogFtCaLg4PvLP7j && export VAULT_LOG_LEVEL=trace && vault secrets enable --version=1 -tls-skip-verify -path=secret kv && vault audit enable file file_path=/vault/vault-audit.log' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Uf5ULWAwqE Success! Enabled the kv secrets engine at: secret/ Success! Enabled the file audit device at: file/ + cat /tmp/tmp.OPI1vhd2kw Unable to use a TTY - input is not a terminal or the right kind of file + rm /tmp/tmp.Uf5ULWAwqE /tmp/tmp.OPI1vhd2kw + return 0 + sleep 10 + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/conf/vault-secret.yaml + sed -e s/#token/hvs.rx8IbSwkyogFtCaLg4PvLP7j/ + sed -e 's/#vault_url/http:\/\/vault-service-1-14483.vault-service-1-14483.svc.cluster.local:8200/' + sed -e s/#secret/secret/ + '[' http == https ']' + /usr/sbin/sed -i /#vault_ca/d /tmp/tmp.oBtCviH3md/vault-secret.yaml + kubectl_bin apply --namespace=restore-to-encrypted-cluster-20222 -f /tmp/tmp.oBtCviH3md/vault-secret.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.WLYfwfaBg5 ++ mktemp + local LAST_ERR=/tmp/tmp.AksyS9ZYFT + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply --namespace=restore-to-encrypted-cluster-20222 -f /tmp/tmp.oBtCviH3md/vault-secret.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.WLYfwfaBg5 secret/some-name-vault created + cat /tmp/tmp.AksyS9ZYFT + rm /tmp/tmp.WLYfwfaBg5 /tmp/tmp.AksyS9ZYFT + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.DVZ6Q2pgbc +++ mktemp ++ local LAST_ERR=/tmp/tmp.4EAZUQRsRM ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DVZ6Q2pgbc ++ cat /tmp/tmp.4EAZUQRsRM ++ rm /tmp/tmp.DVZ6Q2pgbc /tmp/tmp.4EAZUQRsRM ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2203-07b4356f-1-cluster8 --namespace=restore-to-encrypted-cluster-20222 ++ mktemp + local LAST_OUT=/tmp/tmp.6h5MLotylQ ++ mktemp + local LAST_ERR=/tmp/tmp.NQ5nvRmmFh + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2203-07b4356f-1-cluster8 --namespace=restore-to-encrypted-cluster-20222 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.6h5MLotylQ Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2203-07b4356f-1-cluster8" modified. + cat /tmp/tmp.NQ5nvRmmFh + rm /tmp/tmp.6h5MLotylQ /tmp/tmp.NQ5nvRmmFh + return 0 + run_recovery_check some-name on-demand-backup-pvc + local cluster=some-name + local backup=on-demand-backup-pvc ++ get_proxy_engine some-name ++ local cluster_name=some-name +++ get_proxy some-name +++ local target_cluster=some-name ++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.YZcU5uZfWO +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.HJ7PSZZysM ++++ local exit_status=0 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 ']' ++++ break ++++ cat /tmp/tmp.YZcU5uZfWO ++++ cat /tmp/tmp.HJ7PSZZysM ++++ rm /tmp/tmp.YZcU5uZfWO /tmp/tmp.HJ7PSZZysM ++++ return 0 +++ [[ '' == \t\r\u\e ]] ++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.tp5kCiL52w +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.nlKkVxo8yX ++++ local exit_status=0 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 ']' ++++ break ++++ cat /tmp/tmp.tp5kCiL52w ++++ cat /tmp/tmp.nlKkVxo8yX ++++ rm /tmp/tmp.tp5kCiL52w /tmp/tmp.nlKkVxo8yX ++++ return 0 +++ [[ true == \t\r\u\e ]] +++ echo some-name-proxysql +++ return ++ local cluster_proxy=some-name-proxysql ++ echo proxysql + local proxy=proxysql + log 'run pxc-restore/on-demand-backup-pvc' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-10-02T02:35:12+0000]' run pxc-restore/on-demand-backup-pvc [2025-10-02T02:35:12+0000] run pxc-restore/on-demand-backup-pvc + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/restore-to-encrypted-cluster/conf/restore-on-demand-backup-pvc.yaml + /usr/sbin/sed -e s~minio-service.#namespace~minio-service.restore-to-encrypted-cluster-20222~ + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.d2RSC8dr4Z ++ mktemp + local LAST_ERR=/tmp/tmp.i4CRmY76k9 + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.d2RSC8dr4Z perconaxtradbclusterrestore.pxc.percona.com/on-demand-backup-pvc created + cat /tmp/tmp.i4CRmY76k9 + rm /tmp/tmp.d2RSC8dr4Z /tmp/tmp.i4CRmY76k9 + return 0 + wait_backup_restore on-demand-backup-pvc + local backup_name=on-demand-backup-pvc + local target_state=Succeeded + local wait_time=720 + set +o xtrace waiting for pxc-restore/on-demand-backup-pvc to reach Succeeded state 2025-10-02T02:35:15 pxc-restore/on-demand-backup-pvc state: Starting 2025-10-02T02:35:17 pxc-restore/on-demand-backup-pvc state: Starting 2025-10-02T02:35:19 pxc-restore/on-demand-backup-pvc state: Starting 2025-10-02T02:35:22 pxc-restore/on-demand-backup-pvc state: Starting 2025-10-02T02:35:24 pxc-restore/on-demand-backup-pvc state: Starting 2025-10-02T02:35:26 pxc-restore/on-demand-backup-pvc state: Starting 2025-10-02T02:35:28 pxc-restore/on-demand-backup-pvc state: Starting 2025-10-02T02:35:30 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-02T02:35:32 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-02T02:35:34 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-02T02:35:36 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-02T02:35:38 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-02T02:35:40 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-02T02:35:42 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-02T02:35:44 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-02T02:35:46 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-02T02:35:48 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-02T02:35:51 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-02T02:35:53 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-02T02:35:55 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-02T02:35:57 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-02T02:35:59 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-02T02:36:01 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-02T02:36:03 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-02T02:36:05 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-02T02:36:07 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-02T02:36:09 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-02T02:36:11 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-02T02:36:13 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-02T02:36:16 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-02T02:36:18 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-02T02:36:20 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-10-02T02:36:22 pxc-restore/on-demand-backup-pvc state: Restoring 2025-10-02T02:36:24 pxc-restore/on-demand-backup-pvc state: Restoring 2025-10-02T02:36:26 pxc-restore/on-demand-backup-pvc state: Restoring 2025-10-02T02:36:28 pxc-restore/on-demand-backup-pvc state: Restoring 2025-10-02T02:36:30 pxc-restore/on-demand-backup-pvc state: Restoring 2025-10-02T02:36:33 pxc-restore/on-demand-backup-pvc state: Restoring 2025-10-02T02:36:35 pxc-restore/on-demand-backup-pvc state: Restoring 2025-10-02T02:36:37 pxc-restore/on-demand-backup-pvc state: Restoring 2025-10-02T02:36:39 pxc-restore/on-demand-backup-pvc state: Restoring 2025-10-02T02:36:41 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:36:43 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:36:45 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:36:47 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:36:49 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:36:51 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:36:53 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:36:55 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:36:58 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:00 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:02 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:04 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:06 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:08 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:10 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:12 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:14 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:17 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:19 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:21 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:23 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:25 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:27 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:29 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:31 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:33 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:35 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:37 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:39 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:41 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:44 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:46 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:48 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:50 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:52 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:54 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:56 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:37:58 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:38:00 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-10-02T02:38:02 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:04 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:06 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:08 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:10 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:12 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:15 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:17 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:19 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:21 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:23 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:25 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:27 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:29 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:31 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:33 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:35 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:37 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:39 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:41 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:43 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:45 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:48 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:50 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:52 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:54 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:56 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:38:58 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:00 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:02 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:04 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:06 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:08 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:10 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:12 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:14 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:16 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:18 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:21 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:23 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:25 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:27 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:29 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:31 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:33 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:35 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:37 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:39 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:41 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:43 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:45 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:47 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:49 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:52 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:54 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:56 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:39:58 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:00 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:02 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:04 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:06 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:08 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:10 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:12 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:15 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:17 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:19 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:21 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:23 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:25 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:27 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:29 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:31 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:33 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:35 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:37 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:39 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:42 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:44 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:46 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:47 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:49 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:51 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:53 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:55 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:57 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:40:59 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:41:01 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:41:04 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:41:06 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:41:08 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:41:10 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-10-02T02:41:12 pxc-restore/on-demand-backup-pvc state: Succeeded + kubectl_bin logs job/restore-job-on-demand-backup-pvc-some-name ++ mktemp + local LAST_OUT=/tmp/tmp.NqBRJrsjts ++ mktemp + local LAST_ERR=/tmp/tmp.y8wmezuma9 + local exit_status=0 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl logs job/restore-job-on-demand-backup-pvc-some-name + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.NqBRJrsjts + LIB_PATH=/opt/percona/backup/lib/pxc + . /opt/percona/backup/lib/pxc/check-version.sh + . /opt/percona/backup/lib/pxc/vault.sh ++ set -o errexit ++ keyring_vault=/etc/mysql/vault-keyring-secret/keyring_vault.conf + SOCAT_OPTS=TCP:restore-src-on-demand-backup-pvc-some-name:3307,retry=30 + check_ssl + CA=/var/run/secrets/kubernetes.io/serviceaccount/ca.crt + '[' -f /var/run/secrets/kubernetes.io/serviceaccount/service-ca.crt ']' + SSL_DIR=/etc/mysql/ssl + '[' -f /etc/mysql/ssl/ca.crt ']' + CA=/etc/mysql/ssl/ca.crt + SSL_INTERNAL_DIR=/etc/mysql/ssl-internal + '[' -f /etc/mysql/ssl-internal/ca.crt ']' + CA=/etc/mysql/ssl-internal/ca.crt + KEY=/etc/mysql/ssl/tls.key + CERT=/etc/mysql/ssl/tls.crt + '[' -f /etc/mysql/ssl-internal/tls.key ']' + '[' -f /etc/mysql/ssl-internal/tls.crt ']' + KEY=/etc/mysql/ssl-internal/tls.key + CERT=/etc/mysql/ssl-internal/tls.crt + '[' -f /etc/mysql/ssl-internal/ca.crt ']' + '[' -f /etc/mysql/ssl-internal/tls.key ']' + '[' -f /etc/mysql/ssl-internal/tls.crt ']' + SOCAT_OPTS='openssl-connect:restore-src-on-demand-backup-pvc-some-name:3307,reuseaddr,cert=/etc/mysql/ssl-internal/tls.crt,key=/etc/mysql/ssl-internal/tls.key,cafile=/etc/mysql/ssl-internal/ca.crt,verify=1,commonname='\'''\'',retry=30,no-sni=1' + ping -c1 restore-src-on-demand-backup-pvc-some-name /opt/percona/backup/recovery-pvc-joiner.sh: line 40: ping: command not found + : + rm -rf /datadir/#ib_16384_0.dblwr /datadir/#ib_16384_1.dblwr /datadir/#ib_16384_10.dblwr /datadir/#ib_16384_11.dblwr /datadir/#ib_16384_12.dblwr /datadir/#ib_16384_13.dblwr /datadir/#ib_16384_14.dblwr /datadir/#ib_16384_15.dblwr /datadir/#ib_16384_2.dblwr /datadir/#ib_16384_3.dblwr /datadir/#ib_16384_4.dblwr /datadir/#ib_16384_5.dblwr /datadir/#ib_16384_6.dblwr /datadir/#ib_16384_7.dblwr /datadir/#ib_16384_8.dblwr /datadir/#ib_16384_9.dblwr /datadir/#innodb_redo /datadir/#innodb_temp /datadir/auth_plugin /datadir/auto.cnf /datadir/binlog.000001 /datadir/binlog.000002 /datadir/binlog.000003 /datadir/binlog.index /datadir/galera.cache /datadir/get-pxc-state /datadir/grastate.dat /datadir/ib_buffer_pool /datadir/ibdata1 /datadir/innobackup.backup.full.log /datadir/innobackup.backup.log /datadir/liveness-check.sh /datadir/myApp /datadir/mysql /datadir/mysql-state-monitor /datadir/mysql-state-monitor.log /datadir/mysql.ibd /datadir/mysql.state /datadir/mysqld-error.log /datadir/notify.sock /datadir/peer-list /datadir/performance_schema /datadir/pmm-prerun.sh /datadir/private_key.pem /datadir/public_key.pem /datadir/pxc-configure-pxc.sh /datadir/pxc-entrypoint.sh /datadir/readiness-check.sh /datadir/sys /datadir/undo_001 /datadir/undo_002 /datadir/version_info /datadir/wsrep_cmd_notify_handler.sh ++ mktemp --directory /datadir/pxc_sst_XXXX + tmp=/datadir/pxc_sst_2DB1 + socat -u 'openssl-connect:restore-src-on-demand-backup-pvc-some-name:3307,reuseaddr,cert=/etc/mysql/ssl-internal/tls.crt,key=/etc/mysql/ssl-internal/tls.key,cafile=/etc/mysql/ssl-internal/ca.crt,verify=1,commonname='\'''\'',retry=30,no-sni=1' stdio ++ parse_ini mysql-version /datadir/pxc_sst_2DB1/sst_info ++ local key=mysql-version ++ local file_path=/datadir/pxc_sst_2DB1/sst_info ++ awk -F '=[ ]*' '/mysql-version[ ]*=/ {print $2}' /datadir/pxc_sst_2DB1/sst_info + MYSQL_VERSION=8.0.42-33.1 + check_for_version 8.0.42-33.1 8.0.0 + '[' -z 8.0.42-33.1 ']' + '[' -z 8.0.0 ']' + local local_version_str + local required_version_str ++ normalize_version 8.0.42-33.1 ++ local major=0 ++ local minor=0 ++ local patch=0 ++ [[ 8.0.42-33.1 =~ ^([0-9]+)\.([0-9]+)\.?([0-9]*)([^ ])* ]] ++ major=8 ++ minor=0 ++ patch=42 ++ printf %02d%02d%02d 8 0 42 + local_version_str=080042 ++ normalize_version 8.0.0 ++ local major=0 ++ local minor=0 ++ local patch=0 ++ [[ 8.0.0 =~ ^([0-9]+)\.([0-9]+)\.?([0-9]*)([^ ])* ]] ++ major=8 ++ minor=0 ++ patch=0 ++ printf %02d%02d%02d 8 0 0 + required_version_str=080000 + [[ 080042 < 080000 ]] + return 0 + XBSTREAM_EXTRA_ARGS=' --decompress' + socat -u 'openssl-connect:restore-src-on-demand-backup-pvc-some-name:3307,reuseaddr,cert=/etc/mysql/ssl-internal/tls.crt,key=/etc/mysql/ssl-internal/tls.key,cafile=/etc/mysql/ssl-internal/ca.crt,verify=1,commonname='\'''\'',retry=30,no-sni=1' stdio ++ grep -c processor /proc/cpuinfo + xbstream -x -C /datadir/pxc_sst_2DB1 --parallel=4 --decompress + set +o xtrace % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 14 100 14 0 0 736 0 --:--:-- --:--:-- --:--:-- 736 + xtrabackup --use-memory=100MB --prepare --binlog-info=ON --rollback-prepared-trx --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_2DB1 2025-10-02T02:36:31.984071-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized server arguments: --innodb_checksum_algorithm=crc32 --innodb_log_checksums=1 --innodb_data_file_path=ibdata1:12M:autoextend --innodb_log_file_size=50331648 --innodb_page_size=16384 --innodb_undo_directory=./ --innodb_undo_tablespaces=2 --server-id=17542352 --innodb_log_checksums=ON --innodb_redo_log_encrypt=0 --innodb_undo_log_encrypt=0 2025-10-02T02:36:31.984154-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized client arguments: --use-memory=100MB --prepare=1 --rollback-prepared-trx=1 --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_2DB1 xtrabackup version 8.0.35-33 based on MySQL server 8.0.35 Linux (x86_64) (revision id: a982afdd) 2025-10-02T02:36:31.984185-00:00 0 [Note] [MY-011825] [Xtrabackup] cd to /datadir/pxc_sst_2DB1/ 2025-10-02T02:36:31.984252-00:00 0 [Note] [MY-011825] [Xtrabackup] This target seems to be not prepared yet. 2025-10-02T02:36:31.993587-00:00 0 [Note] [MY-011825] [Xtrabackup] xtrabackup_logfile detected: size=8388608, start_lsn=(30285583) 2025-10-02T02:36:31.994350-00:00 0 [Note] [MY-011825] [Xtrabackup] using the following InnoDB configuration for recovery: 2025-10-02T02:36:31.994362-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_home_dir = . 2025-10-02T02:36:31.994366-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_file_path = ibdata1:12M:autoextend 2025-10-02T02:36:31.994389-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_group_home_dir = . 2025-10-02T02:36:31.994399-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_files_in_group = 1 2025-10-02T02:36:31.994406-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_file_size = 8388608 2025-10-02T02:36:31.994574-00:00 0 [Note] [MY-011825] [Xtrabackup] inititialize_service_handles suceeded 2025-10-02T02:36:31.994731-00:00 0 [Note] [MY-011825] [Xtrabackup] using the following InnoDB configuration for recovery: 2025-10-02T02:36:31.994741-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_home_dir = . 2025-10-02T02:36:31.994745-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_file_path = ibdata1:12M:autoextend 2025-10-02T02:36:31.994750-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_group_home_dir = . 2025-10-02T02:36:31.994776-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_files_in_group = 1 2025-10-02T02:36:31.994786-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_file_size = 8388608 2025-10-02T02:36:31.994799-00:00 0 [Note] [MY-011825] [Xtrabackup] Starting InnoDB instance for recovery. 2025-10-02T02:36:31.994807-00:00 0 [Note] [MY-011825] [Xtrabackup] Using 104857600 bytes for buffer pool (set by --use-memory parameter) 2025-10-02T02:36:31.994838-00:00 0 [Note] [MY-012932] [InnoDB] PUNCH HOLE support available 2025-10-02T02:36:31.994854-00:00 0 [Note] [MY-012944] [InnoDB] Uses event mutexes 2025-10-02T02:36:31.994860-00:00 0 [Note] [MY-012945] [InnoDB] GCC builtin __atomic_thread_fence() is used for memory barrier 2025-10-02T02:36:31.994866-00:00 0 [Note] [MY-012948] [InnoDB] Compressed tables use zlib 1.2.13 2025-10-02T02:36:31.995015-00:00 0 [Note] [MY-012951] [InnoDB] Using hardware accelerated crc32 and polynomial multiplication. 2025-10-02T02:36:31.995344-00:00 0 [Note] [MY-012203] [InnoDB] Directories to scan './' 2025-10-02T02:36:31.995378-00:00 0 [Note] [MY-012204] [InnoDB] Scanning './' 2025-10-02T02:36:32.002795-00:00 0 [Note] [MY-012208] [InnoDB] Completed space ID check of 8 files. 2025-10-02T02:36:32.003324-00:00 0 [Note] [MY-012955] [InnoDB] Initializing buffer pool, total size = 128.000000M, instances = 1, chunk size =128.000000M 2025-10-02T02:36:32.009789-00:00 0 [Note] [MY-012957] [InnoDB] Completed initialization of buffer pool 2025-10-02T02:36:32.011493-00:00 0 [Note] [MY-011952] [InnoDB] If the mysqld execution user is authorized, page cleaner thread priority can be changed. See the man page of setpriority(). 2025-10-02T02:36:32.038658-00:00 0 [Note] [MY-013883] [InnoDB] The latest found checkpoint is at lsn = 30285583 in redo log file ./#innodb_redo/#ib_redo0. 2025-10-02T02:36:32.038707-00:00 0 [Note] [MY-012560] [InnoDB] The log sequence number 30166941 in the system tablespace does not match the log sequence number 30285583 in the redo log files! 2025-10-02T02:36:32.038719-00:00 0 [Note] [MY-012551] [InnoDB] Database was not shutdown normally! 2025-10-02T02:36:32.038726-00:00 0 [Note] [MY-012552] [InnoDB] Starting crash recovery. 2025-10-02T02:36:32.038835-00:00 0 [Note] [MY-013086] [InnoDB] Starting to parse redo log at lsn = 30285414, whereas checkpoint_lsn = 30285583 and start_lsn = 30285312 2025-10-02T02:36:32.038848-00:00 0 [Note] [MY-012550] [InnoDB] Doing recovery: scanned up to log sequence number 30285593 2025-10-02T02:36:32.053669-00:00 0 [Note] [MY-013083] [InnoDB] Log background threads are being started... 2025-10-02T02:36:32.053995-00:00 0 [Note] [MY-012532] [InnoDB] Applying a batch of 1 redo log records ... 2025-10-02T02:36:32.054039-00:00 0 [Note] [MY-012533] [InnoDB] 100% 2025-10-02T02:36:32.054052-00:00 0 [Note] [MY-012535] [InnoDB] Apply batch completed! 2025-10-02T02:36:32.155469-00:00 0 [Note] [MY-013084] [InnoDB] Log background threads are being closed... 2025-10-02T02:36:32.156921-00:00 0 [Note] [MY-013888] [InnoDB] Upgrading redo log: 1032M, LSN=30285593. 2025-10-02T02:36:32.156932-00:00 0 [Note] [MY-012968] [InnoDB] Starting to delete and rewrite redo log files. 2025-10-02T02:36:32.156972-00:00 0 [Note] [MY-011825] [InnoDB] Removing redo log file: ./#innodb_redo/#ib_redo0 2025-10-02T02:36:32.185829-00:00 0 [Note] [MY-011825] [InnoDB] Creating redo log file at ./#innodb_redo/#ib_redo0_tmp with file_id 0 with size 33554432 bytes 2025-10-02T02:36:32.188691-00:00 0 [Note] [MY-011825] [InnoDB] Renaming redo log file from ./#innodb_redo/#ib_redo0_tmp to ./#innodb_redo/#ib_redo0 2025-10-02T02:36:32.190722-00:00 0 [Note] [MY-012893] [InnoDB] New redo log files created, LSN=30285836 2025-10-02T02:36:32.190789-00:00 0 [Note] [MY-013083] [InnoDB] Log background threads are being started... 2025-10-02T02:36:32.191095-00:00 0 [Note] [MY-013252] [InnoDB] Using undo tablespace './undo_001'. 2025-10-02T02:36:32.191753-00:00 0 [Note] [MY-013252] [InnoDB] Using undo tablespace './undo_002'. 2025-10-02T02:36:32.192814-00:00 0 [Note] [MY-012910] [InnoDB] Opened 2 existing undo tablespaces. 2025-10-02T02:36:32.192861-00:00 0 [Note] [MY-011980] [InnoDB] GTID recovery trx_no: 5200 2025-10-02T02:36:32.254612-00:00 0 [Note] [MY-013776] [InnoDB] Parallel initialization of rseg complete 2025-10-02T02:36:32.254626-00:00 0 [Note] [MY-013777] [InnoDB] Time taken to initialize rseg using 4 thread: 61767 ms. 2025-10-02T02:36:32.254686-00:00 0 [Note] [MY-012923] [InnoDB] Creating shared tablespace for temporary tables 2025-10-02T02:36:32.254773-00:00 0 [Note] [MY-012265] [InnoDB] Setting file './ibtmp1' size to 12 MB. Physically writing the file full; Please wait ... 2025-10-02T02:36:32.283823-00:00 0 [Note] [MY-012266] [InnoDB] File './ibtmp1' size is now 12 MB. 2025-10-02T02:36:32.283939-00:00 0 [Note] [MY-013627] [InnoDB] Scanning temp tablespace dir:'./#innodb_temp/' 2025-10-02T02:36:32.298404-00:00 0 [Note] [MY-013018] [InnoDB] Created 128 and tracked 128 new rollback segment(s) in the temporary tablespace. 128 are now active. 2025-10-02T02:36:32.298537-00:00 0 [Note] [MY-012976] [InnoDB] 8.0.35 started; log sequence number 30285846 2025-10-02T02:36:32.299007-00:00 0 [Warning] [MY-012091] [InnoDB] Allocated tablespace ID 1 for sys/sys_config, old maximum was 0 2025-10-02T02:36:32.301125-00:00 0 [Note] [MY-011825] [Xtrabackup] Completed loading of 6 tablespaces into cache in 0.00256479 seconds 2025-10-02T02:36:32.315687-00:00 0 [Note] [MY-011825] [Xtrabackup] Time taken to build dictionary: 0.0145426 seconds 2025-10-02T02:36:33.316119-00:00 0 [Note] [MY-011825] [Xtrabackup] Recovered WSREP position: 6e951d1e-9f37-11f0-b927-2e04c998bc46:39 2025-10-02T02:36:33.316162-00:00 0 [Note] [MY-011825] [Xtrabackup] starting shutdown with innodb_fast_shutdown = 1 2025-10-02T02:36:33.316206-00:00 0 [Note] [MY-012330] [InnoDB] FTS optimize thread exiting. 2025-10-02T02:36:34.316097-00:00 0 [Note] [MY-013072] [InnoDB] Starting shutdown... 2025-10-02T02:36:34.416929-00:00 0 [Note] [MY-013084] [InnoDB] Log background threads are being closed... 2025-10-02T02:36:34.427816-00:00 0 [Note] [MY-012980] [InnoDB] Shutdown completed; log sequence number 30285846 2025-10-02T02:36:34.430606-00:00 0 [Note] [MY-011825] [Xtrabackup] completed OK! + xtrabackup --defaults-group=mysqld --datadir=/datadir --move-back --binlog-info=ON --force-non-empty-directories --keyring-vault-config=/etc/mysql/vault-keyring-secret/keyring_vault.conf --early-plugin-load=keyring_vault.so --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_2DB1 2025-10-02T02:36:34.444939-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized server arguments: --defaults_group=mysqld --datadir=/datadir 2025-10-02T02:36:34.445000-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized client arguments: --move-back=1 --force-non-empty-directories=1 --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_2DB1 xtrabackup version 8.0.35-33 based on MySQL server 8.0.35 Linux (x86_64) (revision id: a982afdd) 2025-10-02T02:36:34.445034-00:00 0 [Note] [MY-011825] [Xtrabackup] cd to /datadir/pxc_sst_2DB1/ 2025-10-02T02:36:34.445487-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving undo_001 to /datadir/undo_001 2025-10-02T02:36:34.445525-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file undo_001 to /datadir/undo_001 2025-10-02T02:36:34.445556-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving undo_002 to /datadir/undo_002 2025-10-02T02:36:34.445584-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file undo_002 to /datadir/undo_002 2025-10-02T02:36:34.445753-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving ibdata1 to /datadir/ibdata1 2025-10-02T02:36:34.445798-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file ibdata1 to /datadir/ibdata1 2025-10-02T02:36:34.446027-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving binlog.000009 to /datadir//binlog.000009 2025-10-02T02:36:34.446051-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file binlog.000009 to /datadir//binlog.000009 2025-10-02T02:36:34.446126-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving binlog.index to /datadir//binlog.index 2025-10-02T02:36:34.446155-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file binlog.index to /datadir//binlog.index 2025-10-02T02:36:34.446440-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql.ibd to /datadir/mysql.ibd 2025-10-02T02:36:34.446477-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql.ibd to /datadir/mysql.ibd 2025-10-02T02:36:34.446504-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./ib_buffer_pool to /datadir/ib_buffer_pool 2025-10-02T02:36:34.446534-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./ib_buffer_pool to /datadir/ib_buffer_pool 2025-10-02T02:36:34.446622-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_98.sdi to /datadir/performance_schema/events_waits_sum_98.sdi 2025-10-02T02:36:34.446653-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_98.sdi to /datadir/performance_schema/events_waits_sum_98.sdi 2025-10-02T02:36:34.446694-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_143.sdi to /datadir/performance_schema/events_transacti_143.sdi 2025-10-02T02:36:34.446747-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_143.sdi to /datadir/performance_schema/events_transacti_143.sdi 2025-10-02T02:36:34.446797-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/data_lock_waits_170.sdi to /datadir/performance_schema/data_lock_waits_170.sdi 2025-10-02T02:36:34.446857-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/data_lock_waits_170.sdi to /datadir/performance_schema/data_lock_waits_170.sdi 2025-10-02T02:36:34.446883-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_131.sdi to /datadir/performance_schema/events_statement_131.sdi 2025-10-02T02:36:34.446910-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_131.sdi to /datadir/performance_schema/events_statement_131.sdi 2025-10-02T02:36:34.446944-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/data_locks_169.sdi to /datadir/performance_schema/data_locks_169.sdi 2025-10-02T02:36:34.446971-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/data_locks_169.sdi to /datadir/performance_schema/data_locks_169.sdi 2025-10-02T02:36:34.447000-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/rwlock_instances_110.sdi to /datadir/performance_schema/rwlock_instances_110.sdi 2025-10-02T02:36:34.447032-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/rwlock_instances_110.sdi to /datadir/performance_schema/rwlock_instances_110.sdi 2025-10-02T02:36:34.447069-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/pxc_cluster_view_203.sdi to /datadir/performance_schema/pxc_cluster_view_203.sdi 2025-10-02T02:36:34.447096-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/pxc_cluster_view_203.sdi to /datadir/performance_schema/pxc_cluster_view_203.sdi 2025-10-02T02:36:34.447124-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_97.sdi to /datadir/performance_schema/events_waits_sum_97.sdi 2025-10-02T02:36:34.447179-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_97.sdi to /datadir/performance_schema/events_waits_sum_97.sdi 2025-10-02T02:36:34.447212-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_140.sdi to /datadir/performance_schema/events_transacti_140.sdi 2025-10-02T02:36:34.447241-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_140.sdi to /datadir/performance_schema/events_transacti_140.sdi 2025-10-02T02:36:34.447262-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_threads_115.sdi to /datadir/performance_schema/setup_threads_115.sdi 2025-10-02T02:36:34.447281-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_threads_115.sdi to /datadir/performance_schema/setup_threads_115.sdi 2025-10-02T02:36:34.447310-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_accoun_186.sdi to /datadir/performance_schema/status_by_accoun_186.sdi 2025-10-02T02:36:34.447338-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_accoun_186.sdi to /datadir/performance_schema/status_by_accoun_186.sdi 2025-10-02T02:36:34.447370-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_165.sdi to /datadir/performance_schema/memory_summary_b_165.sdi 2025-10-02T02:36:34.447393-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_165.sdi to /datadir/performance_schema/memory_summary_b_165.sdi 2025-10-02T02:36:34.447424-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_175.sdi to /datadir/performance_schema/replication_appl_175.sdi 2025-10-02T02:36:34.447456-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_175.sdi to /datadir/performance_schema/replication_appl_175.sdi 2025-10-02T02:36:34.447486-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_100.sdi to /datadir/performance_schema/events_waits_sum_100.sdi 2025-10-02T02:36:34.447515-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_100.sdi to /datadir/performance_schema/events_waits_sum_100.sdi 2025-10-02T02:36:34.447575-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_thread_188.sdi to /datadir/performance_schema/status_by_thread_188.sdi 2025-10-02T02:36:34.447605-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_thread_188.sdi to /datadir/performance_schema/status_by_thread_188.sdi 2025-10-02T02:36:34.447636-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/user_defined_fun_197.sdi to /datadir/performance_schema/user_defined_fun_197.sdi 2025-10-02T02:36:34.447665-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/user_defined_fun_197.sdi to /datadir/performance_schema/user_defined_fun_197.sdi 2025-10-02T02:36:34.447696-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_asyn_181.sdi to /datadir/performance_schema/replication_asyn_181.sdi 2025-10-02T02:36:34.447722-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_asyn_181.sdi to /datadir/performance_schema/replication_asyn_181.sdi 2025-10-02T02:36:34.447751-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_166.sdi to /datadir/performance_schema/memory_summary_b_166.sdi 2025-10-02T02:36:34.447793-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_166.sdi to /datadir/performance_schema/memory_summary_b_166.sdi 2025-10-02T02:36:34.447825-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/metadata_locks_168.sdi to /datadir/performance_schema/metadata_locks_168.sdi 2025-10-02T02:36:34.447849-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/metadata_locks_168.sdi to /datadir/performance_schema/metadata_locks_168.sdi 2025-10-02T02:36:34.447880-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/hosts_155.sdi to /datadir/performance_schema/hosts_155.sdi 2025-10-02T02:36:34.447905-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/hosts_155.sdi to /datadir/performance_schema/hosts_155.sdi 2025-10-02T02:36:34.447938-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_145.sdi to /datadir/performance_schema/events_transacti_145.sdi 2025-10-02T02:36:34.447961-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_145.sdi to /datadir/performance_schema/events_transacti_145.sdi 2025-10-02T02:36:34.447988-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_his_94.sdi to /datadir/performance_schema/events_waits_his_94.sdi 2025-10-02T02:36:34.448014-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_his_94.sdi to /datadir/performance_schema/events_waits_his_94.sdi 2025-10-02T02:36:34.448063-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_conn_171.sdi to /datadir/performance_schema/replication_conn_171.sdi 2025-10-02T02:36:34.448091-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_conn_171.sdi to /datadir/performance_schema/replication_conn_171.sdi 2025-10-02T02:36:34.448129-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/threads_119.sdi to /datadir/performance_schema/threads_119.sdi 2025-10-02T02:36:34.448155-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/threads_119.sdi to /datadir/performance_schema/threads_119.sdi 2025-10-02T02:36:34.448183-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_176.sdi to /datadir/performance_schema/replication_appl_176.sdi 2025-10-02T02:36:34.448202-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_176.sdi to /datadir/performance_schema/replication_appl_176.sdi 2025-10-02T02:36:34.448224-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/log_status_183.sdi to /datadir/performance_schema/log_status_183.sdi 2025-10-02T02:36:34.448250-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/log_status_183.sdi to /datadir/performance_schema/log_status_183.sdi 2025-10-02T02:36:34.448282-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_99.sdi to /datadir/performance_schema/events_waits_sum_99.sdi 2025-10-02T02:36:34.448311-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_99.sdi to /datadir/performance_schema/events_waits_sum_99.sdi 2025-10-02T02:36:34.448343-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_instances_156.sdi to /datadir/performance_schema/socket_instances_156.sdi 2025-10-02T02:36:34.448371-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_instances_156.sdi to /datadir/performance_schema/socket_instances_156.sdi 2025-10-02T02:36:34.448399-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_account__160.sdi to /datadir/performance_schema/session_account__160.sdi 2025-10-02T02:36:34.448426-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_account__160.sdi to /datadir/performance_schema/session_account__160.sdi 2025-10-02T02:36:34.448457-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/malloc_stats_201.sdi to /datadir/performance_schema/malloc_stats_201.sdi 2025-10-02T02:36:34.448482-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/malloc_stats_201.sdi to /datadir/performance_schema/malloc_stats_201.sdi 2025-10-02T02:36:34.448514-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_128.sdi to /datadir/performance_schema/events_statement_128.sdi 2025-10-02T02:36:34.448566-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_128.sdi to /datadir/performance_schema/events_statement_128.sdi 2025-10-02T02:36:34.448595-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/keyring_componen_202.sdi to /datadir/performance_schema/keyring_componen_202.sdi 2025-10-02T02:36:34.448623-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/keyring_componen_202.sdi to /datadir/performance_schema/keyring_componen_202.sdi 2025-10-02T02:36:34.448652-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_132.sdi to /datadir/performance_schema/events_statement_132.sdi 2025-10-02T02:36:34.448684-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_132.sdi to /datadir/performance_schema/events_statement_132.sdi 2025-10-02T02:36:34.448714-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/prepared_stateme_184.sdi to /datadir/performance_schema/prepared_stateme_184.sdi 2025-10-02T02:36:34.448742-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/prepared_stateme_184.sdi to /datadir/performance_schema/prepared_stateme_184.sdi 2025-10-02T02:36:34.448785-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_his_95.sdi to /datadir/performance_schema/events_waits_his_95.sdi 2025-10-02T02:36:34.448809-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_his_95.sdi to /datadir/performance_schema/events_waits_his_95.sdi 2025-10-02T02:36:34.448833-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_174.sdi to /datadir/performance_schema/replication_appl_174.sdi 2025-10-02T02:36:34.448861-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_174.sdi to /datadir/performance_schema/replication_appl_174.sdi 2025-10-02T02:36:34.448886-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_objects_114.sdi to /datadir/performance_schema/setup_objects_114.sdi 2025-10-02T02:36:34.448904-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_objects_114.sdi to /datadir/performance_schema/setup_objects_114.sdi 2025-10-02T02:36:34.448932-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/cond_instances_91.sdi to /datadir/performance_schema/cond_instances_91.sdi 2025-10-02T02:36:34.448979-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/cond_instances_91.sdi to /datadir/performance_schema/cond_instances_91.sdi 2025-10-02T02:36:34.449001-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/mutex_instances_106.sdi to /datadir/performance_schema/mutex_instances_106.sdi 2025-10-02T02:36:34.449025-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/mutex_instances_106.sdi to /datadir/performance_schema/mutex_instances_106.sdi 2025-10-02T02:36:34.449052-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/host_cache_105.sdi to /datadir/performance_schema/host_cache_105.sdi 2025-10-02T02:36:34.449077-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/host_cache_105.sdi to /datadir/performance_schema/host_cache_105.sdi 2025-10-02T02:36:34.449107-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_130.sdi to /datadir/performance_schema/events_statement_130.sdi 2025-10-02T02:36:34.449137-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_130.sdi to /datadir/performance_schema/events_statement_130.sdi 2025-10-02T02:36:34.449169-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_grou_172.sdi to /datadir/performance_schema/replication_grou_172.sdi 2025-10-02T02:36:34.449198-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_grou_172.sdi to /datadir/performance_schema/replication_grou_172.sdi 2025-10-02T02:36:34.449227-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_126.sdi to /datadir/performance_schema/events_stages_su_126.sdi 2025-10-02T02:36:34.449256-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_126.sdi to /datadir/performance_schema/events_stages_su_126.sdi 2025-10-02T02:36:34.449285-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_152.sdi to /datadir/performance_schema/events_errors_su_152.sdi 2025-10-02T02:36:34.449312-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_152.sdi to /datadir/performance_schema/events_errors_su_152.sdi 2025-10-02T02:36:34.449343-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_177.sdi to /datadir/performance_schema/replication_appl_177.sdi 2025-10-02T02:36:34.449372-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_177.sdi to /datadir/performance_schema/replication_appl_177.sdi 2025-10-02T02:36:34.449403-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_139.sdi to /datadir/performance_schema/events_statement_139.sdi 2025-10-02T02:36:34.449433-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_139.sdi to /datadir/performance_schema/events_statement_139.sdi 2025-10-02T02:36:34.449462-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_host_187.sdi to /datadir/performance_schema/status_by_host_187.sdi 2025-10-02T02:36:34.449488-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_host_187.sdi to /datadir/performance_schema/status_by_host_187.sdi 2025-10-02T02:36:34.449515-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/malloc_stats_tot_200.sdi to /datadir/performance_schema/malloc_stats_tot_200.sdi 2025-10-02T02:36:34.449539-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/malloc_stats_tot_200.sdi to /datadir/performance_schema/malloc_stats_tot_200.sdi 2025-10-02T02:36:34.449564-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/global_status_190.sdi to /datadir/performance_schema/global_status_190.sdi 2025-10-02T02:36:34.449587-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/global_status_190.sdi to /datadir/performance_schema/global_status_190.sdi 2025-10-02T02:36:34.449613-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_180.sdi to /datadir/performance_schema/replication_appl_180.sdi 2025-10-02T02:36:34.449655-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_180.sdi to /datadir/performance_schema/replication_appl_180.sdi 2025-10-02T02:36:34.449686-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/keyring_keys_161.sdi to /datadir/performance_schema/keyring_keys_161.sdi 2025-10-02T02:36:34.449713-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/keyring_keys_161.sdi to /datadir/performance_schema/keyring_keys_161.sdi 2025-10-02T02:36:34.449744-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_163.sdi to /datadir/performance_schema/memory_summary_b_163.sdi 2025-10-02T02:36:34.449786-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_163.sdi to /datadir/performance_schema/memory_summary_b_163.sdi 2025-10-02T02:36:34.449816-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_149.sdi to /datadir/performance_schema/events_errors_su_149.sdi 2025-10-02T02:36:34.449848-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_149.sdi to /datadir/performance_schema/events_errors_su_149.sdi 2025-10-02T02:36:34.449880-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_147.sdi to /datadir/performance_schema/events_transacti_147.sdi 2025-10-02T02:36:34.449907-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_147.sdi to /datadir/performance_schema/events_transacti_147.sdi 2025-10-02T02:36:34.449938-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_142.sdi to /datadir/performance_schema/events_transacti_142.sdi 2025-10-02T02:36:34.449965-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_142.sdi to /datadir/performance_schema/events_transacti_142.sdi 2025-10-02T02:36:34.449992-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_133.sdi to /datadir/performance_schema/events_statement_133.sdi 2025-10-02T02:36:34.450015-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_133.sdi to /datadir/performance_schema/events_statement_133.sdi 2025-10-02T02:36:34.450043-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_96.sdi to /datadir/performance_schema/events_waits_sum_96.sdi 2025-10-02T02:36:34.450069-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_96.sdi to /datadir/performance_schema/events_waits_sum_96.sdi 2025-10-02T02:36:34.450117-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_150.sdi to /datadir/performance_schema/events_errors_su_150.sdi 2025-10-02T02:36:34.450139-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_150.sdi to /datadir/performance_schema/events_errors_su_150.sdi 2025-10-02T02:36:34.450165-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_io_waits_s_116.sdi to /datadir/performance_schema/table_io_waits_s_116.sdi 2025-10-02T02:36:34.450192-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_io_waits_s_116.sdi to /datadir/performance_schema/table_io_waits_s_116.sdi 2025-10-02T02:36:34.450214-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/users_153.sdi to /datadir/performance_schema/users_153.sdi 2025-10-02T02:36:34.450232-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/users_153.sdi to /datadir/performance_schema/users_153.sdi 2025-10-02T02:36:34.450258-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_141.sdi to /datadir/performance_schema/events_transacti_141.sdi 2025-10-02T02:36:34.450280-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_141.sdi to /datadir/performance_schema/events_transacti_141.sdi 2025-10-02T02:36:34.450298-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/variables_info_195.sdi to /datadir/performance_schema/variables_info_195.sdi 2025-10-02T02:36:34.450318-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/variables_info_195.sdi to /datadir/performance_schema/variables_info_195.sdi 2025-10-02T02:36:34.450346-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_user_189.sdi to /datadir/performance_schema/status_by_user_189.sdi 2025-10-02T02:36:34.450376-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_user_189.sdi to /datadir/performance_schema/status_by_user_189.sdi 2025-10-02T02:36:34.450409-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_g_162.sdi to /datadir/performance_schema/memory_summary_g_162.sdi 2025-10-02T02:36:34.450438-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_g_162.sdi to /datadir/performance_schema/memory_summary_g_162.sdi 2025-10-02T02:36:34.450469-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/global_variables_193.sdi to /datadir/performance_schema/global_variables_193.sdi 2025-10-02T02:36:34.450497-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/global_variables_193.sdi to /datadir/performance_schema/global_variables_193.sdi 2025-10-02T02:36:34.450528-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/performance_time_108.sdi to /datadir/performance_schema/performance_time_108.sdi 2025-10-02T02:36:34.450556-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/performance_time_108.sdi to /datadir/performance_schema/performance_time_108.sdi 2025-10-02T02:36:34.450588-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_lock_waits_118.sdi to /datadir/performance_schema/table_lock_waits_118.sdi 2025-10-02T02:36:34.450612-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_lock_waits_118.sdi to /datadir/performance_schema/table_lock_waits_118.sdi 2025-10-02T02:36:34.450640-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_138.sdi to /datadir/performance_schema/events_statement_138.sdi 2025-10-02T02:36:34.450669-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_138.sdi to /datadir/performance_schema/events_statement_138.sdi 2025-10-02T02:36:34.450701-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_instrument_113.sdi to /datadir/performance_schema/setup_instrument_113.sdi 2025-10-02T02:36:34.450730-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_instrument_113.sdi to /datadir/performance_schema/setup_instrument_113.sdi 2025-10-02T02:36:34.450782-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_179.sdi to /datadir/performance_schema/replication_appl_179.sdi 2025-10-02T02:36:34.450814-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_179.sdi to /datadir/performance_schema/replication_appl_179.sdi 2025-10-02T02:36:34.450844-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_status_191.sdi to /datadir/performance_schema/session_status_191.sdi 2025-10-02T02:36:34.450874-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_status_191.sdi to /datadir/performance_schema/session_status_191.sdi 2025-10-02T02:36:34.450895-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_146.sdi to /datadir/performance_schema/events_transacti_146.sdi 2025-10-02T02:36:34.450917-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_146.sdi to /datadir/performance_schema/events_transacti_146.sdi 2025-10-02T02:36:34.450948-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_125.sdi to /datadir/performance_schema/events_stages_su_125.sdi 2025-10-02T02:36:34.450976-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_125.sdi to /datadir/performance_schema/events_stages_su_125.sdi 2025-10-02T02:36:34.451002-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_summary_by__104.sdi to /datadir/performance_schema/file_summary_by__104.sdi 2025-10-02T02:36:34.451027-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_summary_by__104.sdi to /datadir/performance_schema/file_summary_by__104.sdi 2025-10-02T02:36:34.451053-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_hi_121.sdi to /datadir/performance_schema/events_stages_hi_121.sdi 2025-10-02T02:36:34.451076-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_hi_121.sdi to /datadir/performance_schema/events_stages_hi_121.sdi 2025-10-02T02:36:34.451109-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_135.sdi to /datadir/performance_schema/events_statement_135.sdi 2025-10-02T02:36:34.451134-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_135.sdi to /datadir/performance_schema/events_statement_135.sdi 2025-10-02T02:36:34.451166-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/tls_channel_stat_199.sdi to /datadir/performance_schema/tls_channel_stat_199.sdi 2025-10-02T02:36:34.451193-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/tls_channel_stat_199.sdi to /datadir/performance_schema/tls_channel_stat_199.sdi 2025-10-02T02:36:34.451225-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_actors_111.sdi to /datadir/performance_schema/setup_actors_111.sdi 2025-10-02T02:36:34.451253-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_actors_111.sdi to /datadir/performance_schema/setup_actors_111.sdi 2025-10-02T02:36:34.451288-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_connect__159.sdi to /datadir/performance_schema/session_connect__159.sdi 2025-10-02T02:36:34.451315-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_connect__159.sdi to /datadir/performance_schema/session_connect__159.sdi 2025-10-02T02:36:34.451344-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/error_log_92.sdi to /datadir/performance_schema/error_log_92.sdi 2025-10-02T02:36:34.451366-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/error_log_92.sdi to /datadir/performance_schema/error_log_92.sdi 2025-10-02T02:36:34.451390-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/variables_by_thr_192.sdi to /datadir/performance_schema/variables_by_thr_192.sdi 2025-10-02T02:36:34.451419-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/variables_by_thr_192.sdi to /datadir/performance_schema/variables_by_thr_192.sdi 2025-10-02T02:36:34.451451-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_127.sdi to /datadir/performance_schema/events_stages_su_127.sdi 2025-10-02T02:36:34.451482-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_127.sdi to /datadir/performance_schema/events_stages_su_127.sdi 2025-10-02T02:36:34.451515-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_164.sdi to /datadir/performance_schema/memory_summary_b_164.sdi 2025-10-02T02:36:34.451541-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_164.sdi to /datadir/performance_schema/memory_summary_b_164.sdi 2025-10-02T02:36:34.451572-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/objects_summary__107.sdi to /datadir/performance_schema/objects_summary__107.sdi 2025-10-02T02:36:34.451601-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/objects_summary__107.sdi to /datadir/performance_schema/objects_summary__107.sdi 2025-10-02T02:36:34.451631-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_io_waits_s_117.sdi to /datadir/performance_schema/table_io_waits_s_117.sdi 2025-10-02T02:36:34.451666-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_io_waits_s_117.sdi to /datadir/performance_schema/table_io_waits_s_117.sdi 2025-10-02T02:36:34.451697-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_handles_167.sdi to /datadir/performance_schema/table_handles_167.sdi 2025-10-02T02:36:34.451723-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_handles_167.sdi to /datadir/performance_schema/table_handles_167.sdi 2025-10-02T02:36:34.451752-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_summary_b_157.sdi to /datadir/performance_schema/socket_summary_b_157.sdi 2025-10-02T02:36:34.451787-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_summary_b_157.sdi to /datadir/performance_schema/socket_summary_b_157.sdi 2025-10-02T02:36:34.451817-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_124.sdi to /datadir/performance_schema/events_stages_su_124.sdi 2025-10-02T02:36:34.451845-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_124.sdi to /datadir/performance_schema/events_stages_su_124.sdi 2025-10-02T02:36:34.451878-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_hi_122.sdi to /datadir/performance_schema/events_stages_hi_122.sdi 2025-10-02T02:36:34.451905-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_hi_122.sdi to /datadir/performance_schema/events_stages_hi_122.sdi 2025-10-02T02:36:34.451938-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_grou_178.sdi to /datadir/performance_schema/replication_grou_178.sdi 2025-10-02T02:36:34.451969-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_grou_178.sdi to /datadir/performance_schema/replication_grou_178.sdi 2025-10-02T02:36:34.452003-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_137.sdi to /datadir/performance_schema/events_statement_137.sdi 2025-10-02T02:36:34.452031-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_137.sdi to /datadir/performance_schema/events_statement_137.sdi 2025-10-02T02:36:34.452060-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_134.sdi to /datadir/performance_schema/events_statement_134.sdi 2025-10-02T02:36:34.452083-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_134.sdi to /datadir/performance_schema/events_statement_134.sdi 2025-10-02T02:36:34.452111-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_consumers_112.sdi to /datadir/performance_schema/setup_consumers_112.sdi 2025-10-02T02:36:34.452136-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_consumers_112.sdi to /datadir/performance_schema/setup_consumers_112.sdi 2025-10-02T02:36:34.452166-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/accounts_154.sdi to /datadir/performance_schema/accounts_154.sdi 2025-10-02T02:36:34.452192-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/accounts_154.sdi to /datadir/performance_schema/accounts_154.sdi 2025-10-02T02:36:34.452223-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_101.sdi to /datadir/performance_schema/events_waits_sum_101.sdi 2025-10-02T02:36:34.452251-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_101.sdi to /datadir/performance_schema/events_waits_sum_101.sdi 2025-10-02T02:36:34.452275-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_instances_102.sdi to /datadir/performance_schema/file_instances_102.sdi 2025-10-02T02:36:34.452300-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_instances_102.sdi to /datadir/performance_schema/file_instances_102.sdi 2025-10-02T02:36:34.452331-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_summary_b_158.sdi to /datadir/performance_schema/socket_summary_b_158.sdi 2025-10-02T02:36:34.452363-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_summary_b_158.sdi to /datadir/performance_schema/socket_summary_b_158.sdi 2025-10-02T02:36:34.452385-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_cur_93.sdi to /datadir/performance_schema/events_waits_cur_93.sdi 2025-10-02T02:36:34.452410-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_cur_93.sdi to /datadir/performance_schema/events_waits_cur_93.sdi 2025-10-02T02:36:34.452442-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_variable_194.sdi to /datadir/performance_schema/session_variable_194.sdi 2025-10-02T02:36:34.452468-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_variable_194.sdi to /datadir/performance_schema/session_variable_194.sdi 2025-10-02T02:36:34.452496-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_148.sdi to /datadir/performance_schema/events_errors_su_148.sdi 2025-10-02T02:36:34.452527-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_148.sdi to /datadir/performance_schema/events_errors_su_148.sdi 2025-10-02T02:36:34.452562-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_asyn_182.sdi to /datadir/performance_schema/replication_asyn_182.sdi 2025-10-02T02:36:34.452589-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_asyn_182.sdi to /datadir/performance_schema/replication_asyn_182.sdi 2025-10-02T02:36:34.452616-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/persisted_variab_196.sdi to /datadir/performance_schema/persisted_variab_196.sdi 2025-10-02T02:36:34.452645-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/persisted_variab_196.sdi to /datadir/performance_schema/persisted_variab_196.sdi 2025-10-02T02:36:34.452679-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_144.sdi to /datadir/performance_schema/events_transacti_144.sdi 2025-10-02T02:36:34.452704-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_144.sdi to /datadir/performance_schema/events_transacti_144.sdi 2025-10-02T02:36:34.452736-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_conn_173.sdi to /datadir/performance_schema/replication_conn_173.sdi 2025-10-02T02:36:34.452774-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_conn_173.sdi to /datadir/performance_schema/replication_conn_173.sdi 2025-10-02T02:36:34.452804-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_136.sdi to /datadir/performance_schema/events_statement_136.sdi 2025-10-02T02:36:34.452829-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_136.sdi to /datadir/performance_schema/events_statement_136.sdi 2025-10-02T02:36:34.452855-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/processlist_109.sdi to /datadir/performance_schema/processlist_109.sdi 2025-10-02T02:36:34.452877-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/processlist_109.sdi to /datadir/performance_schema/processlist_109.sdi 2025-10-02T02:36:34.452902-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_summary_by__103.sdi to /datadir/performance_schema/file_summary_by__103.sdi 2025-10-02T02:36:34.452925-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_summary_by__103.sdi to /datadir/performance_schema/file_summary_by__103.sdi 2025-10-02T02:36:34.452953-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_123.sdi to /datadir/performance_schema/events_stages_su_123.sdi 2025-10-02T02:36:34.452982-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_123.sdi to /datadir/performance_schema/events_stages_su_123.sdi 2025-10-02T02:36:34.453011-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_129.sdi to /datadir/performance_schema/events_statement_129.sdi 2025-10-02T02:36:34.453040-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_129.sdi to /datadir/performance_schema/events_statement_129.sdi 2025-10-02T02:36:34.453073-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_cu_120.sdi to /datadir/performance_schema/events_stages_cu_120.sdi 2025-10-02T02:36:34.453102-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_cu_120.sdi to /datadir/performance_schema/events_stages_cu_120.sdi 2025-10-02T02:36:34.453135-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/binary_log_trans_198.sdi to /datadir/performance_schema/binary_log_trans_198.sdi 2025-10-02T02:36:34.453160-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/binary_log_trans_198.sdi to /datadir/performance_schema/binary_log_trans_198.sdi 2025-10-02T02:36:34.453191-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/user_variables_b_185.sdi to /datadir/performance_schema/user_variables_b_185.sdi 2025-10-02T02:36:34.453220-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/user_variables_b_185.sdi to /datadir/performance_schema/user_variables_b_185.sdi 2025-10-02T02:36:34.453249-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_151.sdi to /datadir/performance_schema/events_errors_su_151.sdi 2025-10-02T02:36:34.453269-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_151.sdi to /datadir/performance_schema/events_errors_su_151.sdi 2025-10-02T02:36:34.453336-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./myApp/myApp.ibd to /datadir/myApp/myApp.ibd 2025-10-02T02:36:34.453365-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./myApp/myApp.ibd to /datadir/myApp/myApp.ibd 2025-10-02T02:36:34.453395-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./ibtmp1 to /datadir/ibtmp1 2025-10-02T02:36:34.453421-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./ibtmp1 to /datadir/ibtmp1 2025-10-02T02:36:34.453449-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./xtrabackup_info to /datadir/xtrabackup_info 2025-10-02T02:36:34.453471-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./xtrabackup_info to /datadir/xtrabackup_info 2025-10-02T02:36:34.453542-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./sys/sys_config.ibd to /datadir/sys/sys_config.ibd 2025-10-02T02:36:34.453570-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./sys/sys_config.ibd to /datadir/sys/sys_config.ibd 2025-10-02T02:36:34.453632-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log.CSV to /datadir/mysql/slow_log.CSV 2025-10-02T02:36:34.453666-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log.CSV to /datadir/mysql/slow_log.CSV 2025-10-02T02:36:34.453697-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_streaming_log.ibd to /datadir/mysql/wsrep_streaming_log.ibd 2025-10-02T02:36:34.453720-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_streaming_log.ibd to /datadir/mysql/wsrep_streaming_log.ibd 2025-10-02T02:36:34.453744-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_cluster.ibd to /datadir/mysql/wsrep_cluster.ibd 2025-10-02T02:36:34.453795-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_cluster.ibd to /datadir/mysql/wsrep_cluster.ibd 2025-10-02T02:36:34.453827-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log_226.sdi to /datadir/mysql/slow_log_226.sdi 2025-10-02T02:36:34.453851-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log_226.sdi to /datadir/mysql/slow_log_226.sdi 2025-10-02T02:36:34.453871-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log_225.sdi to /datadir/mysql/general_log_225.sdi 2025-10-02T02:36:34.453887-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log_225.sdi to /datadir/mysql/general_log_225.sdi 2025-10-02T02:36:34.453914-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log.CSM to /datadir/mysql/general_log.CSM 2025-10-02T02:36:34.453936-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log.CSM to /datadir/mysql/general_log.CSM 2025-10-02T02:36:34.453959-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_cluster_members.ibd to /datadir/mysql/wsrep_cluster_members.ibd 2025-10-02T02:36:34.453976-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_cluster_members.ibd to /datadir/mysql/wsrep_cluster_members.ibd 2025-10-02T02:36:34.453995-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log.CSM to /datadir/mysql/slow_log.CSM 2025-10-02T02:36:34.454018-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log.CSM to /datadir/mysql/slow_log.CSM 2025-10-02T02:36:34.454044-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log.CSV to /datadir/mysql/general_log.CSV 2025-10-02T02:36:34.454067-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log.CSV to /datadir/mysql/general_log.CSV 2025-10-02T02:36:34.454091-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./xtrabackup_galera_info to /datadir/xtrabackup_galera_info 2025-10-02T02:36:34.454115-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./xtrabackup_galera_info to /datadir/xtrabackup_galera_info 2025-10-02T02:36:34.454144-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./sst_info to /datadir/sst_info 2025-10-02T02:36:34.454167-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./sst_info to /datadir/sst_info 2025-10-02T02:36:34.454173-00:00 1 [Note] [MY-011825] [Xtrabackup] Creating directory ./#innodb_redo 2025-10-02T02:36:34.454182-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: creating directory ./#innodb_redo 2025-10-02T02:36:34.547309-00:00 0 [Note] [MY-011825] [Xtrabackup] completed OK! + cat /tmp/tmp.y8wmezuma9 Defaulted container "xtrabackup" out of: xtrabackup, backup-init (init) + rm /tmp/tmp.NqBRJrsjts /tmp/tmp.y8wmezuma9 + return 0 + wait_for_running some-name-proxysql 1 + local name=some-name-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in $(seq 0 $last_pod) + wait_pod some-name-proxysql-0 480 + local pod=some-name-proxysql-0 + local max_retry=480 + local ns= ++ echo some-name-proxysql-0 ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container=proxysql + set +o xtrace pod/some-name-proxysql-0 condition met waiting for pod/some-name-proxysql-0 to become Ready.Ok + wait_for_running some-name-pxc 3 + local name=some-name-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in $(seq 0 $last_pod) + wait_pod some-name-pxc-0 480 + local pod=some-name-pxc-0 + local max_retry=480 + local ns= ++ echo some-name-pxc-0 ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container=pxc + set +o xtrace pod/some-name-pxc-0 condition met waiting for pod/some-name-pxc-0 to become Ready.Ok + for i in $(seq 0 $last_pod) + wait_pod some-name-pxc-1 480 + local pod=some-name-pxc-1 + local max_retry=480 + local ns= ++ echo some-name-pxc-1 ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container=pxc + set +o xtrace pod/some-name-pxc-1 condition met waiting for pod/some-name-pxc-1 to become Ready.Ok + for i in $(seq 0 $last_pod) + wait_pod some-name-pxc-2 480 + local pod=some-name-pxc-2 + local max_retry=480 + local ns= ++ echo some-name-pxc-2 ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container=pxc + set +o xtrace pod/some-name-pxc-2 condition met waiting for pod/some-name-pxc-2 to become Ready.Ok ++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.secretsName}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qTYm9b9bFE +++ mktemp ++ local LAST_ERR=/tmp/tmp.1eQNvKv354 ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.spec.secretsName}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qTYm9b9bFE ++ cat /tmp/tmp.1eQNvKv354 ++ rm /tmp/tmp.qTYm9b9bFE /tmp/tmp.1eQNvKv354 ++ return 0 + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.kPbStK188H +++ mktemp ++ local LAST_ERR=/tmp/tmp.GmoJXxw5wD ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kPbStK188H ++ cat /tmp/tmp.GmoJXxw5wD ++ rm /tmp/tmp.kPbStK188H /tmp/tmp.GmoJXxw5wD ++ return 0 + local root_pass=root_password + sleep 35 + log 'check data after pxc-restore/on-demand-backup-pvc' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-10-02T02:42:11+0000]' check data after pxc-restore/on-demand-backup-pvc [2025-10-02T02:42:11+0000] check data after pxc-restore/on-demand-backup-pvc + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\''' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/restore-to-encrypted-cluster/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\''' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.K893mVpN3J +++ mktemp ++ local LAST_ERR=/tmp/tmp.pvlwHtnIde ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.K893mVpN3J ++ cat /tmp/tmp.pvlwHtnIde ++ rm /tmp/tmp.K893mVpN3J /tmp/tmp.pvlwHtnIde ++ return 0 + client_pod=pxc-client-59944c5bbf-jrsmf + wait_pod pxc-client-59944c5bbf-jrsmf + local pod=pxc-client-59944c5bbf-jrsmf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-jrsmf ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container= + set +o xtrace pod/pxc-client-59944c5bbf-jrsmf condition met waiting for pod/pxc-client-59944c5bbf-jrsmf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.oBtCviH3md/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql /tmp/tmp.oBtCviH3md/select-1.sql + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\''' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/restore-to-encrypted-cluster/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\''' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.a8FtgDssVA +++ mktemp ++ local LAST_ERR=/tmp/tmp.7ISVb5Fwq7 ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.a8FtgDssVA ++ cat /tmp/tmp.7ISVb5Fwq7 ++ rm /tmp/tmp.a8FtgDssVA /tmp/tmp.7ISVb5Fwq7 ++ return 0 + client_pod=pxc-client-59944c5bbf-jrsmf + wait_pod pxc-client-59944c5bbf-jrsmf + local pod=pxc-client-59944c5bbf-jrsmf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-jrsmf ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container= + set +o xtrace pod/pxc-client-59944c5bbf-jrsmf condition met waiting for pod/pxc-client-59944c5bbf-jrsmf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.oBtCviH3md/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql /tmp/tmp.oBtCviH3md/select-1.sql + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\''' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/restore-to-encrypted-cluster/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\''' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1Z22oFyyA7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.puLHEjK0Yo ++ local exit_status=0 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1Z22oFyyA7 ++ cat /tmp/tmp.puLHEjK0Yo ++ rm /tmp/tmp.1Z22oFyyA7 /tmp/tmp.puLHEjK0Yo ++ return 0 + client_pod=pxc-client-59944c5bbf-jrsmf + wait_pod pxc-client-59944c5bbf-jrsmf + local pod=pxc-client-59944c5bbf-jrsmf + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-jrsmf ++ /usr/sbin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' egrep: warning: egrep is obsolescent; using grep -E + local container= + set +o xtrace pod/pxc-client-59944c5bbf-jrsmf condition met waiting for pod/pxc-client-59944c5bbf-jrsmf to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.oBtCviH3md/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql /tmp/tmp.oBtCviH3md/select-1.sql + '[' on-demand-backup-pvc '!=' on-demand-backup-minio ']' + log 'copy backup' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-10-02T02:42:29+0000]' copy backup [2025-10-02T02:42:29+0000] copy backup + '[' -n '' ']' + bash /mnt/jenkins/workspace/cloud-pxc-operator_PR-2203/deploy/backup/copy-backup.sh on-demand-backup-pvc /tmp/tmp.oBtCviH3md/backup which: no xbcloud in (/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) No xtrabackup binaries found, please install them: https://www.percona.com/downloads/Percona-XtraBackup-LATEST https://formulae.brew.sh/formula/percona-xtrabackup