Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/logs/demand-backup-encrypted-with-tls-8-0.log Warning: version difference between client (1.34) and server (1.31) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.34) and server (1.31) exceeds the supported minor version skew of +/-1 + main + create_infra demand-backup-encrypted-with-tls-13740 + local ns=demand-backup-encrypted-with-tls-13740 + '[' -n pxc-operator ']' + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl get pxc --all-namespaces -o wide + kubectl patch pxc -n demand-backup-encrypted-with-tls-6814 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.khvMURae5P ++ mktemp + local LAST_ERR=/tmp/tmp.UqiNlCcTPT + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.khvMURae5P perconaxtradbcluster.pxc.percona.com "some-name" deleted from demand-backup-encrypted-with-tls-6814 namespace + cat /tmp/tmp.UqiNlCcTPT + rm /tmp/tmp.khvMURae5P /tmp/tmp.UqiNlCcTPT + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.sqPXM3DFv6 ++ mktemp + local LAST_ERR=/tmp/tmp.Zfy5V7ZPES + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.sqPXM3DFv6 perconaxtradbclusterbackup.pxc.percona.com "on-demand-backup-pvc" deleted from demand-backup-encrypted-with-tls-6814 namespace + cat /tmp/tmp.Zfy5V7ZPES + rm /tmp/tmp.sqPXM3DFv6 /tmp/tmp.Zfy5V7ZPES + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.J1RkoFUQ4O ++ mktemp + local LAST_ERR=/tmp/tmp.WSjWTHlk7r + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.J1RkoFUQ4O perconaxtradbclusterrestore.pxc.percona.com "on-demand-backup-pvc" deleted from demand-backup-encrypted-with-tls-6814 namespace + cat /tmp/tmp.WSjWTHlk7r + rm /tmp/tmp.J1RkoFUQ4O /tmp/tmp.WSjWTHlk7r + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ sed s/NAMESPACE// ++ awk '-F ' '{print $2}' ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 + local chaos_mesh_ns= + '[' -n '' ']' ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get MutatingWebhookConfiguration + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get ValidatingWebhookConfiguration + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep validate-auth + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl api-resources ++ grep chaos-mesh.org ++ kubectl get crd ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ kubectl get clusterrolebinding ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get clusterrole + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + '[' -n '' ']' + desc 'cleaned up old namespaces pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pxc-operator + awk '{print$1}' + xargs kubectl delete ns ++ mktemp + local LAST_OUT=/tmp/tmp.5gwqQZNrW9 + egrep -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + kubectl_bin get ns ++ mktemp + local LAST_ERR=/tmp/tmp.z3zN7uglkn + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.E7kh8JgjgR ++ mktemp + local LAST_ERR=/tmp/tmp.l0pZWpg299 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.E7kh8JgjgR + cat /tmp/tmp.l0pZWpg299 + rm /tmp/tmp.E7kh8JgjgR /tmp/tmp.l0pZWpg299 + return 0 namespace "demand-backup-encrypted-with-tls-6814" deleted namespace "vault-service-1-28748" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.5gwqQZNrW9 namespace "pxc-operator" deleted + cat /tmp/tmp.z3zN7uglkn + rm /tmp/tmp.5gwqQZNrW9 /tmp/tmp.z3zN7uglkn + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + echo -n 'waiting for namespace/pxc-operator to be deleted' waiting for namespace/pxc-operator to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pxc-operator" not found + desc 'create namespace pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.jMEgCCPcSm ++ mktemp + local LAST_ERR=/tmp/tmp.JhsLKCp0LF + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.jMEgCCPcSm namespace/pxc-operator created + cat /tmp/tmp.JhsLKCp0LF + rm /tmp/tmp.jMEgCCPcSm /tmp/tmp.JhsLKCp0LF + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.IXU9LOx8ak +++ mktemp ++ local LAST_ERR=/tmp/tmp.lCl6A0UIla ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.IXU9LOx8ak ++ cat /tmp/tmp.lCl6A0UIla ++ rm /tmp/tmp.IXU9LOx8ak /tmp/tmp.lCl6A0UIla ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2207-89209ce1-16-cluster6 --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.sKZk9RKdT7 ++ mktemp + local LAST_ERR=/tmp/tmp.yQt2wqP8TT + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2207-89209ce1-16-cluster6 --namespace=pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.sKZk9RKdT7 Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2207-89209ce1-16-cluster6" modified. + cat /tmp/tmp.yQt2wqP8TT + rm /tmp/tmp.sKZk9RKdT7 /tmp/tmp.yQt2wqP8TT + return 0 + deploy_operator + desc 'start PXC operator' + set +o xtrace ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.JcYv7mZRuY ++ mktemp + local LAST_ERR=/tmp/tmp.ZwPsXSiGpO + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.JcYv7mZRuY customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied + cat /tmp/tmp.ZwPsXSiGpO + rm /tmp/tmp.JcYv7mZRuY /tmp/tmp.ZwPsXSiGpO + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/deploy/cw-rbac.yaml + kubectl_bin apply -f - + sed -e 's^namespace: .*^namespace: pxc-operator^' ++ mktemp + local LAST_OUT=/tmp/tmp.pjEDTHVL7i ++ mktemp + local LAST_ERR=/tmp/tmp.yBSpiDI9nF + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.pjEDTHVL7i clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.yBSpiDI9nF + rm /tmp/tmp.pjEDTHVL7i /tmp/tmp.yBSpiDI9nF + return 0 + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "DISABLE_TELEMETRY").value) = "true"' - + sed -e 's^failureThreshold: .*^failureThreshold: 10^' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/deploy/cw-operator.yaml + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "LOG_LEVEL").value) = "VERBOSE"' - + kubectl_bin apply -f - + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-2207-89209ce1^' ++ mktemp + local LAST_OUT=/tmp/tmp.Lri7WHY9tk ++ mktemp + local LAST_ERR=/tmp/tmp.ZIcrniST37 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Lri7WHY9tk deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.ZIcrniST37 + rm /tmp/tmp.Lri7WHY9tk /tmp/tmp.ZIcrniST37 + return 0 + sleep 10 + kubectl_bin wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s ++ mktemp + local LAST_OUT=/tmp/tmp.VxwozcgASg ++ mktemp + local LAST_ERR=/tmp/tmp.8qzWUAPYVY + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.VxwozcgASg pod/percona-xtradb-cluster-operator-58dd9fd94c-z8k2q condition met + cat /tmp/tmp.8qzWUAPYVY + rm /tmp/tmp.VxwozcgASg /tmp/tmp.8qzWUAPYVY + return 0 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.716pYkoCdx +++ mktemp ++ local LAST_ERR=/tmp/tmp.Lh70zbgIME ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.716pYkoCdx ++ cat /tmp/tmp.Lh70zbgIME ++ rm /tmp/tmp.716pYkoCdx /tmp/tmp.Lh70zbgIME ++ return 0 + wait_pod percona-xtradb-cluster-operator-58dd9fd94c-z8k2q 480 pxc-operator + local pod=percona-xtradb-cluster-operator-58dd9fd94c-z8k2q + local max_retry=480 + local ns=pxc-operator ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo percona-xtradb-cluster-operator-58dd9fd94c-z8k2q ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-58dd9fd94c-z8k2q condition met waiting for pod/percona-xtradb-cluster-operator-58dd9fd94c-z8k2q to become Ready.Ok + sleep 3 + create_namespace demand-backup-encrypted-with-tls-13740 + local namespace=demand-backup-encrypted-with-tls-13740 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// ++ helm list --all-namespaces --filter chaos-mesh + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep validate-auth + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ awk '{print $1}' ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get clusterrolebinding + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ kubectl get clusterrole ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + '[' -n '' ']' + desc 'cleaned up old namespaces demand-backup-encrypted-with-tls-13740' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces demand-backup-encrypted-with-tls-13740 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace demand-backup-encrypted-with-tls-13740 ++ mktemp + local LAST_OUT=/tmp/tmp.1f8xirORUQ + kubectl_bin get ns ++ mktemp + local LAST_ERR=/tmp/tmp.3jWH9kElM9 + local exit_status=0 ++ mktemp + local LAST_OUT=/tmp/tmp.c3aTyjT2Qe ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace demand-backup-encrypted-with-tls-13740 + xargs kubectl delete ns ++ mktemp + local LAST_ERR=/tmp/tmp.mv3ccC9q2T + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + awk '{print$1}' + egrep -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.c3aTyjT2Qe + cat /tmp/tmp.mv3ccC9q2T + rm /tmp/tmp.c3aTyjT2Qe /tmp/tmp.mv3ccC9q2T + return 0 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace demand-backup-encrypted-with-tls-13740 error: resource(s) were provided, but no name was specified + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace demand-backup-encrypted-with-tls-13740 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.1f8xirORUQ + cat /tmp/tmp.3jWH9kElM9 Error from server (NotFound): namespaces "demand-backup-encrypted-with-tls-13740" not found + rm /tmp/tmp.1f8xirORUQ /tmp/tmp.3jWH9kElM9 + return 1 + : + wait_for_delete namespace/demand-backup-encrypted-with-tls-13740 + local res=namespace/demand-backup-encrypted-with-tls-13740 + echo -n 'waiting for namespace/demand-backup-encrypted-with-tls-13740 to be deleted' waiting for namespace/demand-backup-encrypted-with-tls-13740 to be deleted+ set +o xtrace Error from server (NotFound): namespaces "demand-backup-encrypted-with-tls-13740" not found + desc 'create namespace demand-backup-encrypted-with-tls-13740' + set +o xtrace ----------------------------------------------------------------------------------- create namespace demand-backup-encrypted-with-tls-13740 ----------------------------------------------------------------------------------- + kubectl_bin create namespace demand-backup-encrypted-with-tls-13740 ++ mktemp + local LAST_OUT=/tmp/tmp.gkXlBCpAUl ++ mktemp + local LAST_ERR=/tmp/tmp.brLT45Jnqw + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace demand-backup-encrypted-with-tls-13740 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.gkXlBCpAUl namespace/demand-backup-encrypted-with-tls-13740 created + cat /tmp/tmp.brLT45Jnqw + rm /tmp/tmp.gkXlBCpAUl /tmp/tmp.brLT45Jnqw + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.5Xc3TqchZi +++ mktemp ++ local LAST_ERR=/tmp/tmp.8KkVKQZ8FK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.5Xc3TqchZi ++ cat /tmp/tmp.8KkVKQZ8FK ++ rm /tmp/tmp.5Xc3TqchZi /tmp/tmp.8KkVKQZ8FK ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2207-89209ce1-16-cluster6 --namespace=demand-backup-encrypted-with-tls-13740 ++ mktemp + local LAST_OUT=/tmp/tmp.31LJpXQ3zO ++ mktemp + local LAST_ERR=/tmp/tmp.lPtb5dIcYJ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2207-89209ce1-16-cluster6 --namespace=demand-backup-encrypted-with-tls-13740 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.31LJpXQ3zO Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2207-89209ce1-16-cluster6" modified. + cat /tmp/tmp.lPtb5dIcYJ + rm /tmp/tmp.31LJpXQ3zO /tmp/tmp.lPtb5dIcYJ + return 0 + apply_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.VH74XlNars ++ mktemp + local LAST_ERR=/tmp/tmp.wKiqf7ozV5 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.VH74XlNars secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.wKiqf7ozV5 + rm /tmp/tmp.VH74XlNars /tmp/tmp.wKiqf7ozV5 + return 0 + vault1=vault-service-1-14102 + protocol=https + start_vault vault-service-1-14102 https + name=vault-service-1-14102 + protocol=https + local platform=kubernetes + [[ -n '' ]] + create_namespace vault-service-1-14102 skip_clean + local namespace=vault-service-1-14102 + local skip_clean_namespace=skip_clean + [[ 1 == 1 ]] + [[ -z skip_clean ]] + '[' -n '' ']' + desc 'cleaned up old namespaces vault-service-1-14102' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces vault-service-1-14102 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace vault-service-1-14102 ++ mktemp + local LAST_OUT=/tmp/tmp.GHQv5iZk4W ++ mktemp + local LAST_ERR=/tmp/tmp.vyusOduO5d + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace vault-service-1-14102 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace vault-service-1-14102 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace vault-service-1-14102 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.GHQv5iZk4W + cat /tmp/tmp.vyusOduO5d Error from server (NotFound): namespaces "vault-service-1-14102" not found + rm /tmp/tmp.GHQv5iZk4W /tmp/tmp.vyusOduO5d + return 1 + : + wait_for_delete namespace/vault-service-1-14102 + local res=namespace/vault-service-1-14102 + echo -n 'waiting for namespace/vault-service-1-14102 to be deleted' waiting for namespace/vault-service-1-14102 to be deleted+ set +o xtrace Error from server (NotFound): namespaces "vault-service-1-14102" not found + desc 'create namespace vault-service-1-14102' + set +o xtrace ----------------------------------------------------------------------------------- create namespace vault-service-1-14102 ----------------------------------------------------------------------------------- + kubectl_bin create namespace vault-service-1-14102 ++ mktemp + local LAST_OUT=/tmp/tmp.q9bV7AOgrc ++ mktemp + local LAST_ERR=/tmp/tmp.fTHI7BzAuz + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace vault-service-1-14102 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.q9bV7AOgrc namespace/vault-service-1-14102 created + cat /tmp/tmp.fTHI7BzAuz + rm /tmp/tmp.q9bV7AOgrc /tmp/tmp.fTHI7BzAuz + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.k5RaNG7qJS +++ mktemp ++ local LAST_ERR=/tmp/tmp.riw9gsGu9F ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.k5RaNG7qJS ++ cat /tmp/tmp.riw9gsGu9F ++ rm /tmp/tmp.k5RaNG7qJS /tmp/tmp.riw9gsGu9F ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2207-89209ce1-16-cluster6 --namespace=vault-service-1-14102 ++ mktemp + local LAST_OUT=/tmp/tmp.G68O3ap8GN ++ mktemp + local LAST_ERR=/tmp/tmp.D0R6SP3opO + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2207-89209ce1-16-cluster6 --namespace=vault-service-1-14102 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.G68O3ap8GN Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2207-89209ce1-16-cluster6" modified. + cat /tmp/tmp.D0R6SP3opO + rm /tmp/tmp.G68O3ap8GN /tmp/tmp.D0R6SP3opO + return 0 + deploy_helm vault-service-1-14102 + helm repo add hashicorp https://helm.releases.hashicorp.com "hashicorp" already exists with the same configuration, skipping + helm repo add minio https://charts.min.io/ "minio" already exists with the same configuration, skipping + helm repo update Hang tight while we grab the latest from your chart repositories... ...Successfully got an update from the "minio" chart repository ...Successfully got an update from the "chaos-mesh" chart repository ...Successfully got an update from the "hashicorp" chart repository Update Complete. ⎈Happy Helming!⎈ + helm uninstall vault-service-1-14102 Error: uninstall: Release not loaded: vault-service-1-14102: release: not found + : + desc 'install Vault vault-service-1-14102' + set +o xtrace ----------------------------------------------------------------------------------- install Vault vault-service-1-14102 ----------------------------------------------------------------------------------- + '[' https == https ']' + vault_tls vault-service-1-14102 + local name=vault-service-1-14102 + SERVICE=vault-service-1-14102 + NAMESPACE=vault-service-1-14102 + SECRET_NAME=vault-service-1-14102 + CSR_NAME=vault-csr-6517 + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.31 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + CSR_API_VER=v1 + '[' 0 = 1 ']' + CSR_SIGNER=' signerName: kubernetes.io/kubelet-serving' + openssl genrsa -out /tmp/tmp.xEOJuOxVRz/vault.key 2048 Generating RSA private key, 2048 bit long modulus ...........+++ ........................+++ e is 65537 (0x10001) + cat + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.31 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + openssl req -new -key /tmp/tmp.xEOJuOxVRz/vault.key -subj '/CN=system:node:vault-service-1-14102.vault-service-1-14102.svc;/O=system:nodes' -out /tmp/tmp.xEOJuOxVRz/server.csr -config /tmp/tmp.xEOJuOxVRz/csr.conf + cat ++ cat /tmp/tmp.xEOJuOxVRz/server.csr ++ base64 ++ tr -d '\n' + kubectl_bin create -f /tmp/tmp.xEOJuOxVRz/csr.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.L29gQcvFsj ++ mktemp + local LAST_ERR=/tmp/tmp.tG9L49uV0s + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create -f /tmp/tmp.xEOJuOxVRz/csr.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.L29gQcvFsj certificatesigningrequest.certificates.k8s.io/vault-csr-6517 created + cat /tmp/tmp.tG9L49uV0s + rm /tmp/tmp.L29gQcvFsj /tmp/tmp.tG9L49uV0s + return 0 + sleep 10 + kubectl_bin certificate approve vault-csr-6517 ++ mktemp + local LAST_OUT=/tmp/tmp.KVptTJef06 ++ mktemp + local LAST_ERR=/tmp/tmp.fe7spUdhAT + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl certificate approve vault-csr-6517 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.KVptTJef06 certificatesigningrequest.certificates.k8s.io/vault-csr-6517 approved + cat /tmp/tmp.fe7spUdhAT + rm /tmp/tmp.KVptTJef06 /tmp/tmp.fe7spUdhAT + return 0 + kubectl_bin get csr vault-csr-6517 -o 'jsonpath={.status.certificate}' ++ mktemp + local LAST_OUT=/tmp/tmp.1PUcRBIiWz ++ mktemp + local LAST_ERR=/tmp/tmp.x6jg3AuxBW + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get csr vault-csr-6517 -o 'jsonpath={.status.certificate}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.1PUcRBIiWz + cat /tmp/tmp.x6jg3AuxBW + rm /tmp/tmp.1PUcRBIiWz /tmp/tmp.x6jg3AuxBW + return 0 + openssl base64 -in /tmp/tmp.xEOJuOxVRz/serverCert -d -A -out /tmp/tmp.xEOJuOxVRz/vault.crt + kubectl_bin config view --raw --minify --flatten -o 'jsonpath={.clusters[].cluster.certificate-authority-data}' + base64 -d ++ mktemp + local LAST_OUT=/tmp/tmp.GRaJB7ya6P ++ mktemp + local LAST_ERR=/tmp/tmp.wqtc41TwF2 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config view --raw --minify --flatten -o 'jsonpath={.clusters[].cluster.certificate-authority-data}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.GRaJB7ya6P + cat /tmp/tmp.wqtc41TwF2 + rm /tmp/tmp.GRaJB7ya6P /tmp/tmp.wqtc41TwF2 + return 0 + [[ -n '' ]] + kubectl_bin create secret generic vault-service-1-14102 --namespace vault-service-1-14102 --from-file=vault.key=/tmp/tmp.xEOJuOxVRz/vault.key --from-file=vault.crt=/tmp/tmp.xEOJuOxVRz/vault.crt --from-file=vault.ca=/tmp/tmp.xEOJuOxVRz/vault.ca ++ mktemp + local LAST_OUT=/tmp/tmp.O0jpvZTCti ++ mktemp + local LAST_ERR=/tmp/tmp.U2qsrvRiJn + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create secret generic vault-service-1-14102 --namespace vault-service-1-14102 --from-file=vault.key=/tmp/tmp.xEOJuOxVRz/vault.key --from-file=vault.crt=/tmp/tmp.xEOJuOxVRz/vault.crt --from-file=vault.ca=/tmp/tmp.xEOJuOxVRz/vault.ca + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.O0jpvZTCti secret/vault-service-1-14102 created + cat /tmp/tmp.U2qsrvRiJn + rm /tmp/tmp.O0jpvZTCti /tmp/tmp.U2qsrvRiJn + return 0 + helm install vault-service-1-14102 hashicorp/vault --disable-openapi-validation --version 0.30.0 --namespace vault-service-1-14102 --set dataStorage.enabled=false --set global.tlsDisable=false --set global.platform=kubernetes --set 'server.extraVolumes[0].type=secret' --set 'server.extraVolumes[0].name=vault-service-1-14102' --set server.extraEnvironmentVars.VAULT_CACERT=/vault/userconfig/vault-service-1-14102/vault.ca --set 'server.standalone.config= listener "tcp" { address = "[::]:8200" cluster_address = "[::]:8201" tls_cert_file = "/vault/userconfig/vault-service-1-14102/vault.crt" tls_key_file = "/vault/userconfig/vault-service-1-14102/vault.key" tls_client_ca_file = "/vault/userconfig/vault-service-1-14102/vault.ca" } storage "file" { path = "/vault/data" }' NAME: vault-service-1-14102 LAST DEPLOYED: Thu Nov 6 12:27:03 2025 NAMESPACE: vault-service-1-14102 STATUS: deployed REVISION: 1 NOTES: Thank you for installing HashiCorp Vault! Now that you have deployed Vault, you should look over the docs on using Vault with Kubernetes available here: https://developer.hashicorp.com/vault/docs Your release is named vault-service-1-14102. To learn more about the release, try: $ helm status vault-service-1-14102 $ helm get manifest vault-service-1-14102 + [[ -n '' ]] + set +o xtrace pod/vault-service-1-14102-0......{"running":{"startedAt":"2025-11-06T12:27:24Z"}} + kubectl_bin exec -it vault-service-1-14102-0 -- vault operator init -tls-skip-verify -key-shares=1 -key-threshold=1 -format=json ++ mktemp + local LAST_OUT=/tmp/tmp.rpF67vjtXW ++ mktemp + local LAST_ERR=/tmp/tmp.JRGT37Yalw + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec -it vault-service-1-14102-0 -- vault operator init -tls-skip-verify -key-shares=1 -key-threshold=1 -format=json + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.rpF67vjtXW + cat /tmp/tmp.JRGT37Yalw Unable to use a TTY - input is not a terminal or the right kind of file + rm /tmp/tmp.rpF67vjtXW /tmp/tmp.JRGT37Yalw + return 0 ++ jq -r '.unseal_keys_b64[]' + unsealKey=tGASdHkJl1SVhUnGUmSGr8IrLoaAfp5lwF2BuF6y6OQ= ++ jq -r .root_token + token=hvs.2subK9euJvHwF1aNZxvgIstb + sleep 10 + kubectl_bin exec -it vault-service-1-14102-0 -- vault operator unseal -tls-skip-verify tGASdHkJl1SVhUnGUmSGr8IrLoaAfp5lwF2BuF6y6OQ= ++ mktemp + local LAST_OUT=/tmp/tmp.KhoBBVyLby ++ mktemp + local LAST_ERR=/tmp/tmp.d81TxSyVGd + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec -it vault-service-1-14102-0 -- vault operator unseal -tls-skip-verify tGASdHkJl1SVhUnGUmSGr8IrLoaAfp5lwF2BuF6y6OQ= + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.KhoBBVyLby Key Value --- ----- Seal Type shamir Initialized true Sealed false Total Shares 1 Threshold 1 Version 1.19.0 Build Date 2025-03-04T12:36:40Z Storage Type file Cluster Name vault-cluster-04402b89 Cluster ID 1d003591-912b-03a6-041d-5d3fd500a2af HA Enabled false + cat /tmp/tmp.d81TxSyVGd Unable to use a TTY - input is not a terminal or the right kind of file + rm /tmp/tmp.KhoBBVyLby /tmp/tmp.d81TxSyVGd + return 0 + kubectl_bin exec -it vault-service-1-14102-0 -- sh -c 'export VAULT_TOKEN=hvs.2subK9euJvHwF1aNZxvgIstb && export VAULT_LOG_LEVEL=trace && vault secrets enable --version=1 -tls-skip-verify -path=secret kv && vault audit enable file file_path=/vault/vault-audit.log' ++ mktemp + local LAST_OUT=/tmp/tmp.aYDO3s5w78 ++ mktemp + local LAST_ERR=/tmp/tmp.i53lc9Ih3J + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec -it vault-service-1-14102-0 -- sh -c 'export VAULT_TOKEN=hvs.2subK9euJvHwF1aNZxvgIstb && export VAULT_LOG_LEVEL=trace && vault secrets enable --version=1 -tls-skip-verify -path=secret kv && vault audit enable file file_path=/vault/vault-audit.log' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.aYDO3s5w78 Success! Enabled the kv secrets engine at: secret/ Success! Enabled the file audit device at: file/ + cat /tmp/tmp.i53lc9Ih3J Unable to use a TTY - input is not a terminal or the right kind of file + rm /tmp/tmp.aYDO3s5w78 /tmp/tmp.i53lc9Ih3J + return 0 + sleep 10 + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/vault-secret.yaml + sed -e s/#token/hvs.2subK9euJvHwF1aNZxvgIstb/ + sed -e s/#secret/secret/ + sed -e 's/#vault_url/https:\/\/vault-service-1-14102.vault-service-1-14102.svc.cluster.local:8200/' + '[' https == https ']' + sed -e 's/^/ /' /tmp/tmp.xEOJuOxVRz/vault.ca + /usr/bin/sed -i s/#vault_ca/vault_ca/ /tmp/tmp.xEOJuOxVRz/vault-secret.yaml + /usr/bin/sed -i '/#certVal/r /tmp/tmp.xEOJuOxVRz/vault.new.ca' /tmp/tmp.xEOJuOxVRz/vault-secret.yaml + /usr/bin/sed -i /#certVal/d /tmp/tmp.xEOJuOxVRz/vault-secret.yaml + kubectl_bin apply --namespace=demand-backup-encrypted-with-tls-13740 -f /tmp/tmp.xEOJuOxVRz/vault-secret.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.ZY8xvS5Q9d ++ mktemp + local LAST_ERR=/tmp/tmp.imwoZAqdFb + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --namespace=demand-backup-encrypted-with-tls-13740 -f /tmp/tmp.xEOJuOxVRz/vault-secret.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ZY8xvS5Q9d secret/some-name-vault created + cat /tmp/tmp.imwoZAqdFb + rm /tmp/tmp.ZY8xvS5Q9d /tmp/tmp.imwoZAqdFb + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.Q9j8HGx9zI +++ mktemp ++ local LAST_ERR=/tmp/tmp.anGl2xaxLb ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Q9j8HGx9zI ++ cat /tmp/tmp.anGl2xaxLb ++ rm /tmp/tmp.Q9j8HGx9zI /tmp/tmp.anGl2xaxLb ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2207-89209ce1-16-cluster6 --namespace=demand-backup-encrypted-with-tls-13740 ++ mktemp + local LAST_OUT=/tmp/tmp.MgXaM3fZmO ++ mktemp + local LAST_ERR=/tmp/tmp.dE2RanVsIu + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2207-89209ce1-16-cluster6 --namespace=demand-backup-encrypted-with-tls-13740 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.MgXaM3fZmO Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2207-89209ce1-16-cluster6" modified. + cat /tmp/tmp.dE2RanVsIu + rm /tmp/tmp.MgXaM3fZmO /tmp/tmp.dE2RanVsIu + return 0 ++ jq -r .root_token + token1=hvs.2subK9euJvHwF1aNZxvgIstb + ip1=https://vault-service-1-14102.vault-service-1-14102.svc.cluster.local + cluster=some-name + spinup_pxc some-name /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/some-name.yml + local cluster=some-name + local config=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/some-name.yml + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/secrets.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/client.yml + local port=3306 + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.YJJMm9hrhU ++ mktemp + local LAST_ERR=/tmp/tmp.LQzTj6qit9 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.YJJMm9hrhU secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created + cat /tmp/tmp.LQzTj6qit9 + rm /tmp/tmp.YJJMm9hrhU /tmp/tmp.LQzTj6qit9 + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/client.yml + '[' -z '' ']' + kubectl_bin apply -f - + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/client.yml + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.demand-backup-encrypted-with-tls-13740~ + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2207-89209ce1#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' ++ mktemp + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + local LAST_OUT=/tmp/tmp.YtbCw8pa45 ++ mktemp + local LAST_ERR=/tmp/tmp.kiy9j7D9bz + local exit_status=0 ++ seq 0 2 + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/client.yml + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.YtbCw8pa45 deployment.apps/pxc-client created + cat /tmp/tmp.kiy9j7D9bz + rm /tmp/tmp.YtbCw8pa45 /tmp/tmp.kiy9j7D9bz + return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/some-name.yml + '[' -z '' ']' + kubectl_bin apply -f - + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/some-name.yml + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/conf/some-name.yml ++ mktemp + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + local LAST_OUT=/tmp/tmp.V51CEL9DeJ + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2207-89209ce1#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.demand-backup-encrypted-with-tls-13740~ + /usr/bin/sed -e 's#apply:.*#apply: Never#' ++ mktemp + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + local LAST_ERR=/tmp/tmp.GavSAfsNw7 + local exit_status=0 + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' ++ seq 0 2 + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.V51CEL9DeJ perconaxtradbcluster.pxc.percona.com/some-name created + cat /tmp/tmp.GavSAfsNw7 + rm /tmp/tmp.V51CEL9DeJ /tmp/tmp.GavSAfsNw7 + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- ++ get_proxy some-name ++ local target_cluster=some-name +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Jr31PgfM53 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ZUO0TJHUZY +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.Jr31PgfM53 +++ cat /tmp/tmp.ZUO0TJHUZY +++ rm /tmp/tmp.Jr31PgfM53 /tmp/tmp.ZUO0TJHUZY +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.aKsA7GxVhK ++++ mktemp +++ local LAST_ERR=/tmp/tmp.2U9euuWCdO +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.aKsA7GxVhK +++ cat /tmp/tmp.2U9euuWCdO +++ rm /tmp/tmp.aKsA7GxVhK /tmp/tmp.2U9euuWCdO +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo some-name-proxysql ++ return + local proxy=some-name-proxysql + kubectl_bin wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n demand-backup-encrypted-with-tls-13740 ++ mktemp + local LAST_OUT=/tmp/tmp.0SUzDEeMfH ++ mktemp + local LAST_ERR=/tmp/tmp.DINuSafyCQ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n demand-backup-encrypted-with-tls-13740 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n demand-backup-encrypted-with-tls-13740 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n demand-backup-encrypted-with-tls-13740 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.0SUzDEeMfH + cat /tmp/tmp.DINuSafyCQ error: no matching resources found + rm /tmp/tmp.0SUzDEeMfH /tmp/tmp.DINuSafyCQ + return 1 + true + wait_for_running some-name-proxysql 1 + local name=some-name-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-proxysql-0 480 + local pod=some-name-proxysql-0 + local max_retry=480 + local ns= ++ echo some-name-proxysql-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=proxysql + set +o xtrace pod/some-name-proxysql-0 condition met waiting for pod/some-name-proxysql-0 to become Ready.Ok + wait_for_running some-name-pxc 3 + local name=some-name-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-0 480 + local pod=some-name-pxc-0 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' ++ echo some-name-pxc-0 + local container=pxc + set +o xtrace pod/some-name-pxc-0 condition met waiting for pod/some-name-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-1 480 + local pod=some-name-pxc-1 + local max_retry=480 + local ns= ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo some-name-pxc-1 + local container=pxc + set +o xtrace pod/some-name-pxc-1 condition met waiting for pod/some-name-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-2 480 + local pod=some-name-pxc-2 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo some-name-pxc-2 ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-2 condition met waiting for pod/some-name-pxc-2 to become Ready.Ok + sleep 10 ++ kubectl get pxc some-name -o 'jsonpath={.spec.secretsName}' + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.DI7yttEC0R +++ mktemp ++ local LAST_ERR=/tmp/tmp.vVHVYrwh8f ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DI7yttEC0R ++ cat /tmp/tmp.vVHVYrwh8f ++ rm /tmp/tmp.DI7yttEC0R /tmp/tmp.vVHVYrwh8f ++ return 0 + local root_pass=root_password + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' '-h some-name-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' + local 'uri=-h some-name-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2qD7tx5OAf +++ mktemp ++ local LAST_ERR=/tmp/tmp.6yI8wV2TJz ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2qD7tx5OAf ++ cat /tmp/tmp.6yI8wV2TJz ++ rm /tmp/tmp.2qD7tx5OAf /tmp/tmp.6yI8wV2TJz ++ return 0 + client_pod=pxc-client-59944c5bbf-sh4xv + wait_pod pxc-client-59944c5bbf-sh4xv + local pod=pxc-client-59944c5bbf-sh4xv + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' ++ echo pxc-client-59944c5bbf-sh4xv + local container= + set +o xtrace pod/pxc-client-59944c5bbf-sh4xv condition met waiting for pod/pxc-client-59944c5bbf-sh4xv to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h some-name-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h some-name-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HvffRYe3HM +++ mktemp ++ local LAST_ERR=/tmp/tmp.FC205GZGRE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.HvffRYe3HM ++ cat /tmp/tmp.FC205GZGRE ++ rm /tmp/tmp.HvffRYe3HM /tmp/tmp.FC205GZGRE ++ return 0 + client_pod=pxc-client-59944c5bbf-sh4xv + wait_pod pxc-client-59944c5bbf-sh4xv + local pod=pxc-client-59944c5bbf-sh4xv + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-sh4xv ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-sh4xv condition met waiting for pod/pxc-client-59944c5bbf-sh4xv to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/demand-backup-encrypted-with-tls/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/demand-backup-encrypted-with-tls/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HEvbYQ4e27 +++ mktemp ++ local LAST_ERR=/tmp/tmp.qjHdj9dHKR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.HEvbYQ4e27 ++ cat /tmp/tmp.qjHdj9dHKR ++ rm /tmp/tmp.HEvbYQ4e27 /tmp/tmp.qjHdj9dHKR ++ return 0 + client_pod=pxc-client-59944c5bbf-sh4xv + wait_pod pxc-client-59944c5bbf-sh4xv + local pod=pxc-client-59944c5bbf-sh4xv + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' ++ echo pxc-client-59944c5bbf-sh4xv + local container= + set +o xtrace pod/pxc-client-59944c5bbf-sh4xv condition met waiting for pod/pxc-client-59944c5bbf-sh4xv to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.xEOJuOxVRz/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/demand-backup-encrypted-with-tls/compare/select-1.sql /tmp/tmp.xEOJuOxVRz/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/demand-backup-encrypted-with-tls/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/demand-backup-encrypted-with-tls/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XtbwZxpKkK +++ mktemp ++ local LAST_ERR=/tmp/tmp.TdNlCJmndS ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XtbwZxpKkK ++ cat /tmp/tmp.TdNlCJmndS ++ rm /tmp/tmp.XtbwZxpKkK /tmp/tmp.TdNlCJmndS ++ return 0 + client_pod=pxc-client-59944c5bbf-sh4xv + wait_pod pxc-client-59944c5bbf-sh4xv + local pod=pxc-client-59944c5bbf-sh4xv + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-sh4xv ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-sh4xv condition met waiting for pod/pxc-client-59944c5bbf-sh4xv to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.xEOJuOxVRz/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/demand-backup-encrypted-with-tls/compare/select-1.sql /tmp/tmp.xEOJuOxVRz/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/demand-backup-encrypted-with-tls/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/demand-backup-encrypted-with-tls/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QUL5lMOYNm +++ mktemp ++ local LAST_ERR=/tmp/tmp.NDyX4ArArB ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.QUL5lMOYNm ++ cat /tmp/tmp.NDyX4ArArB ++ rm /tmp/tmp.QUL5lMOYNm /tmp/tmp.NDyX4ArArB ++ return 0 + client_pod=pxc-client-59944c5bbf-sh4xv + wait_pod pxc-client-59944c5bbf-sh4xv + local pod=pxc-client-59944c5bbf-sh4xv + local max_retry=480 + local ns= ++ egrep '^(pxc|proxysql)$' ++ echo pxc-client-59944c5bbf-sh4xv ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-sh4xv condition met waiting for pod/pxc-client-59944c5bbf-sh4xv to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.xEOJuOxVRz/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/demand-backup-encrypted-with-tls/compare/select-1.sql /tmp/tmp.xEOJuOxVRz/select-1.sql ++ is_keyring_plugin_in_use some-name ++ local cluster=some-name ++ kubectl_bin exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ egrep -o 'early-plugin-load=keyring_\w+.so' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9XC6fr9qfv +++ mktemp ++ local LAST_ERR=/tmp/tmp.BUAVWpoogr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9XC6fr9qfv ++ cat /tmp/tmp.BUAVWpoogr Unable to use a TTY - input is not a terminal or the right kind of file ++ rm /tmp/tmp.9XC6fr9qfv /tmp/tmp.BUAVWpoogr ++ return 0 + '[' early-plugin-load=keyring_vault.so ']' + table_must_be_encrypted some-name myApp + desc 'check table encryption' + set +o xtrace ----------------------------------------------------------------------------------- check table encryption ----------------------------------------------------------------------------------- + local cluster=some-name + local table=myApp + is_table_encrypted some-name myApp + local cluster=some-name + local table=myApp + run_mysql 'SELECT CREATE_OPTIONS FROM INFORMATION_SCHEMA.TABLES WHERE TABLE_NAME=\"myApp\";' '-h some-name-proxysql -uroot -proot_password' + local 'command=SELECT CREATE_OPTIONS FROM INFORMATION_SCHEMA.TABLES WHERE TABLE_NAME=\"myApp\";' + local 'uri=-h some-name-proxysql -uroot -proot_password' + egrep -o 'ENCRYPTION=('\''Y'\''|"Y")' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vakrjk9faO +++ mktemp ++ local LAST_ERR=/tmp/tmp.9yvHPYYMBu ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.vakrjk9faO ++ cat /tmp/tmp.9yvHPYYMBu ++ rm /tmp/tmp.vakrjk9faO /tmp/tmp.9yvHPYYMBu ++ return 0 + client_pod=pxc-client-59944c5bbf-sh4xv + wait_pod pxc-client-59944c5bbf-sh4xv + local pod=pxc-client-59944c5bbf-sh4xv + local max_retry=480 + local ns= ++ egrep '^(pxc|proxysql)$' ++ echo pxc-client-59944c5bbf-sh4xv ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-sh4xv condition met waiting for pod/pxc-client-59944c5bbf-sh4xv to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace ENCRYPTION='Y' + keyring_plugin_must_be_in_use some-name + local cluster=some-name + desc 'check keyring plugin usage' + set +o xtrace ----------------------------------------------------------------------------------- check keyring plugin usage ----------------------------------------------------------------------------------- + is_keyring_plugin_in_use some-name + local cluster=some-name + kubectl_bin exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' + egrep -o 'early-plugin-load=keyring_\w+.so' ++ mktemp + local LAST_OUT=/tmp/tmp.vYNxeqpOfO ++ mktemp + local LAST_ERR=/tmp/tmp.ocsyOMPhuX + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.vYNxeqpOfO + cat /tmp/tmp.ocsyOMPhuX Unable to use a TTY - input is not a terminal or the right kind of file + rm /tmp/tmp.vYNxeqpOfO /tmp/tmp.ocsyOMPhuX + return 0 early-plugin-load=keyring_vault.so + table_must_be_encrypted some-name myApp + desc 'check table encryption' + set +o xtrace ----------------------------------------------------------------------------------- check table encryption ----------------------------------------------------------------------------------- + local cluster=some-name + local table=myApp + is_table_encrypted some-name myApp + local cluster=some-name + local table=myApp + egrep -o 'ENCRYPTION=('\''Y'\''|"Y")' + run_mysql 'SELECT CREATE_OPTIONS FROM INFORMATION_SCHEMA.TABLES WHERE TABLE_NAME=\"myApp\";' '-h some-name-proxysql -uroot -proot_password' + local 'command=SELECT CREATE_OPTIONS FROM INFORMATION_SCHEMA.TABLES WHERE TABLE_NAME=\"myApp\";' + local 'uri=-h some-name-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.np0HQHiPoW +++ mktemp ++ local LAST_ERR=/tmp/tmp.BlgiwAwbDw ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.np0HQHiPoW ++ cat /tmp/tmp.BlgiwAwbDw ++ rm /tmp/tmp.np0HQHiPoW /tmp/tmp.BlgiwAwbDw ++ return 0 + client_pod=pxc-client-59944c5bbf-sh4xv + wait_pod pxc-client-59944c5bbf-sh4xv + local pod=pxc-client-59944c5bbf-sh4xv + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-sh4xv ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-sh4xv condition met waiting for pod/pxc-client-59944c5bbf-sh4xv to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace ENCRYPTION='Y' + run_backup some-name on-demand-backup-pvc + local cluster=some-name + local backup=on-demand-backup-pvc + log 'run pxc-backup/on-demand-backup-pvc' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-11-06T12:33:25+0000]' run pxc-backup/on-demand-backup-pvc [2025-11-06T12:33:25+0000] run pxc-backup/on-demand-backup-pvc + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/demand-backup-encrypted-with-tls/conf/on-demand-backup-pvc.yml ++ mktemp + local LAST_OUT=/tmp/tmp.3cBPdcrPU2 ++ mktemp + local LAST_ERR=/tmp/tmp.Xz0VMqCdtM + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/demand-backup-encrypted-with-tls/conf/on-demand-backup-pvc.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.3cBPdcrPU2 perconaxtradbclusterbackup.pxc.percona.com/on-demand-backup-pvc created + cat /tmp/tmp.Xz0VMqCdtM + rm /tmp/tmp.3cBPdcrPU2 /tmp/tmp.Xz0VMqCdtM + return 0 + wait_backup on-demand-backup-pvc + local backup=on-demand-backup-pvc + local status=Succeeded + set +o xtrace waiting for pxc-backup/on-demand-backup-pvc to reach Succeeded state........................Succeeded + run_recovery_check some-name on-demand-backup-pvc + local cluster=some-name + local backup=on-demand-backup-pvc ++ get_proxy_engine some-name ++ local cluster_name=some-name +++ get_proxy some-name +++ local target_cluster=some-name ++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.cnjT2a3GlA +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.JKpil3Zag6 ++++ local exit_status=0 +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ set +e ++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 ']' ++++ break ++++ cat /tmp/tmp.cnjT2a3GlA ++++ cat /tmp/tmp.JKpil3Zag6 ++++ rm /tmp/tmp.cnjT2a3GlA /tmp/tmp.JKpil3Zag6 ++++ return 0 +++ [[ '' == \t\r\u\e ]] ++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.kopDQPHKoF +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.9f3E8vAFN4 ++++ local exit_status=0 +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ set +e ++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 ']' ++++ break ++++ cat /tmp/tmp.kopDQPHKoF ++++ cat /tmp/tmp.9f3E8vAFN4 ++++ rm /tmp/tmp.kopDQPHKoF /tmp/tmp.9f3E8vAFN4 ++++ return 0 +++ [[ true == \t\r\u\e ]] +++ echo some-name-proxysql +++ return ++ local cluster_proxy=some-name-proxysql ++ echo proxysql + local proxy=proxysql + log 'run pxc-restore/on-demand-backup-pvc' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-11-06T12:34:10+0000]' run pxc-restore/on-demand-backup-pvc [2025-11-06T12:34:10+0000] run pxc-restore/on-demand-backup-pvc + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/demand-backup-encrypted-with-tls/conf/restore-on-demand-backup-pvc.yaml + /usr/bin/sed -e s~minio-service.#namespace~minio-service.demand-backup-encrypted-with-tls-13740~ + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.J4HWzGJHIi ++ mktemp + local LAST_ERR=/tmp/tmp.86Yr43i4yk + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.J4HWzGJHIi perconaxtradbclusterrestore.pxc.percona.com/on-demand-backup-pvc created + cat /tmp/tmp.86Yr43i4yk + rm /tmp/tmp.J4HWzGJHIi /tmp/tmp.86Yr43i4yk + return 0 + wait_backup_restore on-demand-backup-pvc + local backup_name=on-demand-backup-pvc + local target_state=Succeeded + local wait_time=720 + set +o xtrace waiting for pxc-restore/on-demand-backup-pvc to reach Succeeded state 2025-11-06T12:34:14 pxc-restore/on-demand-backup-pvc state: Starting 2025-11-06T12:34:16 pxc-restore/on-demand-backup-pvc state: Starting 2025-11-06T12:34:18 pxc-restore/on-demand-backup-pvc state: Starting 2025-11-06T12:34:20 pxc-restore/on-demand-backup-pvc state: Starting 2025-11-06T12:34:22 pxc-restore/on-demand-backup-pvc state: Starting 2025-11-06T12:34:25 pxc-restore/on-demand-backup-pvc state: Starting 2025-11-06T12:34:27 pxc-restore/on-demand-backup-pvc state: Starting 2025-11-06T12:34:29 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:34:32 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:34:34 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:34:36 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:34:38 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:34:40 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:34:42 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:34:44 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:34:46 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:34:49 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:34:51 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:34:53 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:34:55 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:34:57 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:34:59 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:35:01 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:35:03 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:35:05 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:35:07 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:35:09 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:35:11 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:35:14 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:35:16 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:35:18 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:35:20 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:35:22 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:35:24 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:35:26 pxc-restore/on-demand-backup-pvc state: Stopping Cluster 2025-11-06T12:35:29 pxc-restore/on-demand-backup-pvc state: Restoring 2025-11-06T12:35:31 pxc-restore/on-demand-backup-pvc state: Restoring 2025-11-06T12:35:33 pxc-restore/on-demand-backup-pvc state: Restoring 2025-11-06T12:35:35 pxc-restore/on-demand-backup-pvc state: Restoring 2025-11-06T12:35:37 pxc-restore/on-demand-backup-pvc state: Restoring 2025-11-06T12:35:39 pxc-restore/on-demand-backup-pvc state: Restoring 2025-11-06T12:35:41 pxc-restore/on-demand-backup-pvc state: Restoring 2025-11-06T12:35:43 pxc-restore/on-demand-backup-pvc state: Restoring 2025-11-06T12:35:45 pxc-restore/on-demand-backup-pvc state: Restoring 2025-11-06T12:35:47 pxc-restore/on-demand-backup-pvc state: Restoring 2025-11-06T12:35:49 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:35:51 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:35:53 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:35:55 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:35:58 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:00 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:02 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:04 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:07 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:09 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:11 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:13 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:15 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:17 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:19 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:22 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:24 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:26 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:28 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:30 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:32 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:34 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:36 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:38 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:41 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:43 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:45 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:47 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:49 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:51 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:54 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:56 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:36:58 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:37:00 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:37:03 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:37:05 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:37:08 pxc-restore/on-demand-backup-pvc state: Preparing Cluster 2025-11-06T12:37:10 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:37:12 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:37:14 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:37:16 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:37:19 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:37:21 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:37:24 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:37:26 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:37:28 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:37:30 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:37:32 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:37:34 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:37:37 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:37:39 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:37:41 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:37:44 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:37:46 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:37:48 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:37:51 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:37:53 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:37:55 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:37:57 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:37:59 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:01 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:03 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:05 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:07 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:09 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:11 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:13 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:15 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:17 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:19 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:21 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:23 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:26 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:28 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:30 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:32 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:34 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:36 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:38 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:40 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:42 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:44 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:47 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:49 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:51 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:53 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:56 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:38:58 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:00 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:03 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:05 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:07 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:09 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:11 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:14 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:16 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:18 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:20 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:22 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:24 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:26 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:29 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:31 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:33 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:35 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:37 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:39 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:42 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:45 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:47 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:50 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:53 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:57 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:39:59 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:40:02 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:40:05 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:40:07 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:40:10 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:40:12 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:40:14 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:40:16 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:40:18 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:40:20 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:40:23 pxc-restore/on-demand-backup-pvc state: Starting Cluster 2025-11-06T12:40:25 pxc-restore/on-demand-backup-pvc state: Succeeded + kubectl_bin logs job/restore-job-on-demand-backup-pvc-some-name ++ mktemp + local LAST_OUT=/tmp/tmp.uDUFLivJLR ++ mktemp + local LAST_ERR=/tmp/tmp.cBgDjXpD8S + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl logs job/restore-job-on-demand-backup-pvc-some-name + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.uDUFLivJLR + LIB_PATH=/opt/percona/backup/lib/pxc + . /opt/percona/backup/lib/pxc/check-version.sh + . /opt/percona/backup/lib/pxc/vault.sh ++ set -o errexit ++ keyring_vault=/etc/mysql/vault-keyring-secret/keyring_vault.conf + SOCAT_OPTS=TCP:restore-src-on-demand-backup-pvc-some-name:3307,retry=30 + check_ssl + CA=/var/run/secrets/kubernetes.io/serviceaccount/ca.crt + '[' -f /var/run/secrets/kubernetes.io/serviceaccount/service-ca.crt ']' + SSL_DIR=/etc/mysql/ssl + '[' -f /etc/mysql/ssl/ca.crt ']' + CA=/etc/mysql/ssl/ca.crt + SSL_INTERNAL_DIR=/etc/mysql/ssl-internal + '[' -f /etc/mysql/ssl-internal/ca.crt ']' + CA=/etc/mysql/ssl-internal/ca.crt + KEY=/etc/mysql/ssl/tls.key + CERT=/etc/mysql/ssl/tls.crt + '[' -f /etc/mysql/ssl-internal/tls.key ']' + '[' -f /etc/mysql/ssl-internal/tls.crt ']' + KEY=/etc/mysql/ssl-internal/tls.key + CERT=/etc/mysql/ssl-internal/tls.crt + '[' -f /etc/mysql/ssl-internal/ca.crt ']' + '[' -f /etc/mysql/ssl-internal/tls.key ']' + '[' -f /etc/mysql/ssl-internal/tls.crt ']' + SOCAT_OPTS='openssl-connect:restore-src-on-demand-backup-pvc-some-name:3307,reuseaddr,cert=/etc/mysql/ssl-internal/tls.crt,key=/etc/mysql/ssl-internal/tls.key,cafile=/etc/mysql/ssl-internal/ca.crt,verify=1,commonname='\'''\'',retry=30,no-sni=1' + ping -c1 restore-src-on-demand-backup-pvc-some-name /opt/percona/backup/recovery-pvc-joiner.sh: line 40: ping: command not found + : + rm -rf /datadir/#ib_16384_0.dblwr /datadir/#ib_16384_1.dblwr /datadir/#ib_16384_10.dblwr /datadir/#ib_16384_11.dblwr /datadir/#ib_16384_12.dblwr /datadir/#ib_16384_13.dblwr /datadir/#ib_16384_14.dblwr /datadir/#ib_16384_15.dblwr /datadir/#ib_16384_2.dblwr /datadir/#ib_16384_3.dblwr /datadir/#ib_16384_4.dblwr /datadir/#ib_16384_5.dblwr /datadir/#ib_16384_6.dblwr /datadir/#ib_16384_7.dblwr /datadir/#ib_16384_8.dblwr /datadir/#ib_16384_9.dblwr /datadir/#innodb_redo /datadir/#innodb_temp /datadir/auth_plugin /datadir/auto.cnf /datadir/binlog.000001 /datadir/binlog.000002 /datadir/binlog.000003 /datadir/binlog.index /datadir/galera.cache /datadir/get-pxc-state /datadir/grastate.dat /datadir/ib_buffer_pool /datadir/ibdata1 /datadir/innobackup.backup.full.log /datadir/innobackup.backup.log /datadir/liveness-check.sh /datadir/myApp /datadir/mysql /datadir/mysql-state-monitor /datadir/mysql-state-monitor.log /datadir/mysql.ibd /datadir/mysql.state /datadir/mysqld-error.log /datadir/notify.sock /datadir/peer-list /datadir/performance_schema /datadir/pmm-prerun.sh /datadir/private_key.pem /datadir/public_key.pem /datadir/pxc-configure-pxc.sh /datadir/pxc-entrypoint.sh /datadir/readiness-check.sh /datadir/sys /datadir/undo_001 /datadir/undo_002 /datadir/version_info /datadir/wsrep_cmd_notify_handler.sh ++ mktemp --directory /datadir/pxc_sst_XXXX + tmp=/datadir/pxc_sst_3pnO + socat -u 'openssl-connect:restore-src-on-demand-backup-pvc-some-name:3307,reuseaddr,cert=/etc/mysql/ssl-internal/tls.crt,key=/etc/mysql/ssl-internal/tls.key,cafile=/etc/mysql/ssl-internal/ca.crt,verify=1,commonname='\'''\'',retry=30,no-sni=1' stdio ++ parse_ini mysql-version /datadir/pxc_sst_3pnO/sst_info ++ local key=mysql-version ++ local file_path=/datadir/pxc_sst_3pnO/sst_info ++ awk -F '=[ ]*' '/mysql-version[ ]*=/ {print $2}' /datadir/pxc_sst_3pnO/sst_info + MYSQL_VERSION=8.0.43-34.1 + check_for_version 8.0.43-34.1 8.0.0 + '[' -z 8.0.43-34.1 ']' + '[' -z 8.0.0 ']' + local local_version_str + local required_version_str ++ normalize_version 8.0.43-34.1 ++ local major=0 ++ local minor=0 ++ local patch=0 ++ [[ 8.0.43-34.1 =~ ^([0-9]+)\.([0-9]+)\.?([0-9]*)([^ ])* ]] ++ major=8 ++ minor=0 ++ patch=43 ++ printf %02d%02d%02d 8 0 43 + local_version_str=080043 ++ normalize_version 8.0.0 ++ local major=0 ++ local minor=0 ++ local patch=0 ++ [[ 8.0.0 =~ ^([0-9]+)\.([0-9]+)\.?([0-9]*)([^ ])* ]] ++ major=8 ++ minor=0 ++ patch=0 ++ printf %02d%02d%02d 8 0 0 + required_version_str=080000 + [[ 080043 < 080000 ]] + return 0 + XBSTREAM_EXTRA_ARGS=' --decompress' + socat -u 'openssl-connect:restore-src-on-demand-backup-pvc-some-name:3307,reuseaddr,cert=/etc/mysql/ssl-internal/tls.crt,key=/etc/mysql/ssl-internal/tls.key,cafile=/etc/mysql/ssl-internal/ca.crt,verify=1,commonname='\'''\'',retry=30,no-sni=1' stdio ++ grep -c processor /proc/cpuinfo + xbstream -x -C /datadir/pxc_sst_3pnO --parallel=4 --decompress + set +o xtrace % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 235 100 235 0 0 14687 0 --:--:-- --:--:-- --:--:-- 14687 transition-key exists + xtrabackup --use-memory=100MB --prepare --binlog-info=ON --rollback-prepared-trx --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_3pnO 2025-11-06T12:35:39.420729-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized server arguments: --innodb_checksum_algorithm=crc32 --innodb_log_checksums=1 --innodb_data_file_path=ibdata1:12M:autoextend --innodb_log_file_size=50331648 --innodb_page_size=16384 --innodb_undo_directory=./ --innodb_undo_tablespaces=2 --server-id=29259432 --innodb_log_checksums=ON --innodb_redo_log_encrypt=1 --innodb_undo_log_encrypt=1 2025-11-06T12:35:39.420808-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized client arguments: --use-memory=100MB --prepare=1 --transition-key=* --rollback-prepared-trx=1 --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_3pnO xtrabackup version 8.0.35-34 based on MySQL server 8.0.35 Linux (x86_64) (revision id: c8a25ff9) 2025-11-06T12:35:39.420842-00:00 0 [Note] [MY-011825] [Xtrabackup] cd to /datadir/pxc_sst_3pnO/ 2025-11-06T12:35:39.420903-00:00 0 [Note] [MY-011825] [Xtrabackup] This target seems to be not prepared yet. 2025-11-06T12:35:39.429802-00:00 0 [Note] [MY-011825] [Xtrabackup] xtrabackup_logfile detected: size=8388608, start_lsn=(30287756) 2025-11-06T12:35:39.430541-00:00 0 [Note] [MY-011825] [Xtrabackup] using the following InnoDB configuration for recovery: 2025-11-06T12:35:39.430552-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_home_dir = . 2025-11-06T12:35:39.430556-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_file_path = ibdata1:12M:autoextend 2025-11-06T12:35:39.430580-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_group_home_dir = . 2025-11-06T12:35:39.430611-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_files_in_group = 1 2025-11-06T12:35:39.430622-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_file_size = 8388608 2025-11-06T12:35:39.430706-00:00 0 [Note] [MY-011825] [Xtrabackup] Loading xtrabackup_keys 2025-11-06T12:35:39.436524-00:00 0 [Note] [MY-011825] [Xtrabackup] using the following InnoDB configuration for recovery: 2025-11-06T12:35:39.436538-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_home_dir = . 2025-11-06T12:35:39.436545-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_file_path = ibdata1:12M:autoextend 2025-11-06T12:35:39.436554-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_group_home_dir = . 2025-11-06T12:35:39.436560-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_files_in_group = 1 2025-11-06T12:35:39.436565-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_file_size = 8388608 2025-11-06T12:35:39.436576-00:00 0 [Note] [MY-011825] [Xtrabackup] Starting InnoDB instance for recovery. 2025-11-06T12:35:39.436586-00:00 0 [Note] [MY-011825] [Xtrabackup] Using 104857600 bytes for buffer pool (set by --use-memory parameter) 2025-11-06T12:35:39.436613-00:00 0 [Note] [MY-012932] [InnoDB] PUNCH HOLE support available 2025-11-06T12:35:39.436626-00:00 0 [Note] [MY-012944] [InnoDB] Uses event mutexes 2025-11-06T12:35:39.436632-00:00 0 [Note] [MY-012945] [InnoDB] GCC builtin __atomic_thread_fence() is used for memory barrier 2025-11-06T12:35:39.436638-00:00 0 [Note] [MY-012948] [InnoDB] Compressed tables use zlib 1.2.13 2025-11-06T12:35:39.436785-00:00 0 [Note] [MY-012951] [InnoDB] Using hardware accelerated crc32 and polynomial multiplication. 2025-11-06T12:35:39.437114-00:00 0 [Note] [MY-012203] [InnoDB] Directories to scan './' 2025-11-06T12:35:39.437153-00:00 0 [Note] [MY-012204] [InnoDB] Scanning './' 2025-11-06T12:35:39.439518-00:00 0 [Note] [MY-012208] [InnoDB] Completed space ID check of 8 files. 2025-11-06T12:35:39.440041-00:00 0 [Note] [MY-012955] [InnoDB] Initializing buffer pool, total size = 128.000000M, instances = 1, chunk size =128.000000M 2025-11-06T12:35:39.446573-00:00 0 [Note] [MY-012957] [InnoDB] Completed initialization of buffer pool 2025-11-06T12:35:39.448441-00:00 0 [Note] [MY-011952] [InnoDB] If the mysqld execution user is authorized, page cleaner thread priority can be changed. See the man page of setpriority(). 2025-11-06T12:35:39.475532-00:00 0 [Note] [MY-013883] [InnoDB] The latest found checkpoint is at lsn = 30287756 in redo log file ./#innodb_redo/#ib_redo0. 2025-11-06T12:35:39.475565-00:00 0 [Note] [MY-012560] [InnoDB] The log sequence number 30168989 in the system tablespace does not match the log sequence number 30287756 in the redo log files! 2025-11-06T12:35:39.475572-00:00 0 [Note] [MY-012551] [InnoDB] Database was not shutdown normally! 2025-11-06T12:35:39.475579-00:00 0 [Note] [MY-012552] [InnoDB] Starting crash recovery. 2025-11-06T12:35:39.475701-00:00 0 [Note] [MY-013086] [InnoDB] Starting to parse redo log at lsn = 30287437, whereas checkpoint_lsn = 30287756 and start_lsn = 30287360 2025-11-06T12:35:39.475711-00:00 0 [Note] [MY-012550] [InnoDB] Doing recovery: scanned up to log sequence number 30287766 2025-11-06T12:35:39.491373-00:00 0 [Note] [MY-013083] [InnoDB] Log background threads are being started... 2025-11-06T12:35:39.493316-00:00 0 [Note] [MY-012532] [InnoDB] Applying a batch of 1 redo log records ... 2025-11-06T12:35:39.493362-00:00 0 [Note] [MY-012533] [InnoDB] 100% 2025-11-06T12:35:39.493375-00:00 0 [Note] [MY-012535] [InnoDB] Apply batch completed! 2025-11-06T12:35:39.594661-00:00 0 [Note] [MY-013084] [InnoDB] Log background threads are being closed... 2025-11-06T12:35:39.596115-00:00 0 [Note] [MY-013888] [InnoDB] Upgrading redo log: 1032M, LSN=30287766. 2025-11-06T12:35:39.596125-00:00 0 [Note] [MY-012968] [InnoDB] Starting to delete and rewrite redo log files. 2025-11-06T12:35:39.596159-00:00 0 [Note] [MY-011825] [InnoDB] Removing redo log file: ./#innodb_redo/#ib_redo0 2025-11-06T12:35:39.624197-00:00 0 [Note] [MY-011825] [InnoDB] Creating redo log file at ./#innodb_redo/#ib_redo0_tmp with file_id 0 with size 33554432 bytes 2025-11-06T12:35:39.626513-00:00 0 [Note] [MY-011825] [InnoDB] Renaming redo log file from ./#innodb_redo/#ib_redo0_tmp to ./#innodb_redo/#ib_redo0 2025-11-06T12:35:39.628416-00:00 0 [Note] [MY-012893] [InnoDB] New redo log files created, LSN=30287884 2025-11-06T12:35:39.628475-00:00 0 [Note] [MY-013083] [InnoDB] Log background threads are being started... 2025-11-06T12:35:39.628795-00:00 0 [Note] [MY-013252] [InnoDB] Using undo tablespace './undo_001'. 2025-11-06T12:35:39.629293-00:00 0 [Note] [MY-013795] [InnoDB] Encryption key is loaded for undo tablespace 'innodb_undo_001'. 2025-11-06T12:35:39.629357-00:00 0 [Note] [MY-013252] [InnoDB] Using undo tablespace './undo_002'. 2025-11-06T12:35:39.629716-00:00 0 [Note] [MY-013795] [InnoDB] Encryption key is loaded for undo tablespace 'innodb_undo_002'. 2025-11-06T12:35:39.630281-00:00 0 [Note] [MY-012910] [InnoDB] Opened 2 existing undo tablespaces. 2025-11-06T12:35:39.630329-00:00 0 [Note] [MY-011980] [InnoDB] GTID recovery trx_no: 5200 2025-11-06T12:35:39.677940-00:00 0 [Note] [MY-013776] [InnoDB] Parallel initialization of rseg complete 2025-11-06T12:35:39.677958-00:00 0 [Note] [MY-013777] [InnoDB] Time taken to initialize rseg using 4 thread: 47631 ms. 2025-11-06T12:35:39.678006-00:00 0 [Note] [MY-012923] [InnoDB] Creating shared tablespace for temporary tables 2025-11-06T12:35:39.678065-00:00 0 [Note] [MY-012265] [InnoDB] Setting file './ibtmp1' size to 12 MB. Physically writing the file full; Please wait ... 2025-11-06T12:35:39.706808-00:00 0 [Note] [MY-012266] [InnoDB] File './ibtmp1' size is now 12 MB. 2025-11-06T12:35:39.706935-00:00 0 [Note] [MY-013627] [InnoDB] Scanning temp tablespace dir:'./#innodb_temp/' 2025-11-06T12:35:39.719595-00:00 0 [Note] [MY-013018] [InnoDB] Created 128 and tracked 128 new rollback segment(s) in the temporary tablespace. 128 are now active. 2025-11-06T12:35:39.719725-00:00 0 [Note] [MY-012976] [InnoDB] 8.0.35 started; log sequence number 30287894 2025-11-06T12:35:39.720058-00:00 0 [Warning] [MY-012091] [InnoDB] Allocated tablespace ID 1 for sys/sys_config, old maximum was 0 2025-11-06T12:35:39.721666-00:00 0 [Note] [MY-011825] [Xtrabackup] Completed loading of 6 tablespaces into cache in 0.00191855 seconds 2025-11-06T12:35:39.740505-00:00 0 [Note] [MY-011825] [Xtrabackup] Time taken to build dictionary: 0.0188218 seconds 2025-11-06T12:35:40.740937-00:00 0 [Note] [MY-011825] [Xtrabackup] Recovered WSREP position: 14afd6ae-bb0c-11f0-8efb-0e20134148eb:39 2025-11-06T12:35:40.740986-00:00 0 [Note] [MY-011825] [Xtrabackup] starting shutdown with innodb_fast_shutdown = 1 2025-11-06T12:35:40.741036-00:00 0 [Note] [MY-012330] [InnoDB] FTS optimize thread exiting. 2025-11-06T12:35:41.740914-00:00 0 [Note] [MY-013072] [InnoDB] Starting shutdown... 2025-11-06T12:35:41.841854-00:00 0 [Note] [MY-013084] [InnoDB] Log background threads are being closed... 2025-11-06T12:35:41.851633-00:00 0 [Note] [MY-012980] [InnoDB] Shutdown completed; log sequence number 30287894 2025-11-06T12:35:41.854440-00:00 0 [Note] [MY-011825] [Xtrabackup] completed OK! + xtrabackup --defaults-group=mysqld --datadir=/datadir --move-back --binlog-info=ON --force-non-empty-directories --generate-new-master-key --keyring-vault-config=/etc/mysql/vault-keyring-secret/keyring_vault.conf --early-plugin-load=keyring_vault.so --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_3pnO 2025-11-06T12:35:41.868395-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized server arguments: --datadir=/var/lib/mysql --defaults_group=mysqld --datadir=/datadir 2025-11-06T12:35:41.868497-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized client arguments: --move-back=1 --force-non-empty-directories=1 --transition-key=* --generate-new-master-key=1 --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_3pnO xtrabackup version 8.0.35-34 based on MySQL server 8.0.35 Linux (x86_64) (revision id: c8a25ff9) 2025-11-06T12:35:41.868531-00:00 0 [Note] [MY-011825] [Xtrabackup] cd to /datadir/pxc_sst_3pnO/ 2025-11-06T12:35:41.883997-00:00 0 [Warning] [MY-011197] [InnoDB] Plugin keyring_vault reported: 'Probing secret for being a mount point unsuccessful - skipped.' 2025-11-06T12:35:41.896676-00:00 0 [ERROR] [MY-011197] [InnoDB] Plugin keyring_vault reported: 'Could not decode base64 key's signature' 2025-11-06T12:35:41.896694-00:00 0 [ERROR] [MY-011197] [InnoDB] Plugin keyring_vault reported: 'Could not parse key's signature, skipping the key.' 2025-11-06T12:35:41.896763-00:00 0 [Note] [MY-011825] [Xtrabackup] inititialize_service_handles suceeded 2025-11-06T12:35:41.896780-00:00 0 [Note] [MY-011825] [Xtrabackup] Loading xtrabackup_keys 2025-11-06T12:35:41.901856-00:00 0 [Note] [MY-011825] [Xtrabackup] Loading xtrabackup_keys 2025-11-06T12:35:41.916391-00:00 0 [Note] [MY-011825] [Xtrabackup] Generated new master key 2025-11-06T12:35:41.916561-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving undo_001 to /datadir/undo_001 2025-11-06T12:35:41.916597-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file undo_001 to /datadir/undo_001 2025-11-06T12:35:41.916693-00:00 0 [Note] [MY-011825] [Xtrabackup] Encrypting /datadir/undo_001 tablespace header with new master key. 2025-11-06T12:35:41.916767-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving undo_002 to /datadir/undo_002 2025-11-06T12:35:41.916784-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file undo_002 to /datadir/undo_002 2025-11-06T12:35:41.916804-00:00 0 [Note] [MY-011825] [Xtrabackup] Encrypting /datadir/undo_002 tablespace header with new master key. 2025-11-06T12:35:41.916956-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving ibdata1 to /datadir/ibdata1 2025-11-06T12:35:41.916973-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file ibdata1 to /datadir/ibdata1 2025-11-06T12:35:41.917150-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving binlog.000009 to /datadir//binlog.000009 2025-11-06T12:35:41.917165-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file binlog.000009 to /datadir//binlog.000009 2025-11-06T12:35:41.958606-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving binlog.index to /datadir//binlog.index 2025-11-06T12:35:41.958635-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file binlog.index to /datadir//binlog.index 2025-11-06T12:35:41.958896-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql.ibd to /datadir/mysql.ibd 2025-11-06T12:35:41.958928-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql.ibd to /datadir/mysql.ibd 2025-11-06T12:35:41.959013-00:00 1 [Note] [MY-011825] [Xtrabackup] Encrypting /datadir/mysql.ibd tablespace header with new master key. 2025-11-06T12:35:41.959114-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./ibtmp1 to /datadir/ibtmp1 2025-11-06T12:35:41.959143-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./ibtmp1 to /datadir/ibtmp1 2025-11-06T12:35:41.959245-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_streaming_log.ibd to /datadir/mysql/wsrep_streaming_log.ibd 2025-11-06T12:35:41.959274-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_streaming_log.ibd to /datadir/mysql/wsrep_streaming_log.ibd 2025-11-06T12:35:41.959331-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log.CSV to /datadir/mysql/general_log.CSV 2025-11-06T12:35:41.959359-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log.CSV to /datadir/mysql/general_log.CSV 2025-11-06T12:35:41.959390-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log.CSM to /datadir/mysql/slow_log.CSM 2025-11-06T12:35:41.959419-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log.CSM to /datadir/mysql/slow_log.CSM 2025-11-06T12:35:41.959455-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_cluster.ibd to /datadir/mysql/wsrep_cluster.ibd 2025-11-06T12:35:41.959471-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_cluster.ibd to /datadir/mysql/wsrep_cluster.ibd 2025-11-06T12:35:41.959517-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log_226.sdi to /datadir/mysql/slow_log_226.sdi 2025-11-06T12:35:41.959533-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log_226.sdi to /datadir/mysql/slow_log_226.sdi 2025-11-06T12:35:41.959549-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log_225.sdi to /datadir/mysql/general_log_225.sdi 2025-11-06T12:35:41.959562-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log_225.sdi to /datadir/mysql/general_log_225.sdi 2025-11-06T12:35:41.959577-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_cluster_members.ibd to /datadir/mysql/wsrep_cluster_members.ibd 2025-11-06T12:35:41.959590-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_cluster_members.ibd to /datadir/mysql/wsrep_cluster_members.ibd 2025-11-06T12:35:41.959622-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log.CSM to /datadir/mysql/general_log.CSM 2025-11-06T12:35:41.959636-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log.CSM to /datadir/mysql/general_log.CSM 2025-11-06T12:35:41.959651-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log.CSV to /datadir/mysql/slow_log.CSV 2025-11-06T12:35:41.959663-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log.CSV to /datadir/mysql/slow_log.CSV 2025-11-06T12:35:41.959678-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./sst_info to /datadir/sst_info 2025-11-06T12:35:41.959691-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./sst_info to /datadir/sst_info 2025-11-06T12:35:41.959695-00:00 1 [Note] [MY-011825] [Xtrabackup] Creating directory ./#innodb_redo 2025-11-06T12:35:41.959701-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: creating directory ./#innodb_redo 2025-11-06T12:35:41.959716-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./xtrabackup_keys to /datadir/xtrabackup_keys 2025-11-06T12:35:41.959729-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./xtrabackup_keys to /datadir/xtrabackup_keys 2025-11-06T12:35:41.959780-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./sys/sys_config.ibd to /datadir/sys/sys_config.ibd 2025-11-06T12:35:41.959796-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./sys/sys_config.ibd to /datadir/sys/sys_config.ibd 2025-11-06T12:35:41.959864-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_asyn_181.sdi to /datadir/performance_schema/replication_asyn_181.sdi 2025-11-06T12:35:41.959882-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_asyn_181.sdi to /datadir/performance_schema/replication_asyn_181.sdi 2025-11-06T12:35:41.959902-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/variables_info_195.sdi to /datadir/performance_schema/variables_info_195.sdi 2025-11-06T12:35:41.959917-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/variables_info_195.sdi to /datadir/performance_schema/variables_info_195.sdi 2025-11-06T12:35:41.959934-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/data_locks_169.sdi to /datadir/performance_schema/data_locks_169.sdi 2025-11-06T12:35:41.959949-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/data_locks_169.sdi to /datadir/performance_schema/data_locks_169.sdi 2025-11-06T12:35:41.959965-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/binary_log_trans_198.sdi to /datadir/performance_schema/binary_log_trans_198.sdi 2025-11-06T12:35:41.959978-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/binary_log_trans_198.sdi to /datadir/performance_schema/binary_log_trans_198.sdi 2025-11-06T12:35:41.959994-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_165.sdi to /datadir/performance_schema/memory_summary_b_165.sdi 2025-11-06T12:35:41.960009-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_165.sdi to /datadir/performance_schema/memory_summary_b_165.sdi 2025-11-06T12:35:41.960024-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_139.sdi to /datadir/performance_schema/events_statement_139.sdi 2025-11-06T12:35:41.960037-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_139.sdi to /datadir/performance_schema/events_statement_139.sdi 2025-11-06T12:35:41.960052-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/host_cache_105.sdi to /datadir/performance_schema/host_cache_105.sdi 2025-11-06T12:35:41.960066-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/host_cache_105.sdi to /datadir/performance_schema/host_cache_105.sdi 2025-11-06T12:35:41.960081-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/log_status_183.sdi to /datadir/performance_schema/log_status_183.sdi 2025-11-06T12:35:41.960094-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/log_status_183.sdi to /datadir/performance_schema/log_status_183.sdi 2025-11-06T12:35:41.960109-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_131.sdi to /datadir/performance_schema/events_statement_131.sdi 2025-11-06T12:35:41.960122-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_131.sdi to /datadir/performance_schema/events_statement_131.sdi 2025-11-06T12:35:41.960137-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_148.sdi to /datadir/performance_schema/events_errors_su_148.sdi 2025-11-06T12:35:41.960153-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_148.sdi to /datadir/performance_schema/events_errors_su_148.sdi 2025-11-06T12:35:41.960168-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_conn_173.sdi to /datadir/performance_schema/replication_conn_173.sdi 2025-11-06T12:35:41.960186-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_conn_173.sdi to /datadir/performance_schema/replication_conn_173.sdi 2025-11-06T12:35:41.960202-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_174.sdi to /datadir/performance_schema/replication_appl_174.sdi 2025-11-06T12:35:41.960229-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_174.sdi to /datadir/performance_schema/replication_appl_174.sdi 2025-11-06T12:35:41.960247-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_conn_171.sdi to /datadir/performance_schema/replication_conn_171.sdi 2025-11-06T12:35:41.960261-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_conn_171.sdi to /datadir/performance_schema/replication_conn_171.sdi 2025-11-06T12:35:41.960277-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_144.sdi to /datadir/performance_schema/events_transacti_144.sdi 2025-11-06T12:35:41.960290-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_144.sdi to /datadir/performance_schema/events_transacti_144.sdi 2025-11-06T12:35:41.960305-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/user_defined_fun_197.sdi to /datadir/performance_schema/user_defined_fun_197.sdi 2025-11-06T12:35:41.960319-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/user_defined_fun_197.sdi to /datadir/performance_schema/user_defined_fun_197.sdi 2025-11-06T12:35:41.960335-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_100.sdi to /datadir/performance_schema/events_waits_sum_100.sdi 2025-11-06T12:35:41.960349-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_100.sdi to /datadir/performance_schema/events_waits_sum_100.sdi 2025-11-06T12:35:41.960364-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_instrument_113.sdi to /datadir/performance_schema/setup_instrument_113.sdi 2025-11-06T12:35:41.960378-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_instrument_113.sdi to /datadir/performance_schema/setup_instrument_113.sdi 2025-11-06T12:35:41.960393-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/users_153.sdi to /datadir/performance_schema/users_153.sdi 2025-11-06T12:35:41.960406-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/users_153.sdi to /datadir/performance_schema/users_153.sdi 2025-11-06T12:35:41.960421-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/user_variables_b_185.sdi to /datadir/performance_schema/user_variables_b_185.sdi 2025-11-06T12:35:41.960443-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/user_variables_b_185.sdi to /datadir/performance_schema/user_variables_b_185.sdi 2025-11-06T12:35:41.960459-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/objects_summary__107.sdi to /datadir/performance_schema/objects_summary__107.sdi 2025-11-06T12:35:41.960473-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/objects_summary__107.sdi to /datadir/performance_schema/objects_summary__107.sdi 2025-11-06T12:35:41.960488-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_176.sdi to /datadir/performance_schema/replication_appl_176.sdi 2025-11-06T12:35:41.960502-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_176.sdi to /datadir/performance_schema/replication_appl_176.sdi 2025-11-06T12:35:41.960517-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/accounts_154.sdi to /datadir/performance_schema/accounts_154.sdi 2025-11-06T12:35:41.960531-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/accounts_154.sdi to /datadir/performance_schema/accounts_154.sdi 2025-11-06T12:35:41.960546-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_instances_102.sdi to /datadir/performance_schema/file_instances_102.sdi 2025-11-06T12:35:41.960559-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_instances_102.sdi to /datadir/performance_schema/file_instances_102.sdi 2025-11-06T12:35:41.960574-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_actors_111.sdi to /datadir/performance_schema/setup_actors_111.sdi 2025-11-06T12:35:41.960588-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_actors_111.sdi to /datadir/performance_schema/setup_actors_111.sdi 2025-11-06T12:35:41.960603-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_149.sdi to /datadir/performance_schema/events_errors_su_149.sdi 2025-11-06T12:35:41.960617-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_149.sdi to /datadir/performance_schema/events_errors_su_149.sdi 2025-11-06T12:35:41.960633-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_125.sdi to /datadir/performance_schema/events_stages_su_125.sdi 2025-11-06T12:35:41.960647-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_125.sdi to /datadir/performance_schema/events_stages_su_125.sdi 2025-11-06T12:35:41.960662-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_thread_188.sdi to /datadir/performance_schema/status_by_thread_188.sdi 2025-11-06T12:35:41.960677-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_thread_188.sdi to /datadir/performance_schema/status_by_thread_188.sdi 2025-11-06T12:35:41.960692-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_99.sdi to /datadir/performance_schema/events_waits_sum_99.sdi 2025-11-06T12:35:41.960706-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_99.sdi to /datadir/performance_schema/events_waits_sum_99.sdi 2025-11-06T12:35:41.960720-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/malloc_stats_201.sdi to /datadir/performance_schema/malloc_stats_201.sdi 2025-11-06T12:35:41.960734-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/malloc_stats_201.sdi to /datadir/performance_schema/malloc_stats_201.sdi 2025-11-06T12:35:41.960749-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/variables_by_thr_192.sdi to /datadir/performance_schema/variables_by_thr_192.sdi 2025-11-06T12:35:41.960763-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/variables_by_thr_192.sdi to /datadir/performance_schema/variables_by_thr_192.sdi 2025-11-06T12:35:41.960777-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_cur_93.sdi to /datadir/performance_schema/events_waits_cur_93.sdi 2025-11-06T12:35:41.960791-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_cur_93.sdi to /datadir/performance_schema/events_waits_cur_93.sdi 2025-11-06T12:35:41.960806-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/data_lock_waits_170.sdi to /datadir/performance_schema/data_lock_waits_170.sdi 2025-11-06T12:35:41.960820-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/data_lock_waits_170.sdi to /datadir/performance_schema/data_lock_waits_170.sdi 2025-11-06T12:35:41.960834-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_his_95.sdi to /datadir/performance_schema/events_waits_his_95.sdi 2025-11-06T12:35:41.960848-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_his_95.sdi to /datadir/performance_schema/events_waits_his_95.sdi 2025-11-06T12:35:41.960862-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_user_189.sdi to /datadir/performance_schema/status_by_user_189.sdi 2025-11-06T12:35:41.960877-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_user_189.sdi to /datadir/performance_schema/status_by_user_189.sdi 2025-11-06T12:35:41.960892-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_127.sdi to /datadir/performance_schema/events_stages_su_127.sdi 2025-11-06T12:35:41.960910-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_127.sdi to /datadir/performance_schema/events_stages_su_127.sdi 2025-11-06T12:35:41.960926-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_consumers_112.sdi to /datadir/performance_schema/setup_consumers_112.sdi 2025-11-06T12:35:41.960940-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_consumers_112.sdi to /datadir/performance_schema/setup_consumers_112.sdi 2025-11-06T12:35:41.960955-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/hosts_155.sdi to /datadir/performance_schema/hosts_155.sdi 2025-11-06T12:35:41.960968-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/hosts_155.sdi to /datadir/performance_schema/hosts_155.sdi 2025-11-06T12:35:41.960984-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_cu_120.sdi to /datadir/performance_schema/events_stages_cu_120.sdi 2025-11-06T12:35:41.960999-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_cu_120.sdi to /datadir/performance_schema/events_stages_cu_120.sdi 2025-11-06T12:35:41.961014-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_145.sdi to /datadir/performance_schema/events_transacti_145.sdi 2025-11-06T12:35:41.961029-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_145.sdi to /datadir/performance_schema/events_transacti_145.sdi 2025-11-06T12:35:41.961044-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_accoun_186.sdi to /datadir/performance_schema/status_by_accoun_186.sdi 2025-11-06T12:35:41.961059-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_accoun_186.sdi to /datadir/performance_schema/status_by_accoun_186.sdi 2025-11-06T12:35:41.961074-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_152.sdi to /datadir/performance_schema/events_errors_su_152.sdi 2025-11-06T12:35:41.961089-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_152.sdi to /datadir/performance_schema/events_errors_su_152.sdi 2025-11-06T12:35:41.961104-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_grou_178.sdi to /datadir/performance_schema/replication_grou_178.sdi 2025-11-06T12:35:41.961119-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_grou_178.sdi to /datadir/performance_schema/replication_grou_178.sdi 2025-11-06T12:35:41.961134-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_163.sdi to /datadir/performance_schema/memory_summary_b_163.sdi 2025-11-06T12:35:41.961148-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_163.sdi to /datadir/performance_schema/memory_summary_b_163.sdi 2025-11-06T12:35:41.961163-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_grou_172.sdi to /datadir/performance_schema/replication_grou_172.sdi 2025-11-06T12:35:41.961180-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_grou_172.sdi to /datadir/performance_schema/replication_grou_172.sdi 2025-11-06T12:35:41.961197-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_his_94.sdi to /datadir/performance_schema/events_waits_his_94.sdi 2025-11-06T12:35:41.961238-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_his_94.sdi to /datadir/performance_schema/events_waits_his_94.sdi 2025-11-06T12:35:41.961269-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/mutex_instances_106.sdi to /datadir/performance_schema/mutex_instances_106.sdi 2025-11-06T12:35:41.961297-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/mutex_instances_106.sdi to /datadir/performance_schema/mutex_instances_106.sdi 2025-11-06T12:35:41.961327-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_141.sdi to /datadir/performance_schema/events_transacti_141.sdi 2025-11-06T12:35:41.961352-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_141.sdi to /datadir/performance_schema/events_transacti_141.sdi 2025-11-06T12:35:41.961379-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_147.sdi to /datadir/performance_schema/events_transacti_147.sdi 2025-11-06T12:35:41.961408-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_147.sdi to /datadir/performance_schema/events_transacti_147.sdi 2025-11-06T12:35:41.961438-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_host_187.sdi to /datadir/performance_schema/status_by_host_187.sdi 2025-11-06T12:35:41.961458-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_host_187.sdi to /datadir/performance_schema/status_by_host_187.sdi 2025-11-06T12:35:41.961476-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/global_variables_193.sdi to /datadir/performance_schema/global_variables_193.sdi 2025-11-06T12:35:41.961497-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/global_variables_193.sdi to /datadir/performance_schema/global_variables_193.sdi 2025-11-06T12:35:41.961519-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_connect__159.sdi to /datadir/performance_schema/session_connect__159.sdi 2025-11-06T12:35:41.961535-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_connect__159.sdi to /datadir/performance_schema/session_connect__159.sdi 2025-11-06T12:35:41.961557-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_180.sdi to /datadir/performance_schema/replication_appl_180.sdi 2025-11-06T12:35:41.961574-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_180.sdi to /datadir/performance_schema/replication_appl_180.sdi 2025-11-06T12:35:41.961596-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/metadata_locks_168.sdi to /datadir/performance_schema/metadata_locks_168.sdi 2025-11-06T12:35:41.961611-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/metadata_locks_168.sdi to /datadir/performance_schema/metadata_locks_168.sdi 2025-11-06T12:35:41.961627-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_123.sdi to /datadir/performance_schema/events_stages_su_123.sdi 2025-11-06T12:35:41.961642-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_123.sdi to /datadir/performance_schema/events_stages_su_123.sdi 2025-11-06T12:35:41.961657-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_140.sdi to /datadir/performance_schema/events_transacti_140.sdi 2025-11-06T12:35:41.961672-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_140.sdi to /datadir/performance_schema/events_transacti_140.sdi 2025-11-06T12:35:41.961687-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/cond_instances_91.sdi to /datadir/performance_schema/cond_instances_91.sdi 2025-11-06T12:35:41.961701-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/cond_instances_91.sdi to /datadir/performance_schema/cond_instances_91.sdi 2025-11-06T12:35:41.961716-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_97.sdi to /datadir/performance_schema/events_waits_sum_97.sdi 2025-11-06T12:35:41.961729-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_97.sdi to /datadir/performance_schema/events_waits_sum_97.sdi 2025-11-06T12:35:41.961745-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_136.sdi to /datadir/performance_schema/events_statement_136.sdi 2025-11-06T12:35:41.961761-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_136.sdi to /datadir/performance_schema/events_statement_136.sdi 2025-11-06T12:35:41.961776-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_179.sdi to /datadir/performance_schema/replication_appl_179.sdi 2025-11-06T12:35:41.961791-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_179.sdi to /datadir/performance_schema/replication_appl_179.sdi 2025-11-06T12:35:41.961807-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/performance_time_108.sdi to /datadir/performance_schema/performance_time_108.sdi 2025-11-06T12:35:41.961822-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/performance_time_108.sdi to /datadir/performance_schema/performance_time_108.sdi 2025-11-06T12:35:41.961838-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_166.sdi to /datadir/performance_schema/memory_summary_b_166.sdi 2025-11-06T12:35:41.961852-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_166.sdi to /datadir/performance_schema/memory_summary_b_166.sdi 2025-11-06T12:35:41.961868-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_124.sdi to /datadir/performance_schema/events_stages_su_124.sdi 2025-11-06T12:35:41.961883-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_124.sdi to /datadir/performance_schema/events_stages_su_124.sdi 2025-11-06T12:35:41.961898-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_summary_by__103.sdi to /datadir/performance_schema/file_summary_by__103.sdi 2025-11-06T12:35:41.961913-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_summary_by__103.sdi to /datadir/performance_schema/file_summary_by__103.sdi 2025-11-06T12:35:41.961928-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_129.sdi to /datadir/performance_schema/events_statement_129.sdi 2025-11-06T12:35:41.961943-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_129.sdi to /datadir/performance_schema/events_statement_129.sdi 2025-11-06T12:35:41.961958-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_hi_121.sdi to /datadir/performance_schema/events_stages_hi_121.sdi 2025-11-06T12:35:41.961973-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_hi_121.sdi to /datadir/performance_schema/events_stages_hi_121.sdi 2025-11-06T12:35:41.961988-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_summary_by__104.sdi to /datadir/performance_schema/file_summary_by__104.sdi 2025-11-06T12:35:41.962002-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_summary_by__104.sdi to /datadir/performance_schema/file_summary_by__104.sdi 2025-11-06T12:35:41.962017-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_175.sdi to /datadir/performance_schema/replication_appl_175.sdi 2025-11-06T12:35:41.962032-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_175.sdi to /datadir/performance_schema/replication_appl_175.sdi 2025-11-06T12:35:41.962047-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_hi_122.sdi to /datadir/performance_schema/events_stages_hi_122.sdi 2025-11-06T12:35:41.962062-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_hi_122.sdi to /datadir/performance_schema/events_stages_hi_122.sdi 2025-11-06T12:35:41.962078-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_138.sdi to /datadir/performance_schema/events_statement_138.sdi 2025-11-06T12:35:41.962093-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_138.sdi to /datadir/performance_schema/events_statement_138.sdi 2025-11-06T12:35:41.962108-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_instances_156.sdi to /datadir/performance_schema/socket_instances_156.sdi 2025-11-06T12:35:41.962123-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_instances_156.sdi to /datadir/performance_schema/socket_instances_156.sdi 2025-11-06T12:35:41.962139-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/processlist_109.sdi to /datadir/performance_schema/processlist_109.sdi 2025-11-06T12:35:41.962153-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/processlist_109.sdi to /datadir/performance_schema/processlist_109.sdi 2025-11-06T12:35:41.962168-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_134.sdi to /datadir/performance_schema/events_statement_134.sdi 2025-11-06T12:35:41.962189-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_134.sdi to /datadir/performance_schema/events_statement_134.sdi 2025-11-06T12:35:41.962204-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/prepared_stateme_184.sdi to /datadir/performance_schema/prepared_stateme_184.sdi 2025-11-06T12:35:41.962236-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/prepared_stateme_184.sdi to /datadir/performance_schema/prepared_stateme_184.sdi 2025-11-06T12:35:41.962255-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/threads_119.sdi to /datadir/performance_schema/threads_119.sdi 2025-11-06T12:35:41.962270-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/threads_119.sdi to /datadir/performance_schema/threads_119.sdi 2025-11-06T12:35:41.962286-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_lock_waits_118.sdi to /datadir/performance_schema/table_lock_waits_118.sdi 2025-11-06T12:35:41.962301-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_lock_waits_118.sdi to /datadir/performance_schema/table_lock_waits_118.sdi 2025-11-06T12:35:41.962316-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_summary_b_158.sdi to /datadir/performance_schema/socket_summary_b_158.sdi 2025-11-06T12:35:41.962331-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_summary_b_158.sdi to /datadir/performance_schema/socket_summary_b_158.sdi 2025-11-06T12:35:41.962346-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/persisted_variab_196.sdi to /datadir/performance_schema/persisted_variab_196.sdi 2025-11-06T12:35:41.962361-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/persisted_variab_196.sdi to /datadir/performance_schema/persisted_variab_196.sdi 2025-11-06T12:35:41.962377-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_128.sdi to /datadir/performance_schema/events_statement_128.sdi 2025-11-06T12:35:41.962392-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_128.sdi to /datadir/performance_schema/events_statement_128.sdi 2025-11-06T12:35:41.962407-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_io_waits_s_116.sdi to /datadir/performance_schema/table_io_waits_s_116.sdi 2025-11-06T12:35:41.962423-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_io_waits_s_116.sdi to /datadir/performance_schema/table_io_waits_s_116.sdi 2025-11-06T12:35:41.962441-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_handles_167.sdi to /datadir/performance_schema/table_handles_167.sdi 2025-11-06T12:35:41.962456-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_handles_167.sdi to /datadir/performance_schema/table_handles_167.sdi 2025-11-06T12:35:41.962471-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_threads_115.sdi to /datadir/performance_schema/setup_threads_115.sdi 2025-11-06T12:35:41.962486-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_threads_115.sdi to /datadir/performance_schema/setup_threads_115.sdi 2025-11-06T12:35:41.962501-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_status_191.sdi to /datadir/performance_schema/session_status_191.sdi 2025-11-06T12:35:41.962515-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_status_191.sdi to /datadir/performance_schema/session_status_191.sdi 2025-11-06T12:35:41.962531-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_summary_b_157.sdi to /datadir/performance_schema/socket_summary_b_157.sdi 2025-11-06T12:35:41.962546-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_summary_b_157.sdi to /datadir/performance_schema/socket_summary_b_157.sdi 2025-11-06T12:35:41.962561-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_101.sdi to /datadir/performance_schema/events_waits_sum_101.sdi 2025-11-06T12:35:41.962577-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_101.sdi to /datadir/performance_schema/events_waits_sum_101.sdi 2025-11-06T12:35:41.962592-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_account__160.sdi to /datadir/performance_schema/session_account__160.sdi 2025-11-06T12:35:41.962607-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_account__160.sdi to /datadir/performance_schema/session_account__160.sdi 2025-11-06T12:35:41.962623-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/error_log_92.sdi to /datadir/performance_schema/error_log_92.sdi 2025-11-06T12:35:41.962637-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/error_log_92.sdi to /datadir/performance_schema/error_log_92.sdi 2025-11-06T12:35:41.962652-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/malloc_stats_tot_200.sdi to /datadir/performance_schema/malloc_stats_tot_200.sdi 2025-11-06T12:35:41.962667-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/malloc_stats_tot_200.sdi to /datadir/performance_schema/malloc_stats_tot_200.sdi 2025-11-06T12:35:41.962682-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/tls_channel_stat_199.sdi to /datadir/performance_schema/tls_channel_stat_199.sdi 2025-11-06T12:35:41.962697-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/tls_channel_stat_199.sdi to /datadir/performance_schema/tls_channel_stat_199.sdi 2025-11-06T12:35:41.962712-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/keyring_componen_202.sdi to /datadir/performance_schema/keyring_componen_202.sdi 2025-11-06T12:35:41.962727-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/keyring_componen_202.sdi to /datadir/performance_schema/keyring_componen_202.sdi 2025-11-06T12:35:41.962743-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_177.sdi to /datadir/performance_schema/replication_appl_177.sdi 2025-11-06T12:35:41.962758-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_177.sdi to /datadir/performance_schema/replication_appl_177.sdi 2025-11-06T12:35:41.962773-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_137.sdi to /datadir/performance_schema/events_statement_137.sdi 2025-11-06T12:35:41.962788-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_137.sdi to /datadir/performance_schema/events_statement_137.sdi 2025-11-06T12:35:41.962803-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_143.sdi to /datadir/performance_schema/events_transacti_143.sdi 2025-11-06T12:35:41.962819-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_143.sdi to /datadir/performance_schema/events_transacti_143.sdi 2025-11-06T12:35:41.962834-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_objects_114.sdi to /datadir/performance_schema/setup_objects_114.sdi 2025-11-06T12:35:41.962849-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_objects_114.sdi to /datadir/performance_schema/setup_objects_114.sdi 2025-11-06T12:35:41.962864-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_146.sdi to /datadir/performance_schema/events_transacti_146.sdi 2025-11-06T12:35:41.962879-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_146.sdi to /datadir/performance_schema/events_transacti_146.sdi 2025-11-06T12:35:41.962895-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_126.sdi to /datadir/performance_schema/events_stages_su_126.sdi 2025-11-06T12:35:41.962911-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_126.sdi to /datadir/performance_schema/events_stages_su_126.sdi 2025-11-06T12:35:41.962927-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/pxc_cluster_view_203.sdi to /datadir/performance_schema/pxc_cluster_view_203.sdi 2025-11-06T12:35:41.962942-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/pxc_cluster_view_203.sdi to /datadir/performance_schema/pxc_cluster_view_203.sdi 2025-11-06T12:35:41.962957-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_133.sdi to /datadir/performance_schema/events_statement_133.sdi 2025-11-06T12:35:41.962972-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_133.sdi to /datadir/performance_schema/events_statement_133.sdi 2025-11-06T12:35:41.962988-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_151.sdi to /datadir/performance_schema/events_errors_su_151.sdi 2025-11-06T12:35:41.963003-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_151.sdi to /datadir/performance_schema/events_errors_su_151.sdi 2025-11-06T12:35:41.963020-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_96.sdi to /datadir/performance_schema/events_waits_sum_96.sdi 2025-11-06T12:35:41.963034-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_96.sdi to /datadir/performance_schema/events_waits_sum_96.sdi 2025-11-06T12:35:41.963050-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_132.sdi to /datadir/performance_schema/events_statement_132.sdi 2025-11-06T12:35:41.963065-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_132.sdi to /datadir/performance_schema/events_statement_132.sdi 2025-11-06T12:35:41.963080-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_asyn_182.sdi to /datadir/performance_schema/replication_asyn_182.sdi 2025-11-06T12:35:41.963096-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_asyn_182.sdi to /datadir/performance_schema/replication_asyn_182.sdi 2025-11-06T12:35:41.963111-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_io_waits_s_117.sdi to /datadir/performance_schema/table_io_waits_s_117.sdi 2025-11-06T12:35:41.963126-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_io_waits_s_117.sdi to /datadir/performance_schema/table_io_waits_s_117.sdi 2025-11-06T12:35:41.963141-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_135.sdi to /datadir/performance_schema/events_statement_135.sdi 2025-11-06T12:35:41.963157-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_135.sdi to /datadir/performance_schema/events_statement_135.sdi 2025-11-06T12:35:41.963173-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_142.sdi to /datadir/performance_schema/events_transacti_142.sdi 2025-11-06T12:35:41.963194-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_142.sdi to /datadir/performance_schema/events_transacti_142.sdi 2025-11-06T12:35:41.963231-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/global_status_190.sdi to /datadir/performance_schema/global_status_190.sdi 2025-11-06T12:35:41.963248-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/global_status_190.sdi to /datadir/performance_schema/global_status_190.sdi 2025-11-06T12:35:41.963267-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_98.sdi to /datadir/performance_schema/events_waits_sum_98.sdi 2025-11-06T12:35:41.963282-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_98.sdi to /datadir/performance_schema/events_waits_sum_98.sdi 2025-11-06T12:35:41.963297-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/keyring_keys_161.sdi to /datadir/performance_schema/keyring_keys_161.sdi 2025-11-06T12:35:41.963311-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/keyring_keys_161.sdi to /datadir/performance_schema/keyring_keys_161.sdi 2025-11-06T12:35:41.963326-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_164.sdi to /datadir/performance_schema/memory_summary_b_164.sdi 2025-11-06T12:35:41.963342-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_164.sdi to /datadir/performance_schema/memory_summary_b_164.sdi 2025-11-06T12:35:41.963357-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_variable_194.sdi to /datadir/performance_schema/session_variable_194.sdi 2025-11-06T12:35:41.963373-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_variable_194.sdi to /datadir/performance_schema/session_variable_194.sdi 2025-11-06T12:35:41.963389-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_130.sdi to /datadir/performance_schema/events_statement_130.sdi 2025-11-06T12:35:41.963404-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_130.sdi to /datadir/performance_schema/events_statement_130.sdi 2025-11-06T12:35:41.963421-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_150.sdi to /datadir/performance_schema/events_errors_su_150.sdi 2025-11-06T12:35:41.963440-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_150.sdi to /datadir/performance_schema/events_errors_su_150.sdi 2025-11-06T12:35:41.963456-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/rwlock_instances_110.sdi to /datadir/performance_schema/rwlock_instances_110.sdi 2025-11-06T12:35:41.963472-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/rwlock_instances_110.sdi to /datadir/performance_schema/rwlock_instances_110.sdi 2025-11-06T12:35:41.963488-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_g_162.sdi to /datadir/performance_schema/memory_summary_g_162.sdi 2025-11-06T12:35:41.963503-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_g_162.sdi to /datadir/performance_schema/memory_summary_g_162.sdi 2025-11-06T12:35:41.963516-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./ib_buffer_pool to /datadir/ib_buffer_pool 2025-11-06T12:35:41.963530-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./ib_buffer_pool to /datadir/ib_buffer_pool 2025-11-06T12:35:41.963545-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./xtrabackup_galera_info to /datadir/xtrabackup_galera_info 2025-11-06T12:35:41.963559-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./xtrabackup_galera_info to /datadir/xtrabackup_galera_info 2025-11-06T12:35:41.963610-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./myApp/myApp.ibd to /datadir/myApp/myApp.ibd 2025-11-06T12:35:41.963626-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./myApp/myApp.ibd to /datadir/myApp/myApp.ibd 2025-11-06T12:35:41.963647-00:00 1 [Note] [MY-011825] [Xtrabackup] Encrypting /datadir/myApp/myApp.ibd tablespace header with new master key. 2025-11-06T12:35:41.963700-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./xtrabackup_info to /datadir/xtrabackup_info 2025-11-06T12:35:41.963715-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./xtrabackup_info to /datadir/xtrabackup_info 2025-11-06T12:35:42.059626-00:00 0 [Note] [MY-010733] [Server] Shutting down plugin 'keyring_vault' 2025-11-06T12:35:42.059657-00:00 0 [Note] [MY-010733] [Server] Shutting down plugin 'daemon_keyring_proxy_plugin' 2025-11-06T12:35:42.060059-00:00 0 [Note] [MY-011825] [Xtrabackup] completed OK! + cat /tmp/tmp.cBgDjXpD8S Defaulted container "xtrabackup" out of: xtrabackup, backup-init (init) + rm /tmp/tmp.uDUFLivJLR /tmp/tmp.cBgDjXpD8S + return 0 + wait_for_running some-name-proxysql 1 + local name=some-name-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-proxysql-0 480 + local pod=some-name-proxysql-0 + local max_retry=480 + local ns= ++ egrep '^(pxc|proxysql)$' ++ echo some-name-proxysql-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=proxysql + set +o xtrace pod/some-name-proxysql-0 condition met waiting for pod/some-name-proxysql-0 to become Ready.Ok + wait_for_running some-name-pxc 3 + local name=some-name-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-0 480 + local pod=some-name-pxc-0 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo some-name-pxc-0 ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-0 condition met waiting for pod/some-name-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-1 480 + local pod=some-name-pxc-1 + local max_retry=480 + local ns= ++ echo some-name-pxc-1 ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=pxc + set +o xtrace pod/some-name-pxc-1 condition met waiting for pod/some-name-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-2 480 + local pod=some-name-pxc-2 + local max_retry=480 + local ns= ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo some-name-pxc-2 + local container=pxc + set +o xtrace pod/some-name-pxc-2 condition met waiting for pod/some-name-pxc-2 to become Ready.Ok ++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.secretsName}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Bwky1XTt5s +++ mktemp ++ local LAST_ERR=/tmp/tmp.jSUREByZDu ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.spec.secretsName}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Bwky1XTt5s ++ cat /tmp/tmp.jSUREByZDu ++ rm /tmp/tmp.Bwky1XTt5s /tmp/tmp.jSUREByZDu ++ return 0 + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.EckXMzj2BR +++ mktemp ++ local LAST_ERR=/tmp/tmp.P5syDm1ycK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.EckXMzj2BR ++ cat /tmp/tmp.P5syDm1ycK ++ rm /tmp/tmp.EckXMzj2BR /tmp/tmp.P5syDm1ycK ++ return 0 + local root_pass=root_password + sleep 35 + log 'check data after pxc-restore/on-demand-backup-pvc' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-11-06T12:41:30+0000]' check data after pxc-restore/on-demand-backup-pvc [2025-11-06T12:41:30+0000] check data after pxc-restore/on-demand-backup-pvc + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\''' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/demand-backup-encrypted-with-tls/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/demand-backup-encrypted-with-tls/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\''' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LOeYG1wM3Z +++ mktemp ++ local LAST_ERR=/tmp/tmp.JEZki89ADu ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.LOeYG1wM3Z ++ cat /tmp/tmp.JEZki89ADu ++ rm /tmp/tmp.LOeYG1wM3Z /tmp/tmp.JEZki89ADu ++ return 0 + client_pod=pxc-client-59944c5bbf-sh4xv + wait_pod pxc-client-59944c5bbf-sh4xv + local pod=pxc-client-59944c5bbf-sh4xv + local max_retry=480 + local ns= ++ egrep '^(pxc|proxysql)$' ++ echo pxc-client-59944c5bbf-sh4xv ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-sh4xv condition met waiting for pod/pxc-client-59944c5bbf-sh4xv to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.xEOJuOxVRz/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/demand-backup-encrypted-with-tls/compare/select-1.sql /tmp/tmp.xEOJuOxVRz/select-1.sql + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\''' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/demand-backup-encrypted-with-tls/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/demand-backup-encrypted-with-tls/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\''' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6Z7fqnKxEL +++ mktemp ++ local LAST_ERR=/tmp/tmp.LAI6YP7feq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6Z7fqnKxEL ++ cat /tmp/tmp.LAI6YP7feq ++ rm /tmp/tmp.6Z7fqnKxEL /tmp/tmp.LAI6YP7feq ++ return 0 + client_pod=pxc-client-59944c5bbf-sh4xv + wait_pod pxc-client-59944c5bbf-sh4xv + local pod=pxc-client-59944c5bbf-sh4xv + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-sh4xv ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-sh4xv condition met waiting for pod/pxc-client-59944c5bbf-sh4xv to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.xEOJuOxVRz/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/demand-backup-encrypted-with-tls/compare/select-1.sql /tmp/tmp.xEOJuOxVRz/select-1.sql + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\''' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\''' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/demand-backup-encrypted-with-tls/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/demand-backup-encrypted-with-tls/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\''' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.y65jtb94vg +++ mktemp ++ local LAST_ERR=/tmp/tmp.O3jUhuYCPp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.y65jtb94vg ++ cat /tmp/tmp.O3jUhuYCPp ++ rm /tmp/tmp.y65jtb94vg /tmp/tmp.O3jUhuYCPp ++ return 0 + client_pod=pxc-client-59944c5bbf-sh4xv + wait_pod pxc-client-59944c5bbf-sh4xv + local pod=pxc-client-59944c5bbf-sh4xv + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-sh4xv ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-sh4xv condition met waiting for pod/pxc-client-59944c5bbf-sh4xv to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.xEOJuOxVRz/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/e2e-tests/demand-backup-encrypted-with-tls/compare/select-1.sql /tmp/tmp.xEOJuOxVRz/select-1.sql + '[' on-demand-backup-pvc '!=' on-demand-backup-minio ']' + log 'copy backup' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-11-06T12:41:54+0000]' copy backup [2025-11-06T12:41:54+0000] copy backup + '[' -n '' ']' + bash /mnt/jenkins/workspace/cloud-pxc-operator_PR-2207/deploy/backup/copy-backup.sh on-demand-backup-pvc /tmp/tmp.xEOJuOxVRz/backup Log: /tmp/tmp.wUQmSgDXYh/log pvc/xb-on-demand-backup-pvcpod/backup-access created Starting pod......[done] Downloading started tar: Removing leading `/' from member names Dropping out copy after 0 retries error: unexpected EOF