Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/logs/pitr-8-0.log Warning: version difference between client (1.34) and server (1.31) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.34) and server (1.31) exceeds the supported minor version skew of +/-1 + GTID_PATTERN='[A-F0-9a-f]{8}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{12}:[0-9]+' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + main + create_infra pitr-4490 + local ns=pitr-4490 + '[' -n pxc-operator ']' + kubectl get pxc --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch pxc -n pitr-4818 pitr --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/pitr patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.WQpmXsnqRe ++ mktemp + local LAST_ERR=/tmp/tmp.QWQev2ZfYr + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.WQpmXsnqRe perconaxtradbcluster.pxc.percona.com "pitr" deleted from pitr-4818 namespace + cat /tmp/tmp.QWQev2ZfYr + rm /tmp/tmp.WQpmXsnqRe /tmp/tmp.QWQev2ZfYr + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.EBcPhsHoGT ++ mktemp + local LAST_ERR=/tmp/tmp.HIOhWbVqJv + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.EBcPhsHoGT perconaxtradbclusterbackup.pxc.percona.com "on-pitr-minio" deleted from pitr-4818 namespace + cat /tmp/tmp.HIOhWbVqJv + rm /tmp/tmp.EBcPhsHoGT /tmp/tmp.HIOhWbVqJv + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.wrI0DlsMrd ++ mktemp + local LAST_ERR=/tmp/tmp.Y1XWCYavPI + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.wrI0DlsMrd perconaxtradbclusterrestore.pxc.percona.com "restore-on-pitr-minio-gtid" deleted from pitr-4818 namespace + cat /tmp/tmp.Y1XWCYavPI + rm /tmp/tmp.wrI0DlsMrd /tmp/tmp.Y1XWCYavPI + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep validate-auth + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + egrep -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + awk '{print$1}' + '[' -n '' ']' + desc 'cleaned up old namespaces pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pxc-operator + kubectl_bin get ns ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.xYf1fzxSyV ++ mktemp + local LAST_ERR=/tmp/tmp.iIu7IMKeNA + local exit_status=0 ++ seq 0 2 + local LAST_OUT=/tmp/tmp.kryJMm27az ++ mktemp + for i in '$(seq 0 2)' + xargs kubectl delete ns + set +e + kubectl delete namespace pxc-operator + local LAST_ERR=/tmp/tmp.TSX7t4ltRf + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.kryJMm27az + cat /tmp/tmp.TSX7t4ltRf + rm /tmp/tmp.kryJMm27az /tmp/tmp.TSX7t4ltRf + return 0 namespace "cert-manager" deleted namespace "pitr-4818" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.xYf1fzxSyV namespace "pxc-operator" deleted + cat /tmp/tmp.iIu7IMKeNA + rm /tmp/tmp.xYf1fzxSyV /tmp/tmp.iIu7IMKeNA + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + echo -n 'waiting for namespace/pxc-operator to be deleted' waiting for namespace/pxc-operator to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pxc-operator" not found + desc 'create namespace pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.Wz50Ye7cQc ++ mktemp + local LAST_ERR=/tmp/tmp.lbQRJo8CYj + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Wz50Ye7cQc namespace/pxc-operator created + cat /tmp/tmp.lbQRJo8CYj + rm /tmp/tmp.Wz50Ye7cQc /tmp/tmp.lbQRJo8CYj + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.FtleX3f39x +++ mktemp ++ local LAST_ERR=/tmp/tmp.wyIRwkbPKq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.FtleX3f39x ++ cat /tmp/tmp.wyIRwkbPKq ++ rm /tmp/tmp.FtleX3f39x /tmp/tmp.wyIRwkbPKq ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2234-269f3694-3-cluster7 --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.OFTZD6x21F ++ mktemp + local LAST_ERR=/tmp/tmp.ejg9x84Aps + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2234-269f3694-3-cluster7 --namespace=pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.OFTZD6x21F Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2234-269f3694-3-cluster7" modified. + cat /tmp/tmp.ejg9x84Aps + rm /tmp/tmp.OFTZD6x21F /tmp/tmp.ejg9x84Aps + return 0 + deploy_operator + desc 'start PXC operator' + set +o xtrace ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.9iqzVkf06a ++ mktemp + local LAST_ERR=/tmp/tmp.HfPgk3r58q + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.9iqzVkf06a customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied + cat /tmp/tmp.HfPgk3r58q + rm /tmp/tmp.9iqzVkf06a /tmp/tmp.HfPgk3r58q + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + sed -e 's^namespace: .*^namespace: pxc-operator^' + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/deploy/cw-rbac.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.mwm04q7JnF ++ mktemp + local LAST_ERR=/tmp/tmp.PLw4iZgGGX + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.mwm04q7JnF clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.PLw4iZgGGX + rm /tmp/tmp.mwm04q7JnF /tmp/tmp.PLw4iZgGGX + return 0 + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/deploy/cw-operator.yaml + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-2234-269f3694^' + sed -e 's^failureThreshold: .*^failureThreshold: 10^' + kubectl_bin apply -f - + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "LOG_LEVEL").value) = "VERBOSE"' - + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "DISABLE_TELEMETRY").value) = "true"' - ++ mktemp + local LAST_OUT=/tmp/tmp.jm9YtHLup8 ++ mktemp + local LAST_ERR=/tmp/tmp.Tqqenb5ipc + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.jm9YtHLup8 deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.Tqqenb5ipc + rm /tmp/tmp.jm9YtHLup8 /tmp/tmp.Tqqenb5ipc + return 0 + sleep 10 + kubectl_bin wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s ++ mktemp + local LAST_OUT=/tmp/tmp.DPI9wAHhqW ++ mktemp + local LAST_ERR=/tmp/tmp.jOXlIQY6aQ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.DPI9wAHhqW pod/percona-xtradb-cluster-operator-b5f9c4897-ghlgm condition met + cat /tmp/tmp.jOXlIQY6aQ + rm /tmp/tmp.DPI9wAHhqW /tmp/tmp.jOXlIQY6aQ + return 0 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.uPOZEtLlu1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.NUv4YkMqfz ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.uPOZEtLlu1 ++ cat /tmp/tmp.NUv4YkMqfz ++ rm /tmp/tmp.uPOZEtLlu1 /tmp/tmp.NUv4YkMqfz ++ return 0 + wait_pod percona-xtradb-cluster-operator-b5f9c4897-ghlgm 480 pxc-operator + local pod=percona-xtradb-cluster-operator-b5f9c4897-ghlgm + local max_retry=480 + local ns=pxc-operator ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' ++ echo percona-xtradb-cluster-operator-b5f9c4897-ghlgm + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-b5f9c4897-ghlgm condition met waiting for pod/percona-xtradb-cluster-operator-b5f9c4897-ghlgm to become Ready.Ok + sleep 3 + create_namespace pitr-4490 + local namespace=pitr-4490 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep validate-auth + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + egrep -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + awk '{print$1}' + '[' -n '' ']' + desc 'cleaned up old namespaces pitr-4490' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pitr-4490 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pitr-4490 + xargs kubectl delete ns ++ mktemp + local LAST_OUT=/tmp/tmp.4PI9igaIQt ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.D3kho3bsYS + local exit_status=0 + local LAST_OUT=/tmp/tmp.imMBSZJREo ++ seq 0 2 ++ mktemp + local LAST_ERR=/tmp/tmp.GI0WlK7VxM + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pitr-4490 + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pitr-4490 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.imMBSZJREo + cat /tmp/tmp.GI0WlK7VxM + rm /tmp/tmp.imMBSZJREo /tmp/tmp.GI0WlK7VxM + return 0 error: resource(s) were provided, but no name was specified + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pitr-4490 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.4PI9igaIQt + cat /tmp/tmp.D3kho3bsYS Error from server (NotFound): namespaces "pitr-4490" not found + rm /tmp/tmp.4PI9igaIQt /tmp/tmp.D3kho3bsYS + return 1 + : + wait_for_delete namespace/pitr-4490 + local res=namespace/pitr-4490 + echo -n 'waiting for namespace/pitr-4490 to be deleted' waiting for namespace/pitr-4490 to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pitr-4490" not found + desc 'create namespace pitr-4490' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pitr-4490 ----------------------------------------------------------------------------------- + kubectl_bin create namespace pitr-4490 ++ mktemp + local LAST_OUT=/tmp/tmp.nE6lL8zxUK ++ mktemp + local LAST_ERR=/tmp/tmp.xtNI382bgv + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pitr-4490 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.nE6lL8zxUK namespace/pitr-4490 created + cat /tmp/tmp.xtNI382bgv + rm /tmp/tmp.nE6lL8zxUK /tmp/tmp.xtNI382bgv + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.aQTpgVJL9f +++ mktemp ++ local LAST_ERR=/tmp/tmp.c4HP10keiS ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.aQTpgVJL9f ++ cat /tmp/tmp.c4HP10keiS ++ rm /tmp/tmp.aQTpgVJL9f /tmp/tmp.c4HP10keiS ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2234-269f3694-3-cluster7 --namespace=pitr-4490 ++ mktemp + local LAST_OUT=/tmp/tmp.uvaMJX8fis ++ mktemp + local LAST_ERR=/tmp/tmp.xNYy9z3N8d + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2234-269f3694-3-cluster7 --namespace=pitr-4490 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.uvaMJX8fis Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2234-269f3694-3-cluster7" modified. + cat /tmp/tmp.xNYy9z3N8d + rm /tmp/tmp.uvaMJX8fis /tmp/tmp.xNYy9z3N8d + return 0 + apply_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.4WdNUTzQA2 ++ mktemp + local LAST_ERR=/tmp/tmp.jb0PLLpsNw + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.4WdNUTzQA2 secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.jb0PLLpsNw + rm /tmp/tmp.4WdNUTzQA2 /tmp/tmp.jb0PLLpsNw + return 0 + deploy_cert_manager + desc 'deploy cert manager' + set +o xtrace ----------------------------------------------------------------------------------- deploy cert manager ----------------------------------------------------------------------------------- + kubectl_bin create namespace cert-manager ++ mktemp + local LAST_OUT=/tmp/tmp.ZyEbIpflqZ ++ mktemp + local LAST_ERR=/tmp/tmp.ieG77UdQt0 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace cert-manager + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ZyEbIpflqZ namespace/cert-manager created + cat /tmp/tmp.ieG77UdQt0 + rm /tmp/tmp.ZyEbIpflqZ /tmp/tmp.ieG77UdQt0 + return 0 + kubectl_bin label namespace cert-manager certmanager.k8s.io/disable-validation=true ++ mktemp + local LAST_OUT=/tmp/tmp.C8DQnVqbaz ++ mktemp + local LAST_ERR=/tmp/tmp.sa1i93MLmH + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl label namespace cert-manager certmanager.k8s.io/disable-validation=true + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.C8DQnVqbaz namespace/cert-manager labeled + cat /tmp/tmp.sa1i93MLmH + rm /tmp/tmp.C8DQnVqbaz /tmp/tmp.sa1i93MLmH + return 0 + kubectl_bin apply -f https://github.com/jetstack/cert-manager/releases/download/v1.18.2/cert-manager.yaml --validate=false ++ mktemp + local LAST_OUT=/tmp/tmp.pd5xWLjmdH ++ mktemp + local LAST_ERR=/tmp/tmp.DH8xArvygQ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.18.2/cert-manager.yaml --validate=false + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.pd5xWLjmdH namespace/cert-manager configured customresourcedefinition.apiextensions.k8s.io/certificaterequests.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/certificates.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/challenges.acme.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/clusterissuers.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/issuers.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/orders.acme.cert-manager.io unchanged serviceaccount/cert-manager-cainjector created serviceaccount/cert-manager created serviceaccount/cert-manager-webhook created clusterrole.rbac.authorization.k8s.io/cert-manager-cainjector unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-issuers unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificates unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-orders unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-challenges unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-cluster-view unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-view unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-edit unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-cainjector unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-issuers unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificates unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-orders unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-challenges unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews unchanged role.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection unchanged role.rbac.authorization.k8s.io/cert-manager:leaderelection unchanged role.rbac.authorization.k8s.io/cert-manager-tokenrequest created role.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created rolebinding.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection unchanged rolebinding.rbac.authorization.k8s.io/cert-manager:leaderelection unchanged rolebinding.rbac.authorization.k8s.io/cert-manager-cert-manager-tokenrequest created rolebinding.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created service/cert-manager-cainjector created service/cert-manager created service/cert-manager-webhook created deployment.apps/cert-manager-cainjector created deployment.apps/cert-manager created deployment.apps/cert-manager-webhook created mutatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook configured validatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook configured + cat /tmp/tmp.DH8xArvygQ Warning: resource namespaces/cert-manager is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. + rm /tmp/tmp.pd5xWLjmdH /tmp/tmp.DH8xArvygQ + return 0 + '[' '' == 4.10 ']' + sleep 70 + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/conf/issuer.yml ++ mktemp + local LAST_OUT=/tmp/tmp.kOcot5VbvX ++ mktemp + local LAST_ERR=/tmp/tmp.uYwY33TDZx + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/conf/issuer.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.kOcot5VbvX issuer.cert-manager.io/selfsigning-issuer created + cat /tmp/tmp.uYwY33TDZx + rm /tmp/tmp.kOcot5VbvX /tmp/tmp.uYwY33TDZx + return 0 + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/conf/cert.yml ++ mktemp + local LAST_OUT=/tmp/tmp.oFWSSPtusD ++ mktemp + local LAST_ERR=/tmp/tmp.sIENPafg1o + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/conf/cert.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.oFWSSPtusD certificate.cert-manager.io/tls-minio created + cat /tmp/tmp.sIENPafg1o Warning: spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. + rm /tmp/tmp.oFWSSPtusD /tmp/tmp.sIENPafg1o + return 0 + sleep 25 + start_minio tls-minio + deploy_helm pitr-4490 + helm repo add hashicorp https://helm.releases.hashicorp.com "hashicorp" already exists with the same configuration, skipping + helm repo add minio https://charts.min.io/ "minio" already exists with the same configuration, skipping + helm repo update Hang tight while we grab the latest from your chart repositories... ...Successfully got an update from the "minio" chart repository ...Successfully got an update from the "hashicorp" chart repository Update Complete. ⎈Happy Helming!⎈ + local cert_secret=tls-minio + local endpoint=http://minio-service:9000 + minio_args=(--version $MINIO_VER --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set "users[0].accessKey=some-access-key" --set "users[0].secretKey=some-secret-key" --set "users[0].policy=consoleAdmin" --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G) + local minio_args + [[ -n tls-minio ]] + endpoint=https://minio-service:9000 + minio_args+=(--set tls.enabled=true --set tls.certSecret="$cert_secret" --set tls.publicCrt=tls.crt --set tls.privateKey=tls.key) + desc 'install Minio' + set +o xtrace ----------------------------------------------------------------------------------- install Minio ----------------------------------------------------------------------------------- + helm uninstall minio-service Error: uninstall: Release not loaded: minio-service: release: not found + : + retry 10 60 helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G --set tls.enabled=true --set tls.certSecret=tls-minio --set tls.publicCrt=tls.crt --set tls.privateKey=tls.key minio/minio + local max=10 + local delay=60 + shift 2 + local n=1 + helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G --set tls.enabled=true --set tls.certSecret=tls-minio --set tls.publicCrt=tls.crt --set tls.privateKey=tls.key minio/minio NAME: minio-service LAST DEPLOYED: Wed Nov 12 11:34:58 2025 NAMESPACE: pitr-4490 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: MinIO can be accessed via port 9000 on the following DNS name from within your cluster: minio-service.pitr-4490.cluster.local To access MinIO from localhost, run the below commands: 1. export POD_NAME=$(kubectl get pods --namespace pitr-4490 -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}") 2. kubectl port-forward $POD_NAME 9000 --namespace pitr-4490 Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/ You can now access MinIO server on http://localhost:9000. Follow the below steps to connect to MinIO server with mc client: 1. Download the MinIO mc client - https://min.io/docs/minio/linux/reference/minio-mc.html#quickstart 2. export MC_HOST_minio-service-local=http://$(kubectl get secret --namespace pitr-4490 minio-service -o jsonpath="{.data.rootUser}" | base64 --decode):$(kubectl get secret --namespace pitr-4490 minio-service -o jsonpath="{.data.rootPassword}" | base64 --decode)@localhost:9000 3. mc ls minio-service-local + sleep 30 ++ kubectl_bin get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.96edyr6020 +++ mktemp ++ local LAST_ERR=/tmp/tmp.zItaL4CD1I ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.96edyr6020 ++ cat /tmp/tmp.zItaL4CD1I ++ rm /tmp/tmp.96edyr6020 /tmp/tmp.zItaL4CD1I ++ return 0 + MINIO_POD=minio-service-f887c755-g4f45 + wait_pod minio-service-f887c755-g4f45 + local pod=minio-service-f887c755-g4f45 + local max_retry=480 + local ns= ++ echo minio-service-f887c755-g4f45 ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/minio-service-f887c755-g4f45 condition met waiting for pod/minio-service-f887c755-g4f45 to become Ready.Ok + kubectl_bin run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- /usr/bin/env AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url https://minio-service:9000 --no-verify-ssl s3 mb s3://operator-testing ++ mktemp + local LAST_OUT=/tmp/tmp.dcbW5D4JC0 ++ mktemp + local LAST_ERR=/tmp/tmp.Qiju31Jzpk + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- /usr/bin/env AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url https://minio-service:9000 --no-verify-ssl s3 mb s3://operator-testing + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.dcbW5D4JC0 /usr/lib/python2.7/site-packages/botocore/vendored/requests/packages/urllib3/connectionpool.py:768: InsecureRequestWarning: Unverified HTTPS request is being made. Adding certificate verification is strongly advised. See: https://urllib3.readthedocs.org/en/latest/security.html InsecureRequestWarning) make_bucket: operator-testing pod "aws-cli" deleted from pitr-4490 namespace + cat /tmp/tmp.Qiju31Jzpk All commands and output from this session will be recorded in container logs, including credentials and sensitive information passed through the command prompt. If you don't see a command prompt, try pressing enter. warning: couldn't attach to pod/aws-cli, falling back to streaming logs: Internal error occurred: unable to upgrade connection: container aws-cli not found in pod aws-cli_pitr-4490 + rm /tmp/tmp.dcbW5D4JC0 /tmp/tmp.Qiju31Jzpk + return 0 + cluster=pitr + spinup_pxc pitr /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/conf/pitr.yml + local cluster=pitr + local config=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/conf/pitr.yml + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/secrets.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/client.yml + local port=3306 + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.wZZPuZVAje ++ mktemp + local LAST_ERR=/tmp/tmp.NqnLeXC9bK + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.wZZPuZVAje secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created + cat /tmp/tmp.NqnLeXC9bK + rm /tmp/tmp.wZZPuZVAje /tmp/tmp.NqnLeXC9bK + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/client.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/client.yml + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/client.yml + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' ++ mktemp + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + local LAST_OUT=/tmp/tmp.FJSOi6ZYht + /usr/bin/sed -e s~minio-service.#namespace~minio-service.pitr-4490~ + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2234-269f3694#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' ++ mktemp + local LAST_ERR=/tmp/tmp.QXaXPSws2t + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.FJSOi6ZYht deployment.apps/pxc-client created + cat /tmp/tmp.QXaXPSws2t + rm /tmp/tmp.FJSOi6ZYht /tmp/tmp.QXaXPSws2t + return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/conf/pitr.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/conf/pitr.yml + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/conf/pitr.yml ++ mktemp + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + local LAST_OUT=/tmp/tmp.PHkBYcUF3M + /usr/bin/sed -e s~minio-service.#namespace~minio-service.pitr-4490~ ++ mktemp + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + local LAST_ERR=/tmp/tmp.KlLG9fOs5K + local exit_status=0 ++ seq 0 2 + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2234-269f3694#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.PHkBYcUF3M perconaxtradbcluster.pxc.percona.com/pitr created + cat /tmp/tmp.KlLG9fOs5K + rm /tmp/tmp.PHkBYcUF3M /tmp/tmp.KlLG9fOs5K + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- ++ get_proxy pitr ++ local target_cluster=pitr +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.fEhTk1FrrC ++++ mktemp +++ local LAST_ERR=/tmp/tmp.S0vnmd1Tvy +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.fEhTk1FrrC +++ cat /tmp/tmp.S0vnmd1Tvy +++ rm /tmp/tmp.fEhTk1FrrC /tmp/tmp.S0vnmd1Tvy +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.FGiEpYgfK6 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ngk5mS5QQY +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.FGiEpYgfK6 +++ cat /tmp/tmp.ngk5mS5QQY +++ rm /tmp/tmp.FGiEpYgfK6 /tmp/tmp.ngk5mS5QQY +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo pitr-proxysql ++ return + local proxy=pitr-proxysql + kubectl_bin wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n pitr-4490 ++ mktemp + local LAST_OUT=/tmp/tmp.nvTjsi3gdz ++ mktemp + local LAST_ERR=/tmp/tmp.GKYPsZB4Y0 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n pitr-4490 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n pitr-4490 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n pitr-4490 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.nvTjsi3gdz + cat /tmp/tmp.GKYPsZB4Y0 error: no matching resources found + rm /tmp/tmp.nvTjsi3gdz /tmp/tmp.GKYPsZB4Y0 + return 1 + true + wait_for_running pitr-proxysql 1 + local name=pitr-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod pitr-proxysql-0 480 + local pod=pitr-proxysql-0 + local max_retry=480 + local ns= ++ echo pitr-proxysql-0 ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=proxysql + set +o xtrace Error from server (NotFound): pods "pitr-proxysql-0" not found waiting for pod/pitr-proxysql-0 to become Ready..........Ok + wait_for_running pitr-pxc 3 + local name=pitr-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod pitr-pxc-0 480 + local pod=pitr-pxc-0 + local max_retry=480 + local ns= ++ echo pitr-pxc-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/pitr-pxc-0 condition met waiting for pod/pitr-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod pitr-pxc-1 480 + local pod=pitr-pxc-1 + local max_retry=480 + local ns= ++ echo pitr-pxc-1 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/pitr-pxc-1 condition met waiting for pod/pitr-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod pitr-pxc-2 480 + local pod=pitr-pxc-2 + local max_retry=480 + local ns= ++ echo pitr-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/pitr-pxc-2 condition met waiting for pod/pitr-pxc-2 to become Ready.Ok + sleep 10 ++ kubectl get pxc pitr -o 'jsonpath={.spec.secretsName}' + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.pEOmsI7S6u +++ mktemp ++ local LAST_ERR=/tmp/tmp.gNKKYQPqYp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.pEOmsI7S6u ++ cat /tmp/tmp.gNKKYQPqYp ++ rm /tmp/tmp.pEOmsI7S6u /tmp/tmp.gNKKYQPqYp ++ return 0 + local root_pass=root_password + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' '-h pitr-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' + local 'uri=-h pitr-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.skB5nkla7y +++ mktemp ++ local LAST_ERR=/tmp/tmp.8Hy0BsjwNS ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.skB5nkla7y ++ cat /tmp/tmp.8Hy0BsjwNS ++ rm /tmp/tmp.skB5nkla7y /tmp/tmp.8Hy0BsjwNS ++ return 0 + client_pod=pxc-client-59944c5bbf-c955b + wait_pod pxc-client-59944c5bbf-c955b + local pod=pxc-client-59944c5bbf-c955b + local max_retry=480 + local ns= ++ egrep '^(pxc|proxysql)$' ++ echo pxc-client-59944c5bbf-c955b ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-c955b condition met waiting for pod/pxc-client-59944c5bbf-c955b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h pitr-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h pitr-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hj9Gjq2zH2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.BXlNgceGI5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.hj9Gjq2zH2 ++ cat /tmp/tmp.BXlNgceGI5 ++ rm /tmp/tmp.hj9Gjq2zH2 /tmp/tmp.BXlNgceGI5 ++ return 0 + client_pod=pxc-client-59944c5bbf-c955b + wait_pod pxc-client-59944c5bbf-c955b + local pod=pxc-client-59944c5bbf-c955b + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-c955b ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-c955b condition met waiting for pod/pxc-client-59944c5bbf-c955b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h pitr-pxc-0.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h pitr-pxc-0.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.S1CuQJkzaY +++ mktemp ++ local LAST_ERR=/tmp/tmp.DPzdXWiJEQ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.S1CuQJkzaY ++ cat /tmp/tmp.DPzdXWiJEQ ++ rm /tmp/tmp.S1CuQJkzaY /tmp/tmp.DPzdXWiJEQ ++ return 0 + client_pod=pxc-client-59944c5bbf-c955b + wait_pod pxc-client-59944c5bbf-c955b + local pod=pxc-client-59944c5bbf-c955b + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-c955b ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-c955b condition met waiting for pod/pxc-client-59944c5bbf-c955b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.I8e0A0R49H/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/select-1.sql /tmp/tmp.I8e0A0R49H/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h pitr-pxc-1.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h pitr-pxc-1.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.M7vWUZYlIW +++ mktemp ++ local LAST_ERR=/tmp/tmp.u8j7njRKnv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.M7vWUZYlIW ++ cat /tmp/tmp.u8j7njRKnv ++ rm /tmp/tmp.M7vWUZYlIW /tmp/tmp.u8j7njRKnv ++ return 0 + client_pod=pxc-client-59944c5bbf-c955b + wait_pod pxc-client-59944c5bbf-c955b + local pod=pxc-client-59944c5bbf-c955b + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-c955b ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-c955b condition met waiting for pod/pxc-client-59944c5bbf-c955b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.I8e0A0R49H/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/select-1.sql /tmp/tmp.I8e0A0R49H/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h pitr-pxc-2.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h pitr-pxc-2.pitr-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FajciVRKpD +++ mktemp ++ local LAST_ERR=/tmp/tmp.BA8CYTxCng ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.FajciVRKpD ++ cat /tmp/tmp.BA8CYTxCng ++ rm /tmp/tmp.FajciVRKpD /tmp/tmp.BA8CYTxCng ++ return 0 + client_pod=pxc-client-59944c5bbf-c955b + wait_pod pxc-client-59944c5bbf-c955b + local pod=pxc-client-59944c5bbf-c955b + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-c955b ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-c955b condition met waiting for pod/pxc-client-59944c5bbf-c955b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.I8e0A0R49H/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/select-1.sql /tmp/tmp.I8e0A0R49H/select-1.sql ++ is_keyring_plugin_in_use pitr ++ local cluster=pitr ++ egrep -o 'early-plugin-load=keyring_\w+.so' ++ kubectl_bin exec -it pitr-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wR64KSAFpi +++ mktemp ++ local LAST_ERR=/tmp/tmp.hvXJsn8i5P ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl exec -it pitr-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.wR64KSAFpi ++ cat /tmp/tmp.hvXJsn8i5P Unable to use a TTY - input is not a terminal or the right kind of file ++ rm /tmp/tmp.wR64KSAFpi /tmp/tmp.hvXJsn8i5P ++ return 0 + '[' '' ']' + run_backup pitr on-pitr-minio + local cluster=pitr + local backup=on-pitr-minio + log 'run pxc-backup/on-pitr-minio' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-11-12T11:41:10+0000]' run pxc-backup/on-pitr-minio [2025-11-12T11:41:10+0000] run pxc-backup/on-pitr-minio + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/conf/on-pitr-minio.yml ++ mktemp + local LAST_OUT=/tmp/tmp.QiHuFuFl1Z ++ mktemp + local LAST_ERR=/tmp/tmp.4qvLz3wvgW + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/conf/on-pitr-minio.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.QiHuFuFl1Z perconaxtradbclusterbackup.pxc.percona.com/on-pitr-minio created + cat /tmp/tmp.4qvLz3wvgW + rm /tmp/tmp.QiHuFuFl1Z /tmp/tmp.4qvLz3wvgW + return 0 + wait_backup on-pitr-minio + local backup=on-pitr-minio + local status=Succeeded + set +o xtrace waiting for pxc-backup/on-pitr-minio to reach Succeeded state........................Succeeded + write_test_data pitr + local cluster=pitr + local config= + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/secrets.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/client.yml ++ get_proxy pitr ++ local target_cluster=pitr +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.pAlNbaW61I ++++ mktemp +++ local LAST_ERR=/tmp/tmp.LjpFmvptCc +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.pAlNbaW61I +++ cat /tmp/tmp.LjpFmvptCc +++ rm /tmp/tmp.pAlNbaW61I /tmp/tmp.LjpFmvptCc +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.pRUAWyohXk ++++ mktemp +++ local LAST_ERR=/tmp/tmp.xxqKEDzmKf +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.pRUAWyohXk +++ cat /tmp/tmp.xxqKEDzmKf +++ rm /tmp/tmp.pRUAWyohXk /tmp/tmp.xxqKEDzmKf +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo pitr-proxysql ++ return + local proxy=pitr-proxysql + desc 'write test data' + set +o xtrace ----------------------------------------------------------------------------------- write test data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS test; use test; CREATE TABLE IF NOT EXISTS test (id int PRIMARY KEY) ;' '-h pitr-proxysql -uroot -proot_password' + local 'command=CREATE DATABASE IF NOT EXISTS test; use test; CREATE TABLE IF NOT EXISTS test (id int PRIMARY KEY) ;' + local 'uri=-h pitr-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NMdDK1op7Z +++ mktemp ++ local LAST_ERR=/tmp/tmp.MhZiN3ZhiK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NMdDK1op7Z ++ cat /tmp/tmp.MhZiN3ZhiK ++ rm /tmp/tmp.NMdDK1op7Z /tmp/tmp.MhZiN3ZhiK ++ return 0 + client_pod=pxc-client-59944c5bbf-c955b + wait_pod pxc-client-59944c5bbf-c955b + local pod=pxc-client-59944c5bbf-c955b + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-c955b ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-c955b condition met waiting for pod/pxc-client-59944c5bbf-c955b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT test.test (id) VALUES (100500); INSERT test.test (id) VALUES (100501); INSERT test.test (id) VALUES (100502);' '-h pitr-proxysql -uroot -proot_password' + local 'command=INSERT test.test (id) VALUES (100500); INSERT test.test (id) VALUES (100501); INSERT test.test (id) VALUES (100502);' + local 'uri=-h pitr-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LHkltS4k1M +++ mktemp ++ local LAST_ERR=/tmp/tmp.ek78vNEMS0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.LHkltS4k1M ++ cat /tmp/tmp.ek78vNEMS0 ++ rm /tmp/tmp.LHkltS4k1M /tmp/tmp.ek78vNEMS0 ++ return 0 + client_pod=pxc-client-59944c5bbf-c955b + wait_pod pxc-client-59944c5bbf-c955b + local pod=pxc-client-59944c5bbf-c955b + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-c955b ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-c955b condition met waiting for pod/pxc-client-59944c5bbf-c955b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-3 'SELECT * from test.test;' '-h pitr-pxc-0.pitr-pxc -uroot -proot_password' + local command_id=select-3 + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/select-3.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/select-3-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from test.test;' '-h pitr-pxc-0.pitr-pxc -uroot -proot_password' + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-0.pitr-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.33ms3YV83c +++ mktemp ++ local LAST_ERR=/tmp/tmp.IxhFPM0DKG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.33ms3YV83c ++ cat /tmp/tmp.IxhFPM0DKG ++ rm /tmp/tmp.33ms3YV83c /tmp/tmp.IxhFPM0DKG ++ return 0 + client_pod=pxc-client-59944c5bbf-c955b + wait_pod pxc-client-59944c5bbf-c955b + local pod=pxc-client-59944c5bbf-c955b + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-c955b ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-c955b condition met waiting for pod/pxc-client-59944c5bbf-c955b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.I8e0A0R49H/select-3.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/select-3.sql /tmp/tmp.I8e0A0R49H/select-3.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-3 'SELECT * from test.test;' '-h pitr-pxc-1.pitr-pxc -uroot -proot_password' + local command_id=select-3 + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/select-3.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/select-3-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from test.test;' '-h pitr-pxc-1.pitr-pxc -uroot -proot_password' + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-1.pitr-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8BNFxesGQp +++ mktemp ++ local LAST_ERR=/tmp/tmp.IMfPz6A8sx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8BNFxesGQp ++ cat /tmp/tmp.IMfPz6A8sx ++ rm /tmp/tmp.8BNFxesGQp /tmp/tmp.IMfPz6A8sx ++ return 0 + client_pod=pxc-client-59944c5bbf-c955b + wait_pod pxc-client-59944c5bbf-c955b + local pod=pxc-client-59944c5bbf-c955b + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-c955b ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-c955b condition met waiting for pod/pxc-client-59944c5bbf-c955b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.I8e0A0R49H/select-3.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/select-3.sql /tmp/tmp.I8e0A0R49H/select-3.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-3 'SELECT * from test.test;' '-h pitr-pxc-2.pitr-pxc -uroot -proot_password' + local command_id=select-3 + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/select-3.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/select-3-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from test.test;' '-h pitr-pxc-2.pitr-pxc -uroot -proot_password' + local 'command=SELECT * from test.test;' + local 'uri=-h pitr-pxc-2.pitr-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.26kJx8PHvy +++ mktemp ++ local LAST_ERR=/tmp/tmp.zNKhYV6bGg ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.26kJx8PHvy ++ cat /tmp/tmp.zNKhYV6bGg ++ rm /tmp/tmp.26kJx8PHvy /tmp/tmp.zNKhYV6bGg ++ return 0 + client_pod=pxc-client-59944c5bbf-c955b + wait_pod pxc-client-59944c5bbf-c955b + local pod=pxc-client-59944c5bbf-c955b + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-c955b ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-c955b condition met waiting for pod/pxc-client-59944c5bbf-c955b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.I8e0A0R49H/select-3.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/select-3.sql /tmp/tmp.I8e0A0R49H/select-3.sql ++ is_keyring_plugin_in_use pitr ++ local cluster=pitr ++ kubectl_bin exec -it pitr-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ egrep -o 'early-plugin-load=keyring_\w+.so' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cXuU5HpIzM +++ mktemp ++ local LAST_ERR=/tmp/tmp.MkadA1oRxY ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl exec -it pitr-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cXuU5HpIzM ++ cat /tmp/tmp.MkadA1oRxY Unable to use a TTY - input is not a terminal or the right kind of file ++ rm /tmp/tmp.cXuU5HpIzM /tmp/tmp.MkadA1oRxY ++ return 0 + '[' '' ']' + desc 'changing xtrabackup password multiple times' + set +o xtrace ----------------------------------------------------------------------------------- changing xtrabackup password multiple times ----------------------------------------------------------------------------------- ++ echo -n pass1 ++ base64 + patch_secret my-cluster-secrets xtrabackup cGFzczE= + local secret=my-cluster-secrets + local key=xtrabackup + local value=cGFzczE= + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczE="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.aJEGqgfipD ++ mktemp + local LAST_ERR=/tmp/tmp.SFNfcV3G1M + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczE="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.aJEGqgfipD secret/my-cluster-secrets patched + cat /tmp/tmp.SFNfcV3G1M + rm /tmp/tmp.aJEGqgfipD /tmp/tmp.SFNfcV3G1M + return 0 + wait_cluster_consistency pitr 3 2 + local cluster_name=pitr + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/pitr to be ready' waiting for pxc/pitr to be ready++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8xtBH9b7Sd +++ mktemp ++ local LAST_ERR=/tmp/tmp.5zjBcsoel1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8xtBH9b7Sd ++ cat /tmp/tmp.5zjBcsoel1 ++ rm /tmp/tmp.8xtBH9b7Sd /tmp/tmp.5zjBcsoel1 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TFWqWXPMBC +++ mktemp ++ local LAST_ERR=/tmp/tmp.eF7cpQQYTc ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.TFWqWXPMBC ++ cat /tmp/tmp.eF7cpQQYTc ++ rm /tmp/tmp.TFWqWXPMBC /tmp/tmp.eF7cpQQYTc ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lOO9p8F6J0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.QQXjEmcWya ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.lOO9p8F6J0 ++ cat /tmp/tmp.QQXjEmcWya ++ rm /tmp/tmp.lOO9p8F6J0 /tmp/tmp.QQXjEmcWya ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.37FqPtL1zG +++ mktemp ++ local LAST_ERR=/tmp/tmp.GHzvr1WqR1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.37FqPtL1zG ++ cat /tmp/tmp.GHzvr1WqR1 ++ rm /tmp/tmp.37FqPtL1zG /tmp/tmp.GHzvr1WqR1 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7ts2durAIO +++ mktemp ++ local LAST_ERR=/tmp/tmp.M4W2xhQwMa ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.7ts2durAIO ++ cat /tmp/tmp.M4W2xhQwMa ++ rm /tmp/tmp.7ts2durAIO /tmp/tmp.M4W2xhQwMa ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SanRsukw5u +++ mktemp ++ local LAST_ERR=/tmp/tmp.Kf33sz43tX ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.SanRsukw5u ++ cat /tmp/tmp.Kf33sz43tX ++ rm /tmp/tmp.SanRsukw5u /tmp/tmp.Kf33sz43tX ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oxHmFyfLbe +++ mktemp ++ local LAST_ERR=/tmp/tmp.XrSUp8O9tO ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.oxHmFyfLbe ++ cat /tmp/tmp.XrSUp8O9tO ++ rm /tmp/tmp.oxHmFyfLbe /tmp/tmp.XrSUp8O9tO ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZGgIFSHoN1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.hegBpo85WF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ZGgIFSHoN1 ++ cat /tmp/tmp.hegBpo85WF ++ rm /tmp/tmp.ZGgIFSHoN1 /tmp/tmp.hegBpo85WF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZzUEKSuhO6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.8ID5iva55k ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ZzUEKSuhO6 ++ cat /tmp/tmp.8ID5iva55k ++ rm /tmp/tmp.ZzUEKSuhO6 /tmp/tmp.8ID5iva55k ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gLU1esxwd3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.rw4mngecOn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gLU1esxwd3 ++ cat /tmp/tmp.rw4mngecOn ++ rm /tmp/tmp.gLU1esxwd3 /tmp/tmp.rw4mngecOn ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wj8nbU6cWY +++ mktemp ++ local LAST_ERR=/tmp/tmp.hrWXZK7XEm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.wj8nbU6cWY ++ cat /tmp/tmp.hrWXZK7XEm ++ rm /tmp/tmp.wj8nbU6cWY /tmp/tmp.hrWXZK7XEm ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.obMDxAu4Qb +++ mktemp ++ local LAST_ERR=/tmp/tmp.rNOLsqAujh ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.obMDxAu4Qb ++ cat /tmp/tmp.rNOLsqAujh ++ rm /tmp/tmp.obMDxAu4Qb /tmp/tmp.rNOLsqAujh ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tqZMVxj01B +++ mktemp ++ local LAST_ERR=/tmp/tmp.yczefuy2La ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.tqZMVxj01B ++ cat /tmp/tmp.yczefuy2La ++ rm /tmp/tmp.tqZMVxj01B /tmp/tmp.yczefuy2La ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Yw2hYHsBYq +++ mktemp ++ local LAST_ERR=/tmp/tmp.Z07NwOcZFf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Yw2hYHsBYq ++ cat /tmp/tmp.Z07NwOcZFf ++ rm /tmp/tmp.Yw2hYHsBYq /tmp/tmp.Z07NwOcZFf ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dUtQOAwO3j +++ mktemp ++ local LAST_ERR=/tmp/tmp.GuhnFOhMPr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dUtQOAwO3j ++ cat /tmp/tmp.GuhnFOhMPr ++ rm /tmp/tmp.dUtQOAwO3j /tmp/tmp.GuhnFOhMPr ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XvuvOYqTb3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.b7nhx2BeCl ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XvuvOYqTb3 ++ cat /tmp/tmp.b7nhx2BeCl ++ rm /tmp/tmp.XvuvOYqTb3 /tmp/tmp.b7nhx2BeCl ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NR7uudxuBi +++ mktemp ++ local LAST_ERR=/tmp/tmp.nLQNLJUKYi ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NR7uudxuBi ++ cat /tmp/tmp.nLQNLJUKYi ++ rm /tmp/tmp.NR7uudxuBi /tmp/tmp.nLQNLJUKYi ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Cd6ocPSvIC +++ mktemp ++ local LAST_ERR=/tmp/tmp.E1nukieVlp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Cd6ocPSvIC ++ cat /tmp/tmp.E1nukieVlp ++ rm /tmp/tmp.Cd6ocPSvIC /tmp/tmp.E1nukieVlp ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.USpFP7FcFz +++ mktemp ++ local LAST_ERR=/tmp/tmp.xe6See7EwX ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.USpFP7FcFz ++ cat /tmp/tmp.xe6See7EwX ++ rm /tmp/tmp.USpFP7FcFz /tmp/tmp.xe6See7EwX ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.20lpZlzbpv +++ mktemp ++ local LAST_ERR=/tmp/tmp.yuX0EgfdDk ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.20lpZlzbpv ++ cat /tmp/tmp.yuX0EgfdDk ++ rm /tmp/tmp.20lpZlzbpv /tmp/tmp.yuX0EgfdDk ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KjjBGJf7m1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.nE0q3UPeSl ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KjjBGJf7m1 ++ cat /tmp/tmp.nE0q3UPeSl ++ rm /tmp/tmp.KjjBGJf7m1 /tmp/tmp.nE0q3UPeSl ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.CFbQpBBZ0u +++ mktemp ++ local LAST_ERR=/tmp/tmp.LZsDmZeRMp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.CFbQpBBZ0u ++ cat /tmp/tmp.LZsDmZeRMp ++ rm /tmp/tmp.CFbQpBBZ0u /tmp/tmp.LZsDmZeRMp ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0Fxfhyzw2i +++ mktemp ++ local LAST_ERR=/tmp/tmp.K2cKR7Bvcr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.0Fxfhyzw2i ++ cat /tmp/tmp.K2cKR7Bvcr ++ rm /tmp/tmp.0Fxfhyzw2i /tmp/tmp.K2cKR7Bvcr ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.d9giGkN8RX +++ mktemp ++ local LAST_ERR=/tmp/tmp.TTrVa7v3G8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.d9giGkN8RX ++ cat /tmp/tmp.TTrVa7v3G8 ++ rm /tmp/tmp.d9giGkN8RX /tmp/tmp.TTrVa7v3G8 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LoyNrNLetd +++ mktemp ++ local LAST_ERR=/tmp/tmp.hk7tUkiDkK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.LoyNrNLetd ++ cat /tmp/tmp.hk7tUkiDkK ++ rm /tmp/tmp.LoyNrNLetd /tmp/tmp.hk7tUkiDkK ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 24 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fxakggK5hP +++ mktemp ++ local LAST_ERR=/tmp/tmp.I7Vh3AHRR6 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.fxakggK5hP ++ cat /tmp/tmp.I7Vh3AHRR6 ++ rm /tmp/tmp.fxakggK5hP /tmp/tmp.I7Vh3AHRR6 ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cPuiAbMXoS +++ mktemp ++ local LAST_ERR=/tmp/tmp.WPTO6A6Qzo ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cPuiAbMXoS ++ cat /tmp/tmp.WPTO6A6Qzo ++ rm /tmp/tmp.cPuiAbMXoS /tmp/tmp.WPTO6A6Qzo ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine pitr +++ local cluster_name=pitr ++++ get_proxy pitr ++++ local target_cluster=pitr +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.ItHbWlBvRL ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.gBG7FVwcZI +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.ItHbWlBvRL +++++ cat /tmp/tmp.gBG7FVwcZI +++++ rm /tmp/tmp.ItHbWlBvRL /tmp/tmp.gBG7FVwcZI +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.0JsvWUsAO8 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.UUpg19Odtr +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.0JsvWUsAO8 +++++ cat /tmp/tmp.UUpg19Odtr +++++ rm /tmp/tmp.0JsvWUsAO8 /tmp/tmp.UUpg19Odtr +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo pitr-proxysql ++++ return +++ local cluster_proxy=pitr-proxysql +++ echo proxysql ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.whtdlh9w3W +++ mktemp ++ local LAST_ERR=/tmp/tmp.klGiHgYgJ2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.whtdlh9w3W ++ cat /tmp/tmp.klGiHgYgJ2 ++ rm /tmp/tmp.whtdlh9w3W /tmp/tmp.klGiHgYgJ2 ++ return 0 + [[ 2 == \2 ]] + echo ++ echo -n pass2 ++ base64 + patch_secret my-cluster-secrets xtrabackup cGFzczI= + local secret=my-cluster-secrets + local key=xtrabackup + local value=cGFzczI= + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczI="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.ZNkhZLly80 ++ mktemp + local LAST_ERR=/tmp/tmp.x9Y0plyUUT + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczI="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ZNkhZLly80 secret/my-cluster-secrets patched + cat /tmp/tmp.x9Y0plyUUT + rm /tmp/tmp.ZNkhZLly80 /tmp/tmp.x9Y0plyUUT + return 0 + wait_cluster_consistency pitr 3 2 + local cluster_name=pitr + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/pitr to be ready' waiting for pxc/pitr to be ready++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8XEKCl7gGs +++ mktemp ++ local LAST_ERR=/tmp/tmp.2eAxvuNuVI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8XEKCl7gGs ++ cat /tmp/tmp.2eAxvuNuVI ++ rm /tmp/tmp.8XEKCl7gGs /tmp/tmp.2eAxvuNuVI ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.chYk6dZJtM +++ mktemp ++ local LAST_ERR=/tmp/tmp.8MQ0hJGUo5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.chYk6dZJtM ++ cat /tmp/tmp.8MQ0hJGUo5 ++ rm /tmp/tmp.chYk6dZJtM /tmp/tmp.8MQ0hJGUo5 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.box9m5ZCEZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.RGs3Zke52h ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.box9m5ZCEZ ++ cat /tmp/tmp.RGs3Zke52h ++ rm /tmp/tmp.box9m5ZCEZ /tmp/tmp.RGs3Zke52h ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ls486by98Q +++ mktemp ++ local LAST_ERR=/tmp/tmp.aYnzKs85S9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Ls486by98Q ++ cat /tmp/tmp.aYnzKs85S9 ++ rm /tmp/tmp.Ls486by98Q /tmp/tmp.aYnzKs85S9 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YDz2Oo9hBh +++ mktemp ++ local LAST_ERR=/tmp/tmp.o5U09UapXE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.YDz2Oo9hBh ++ cat /tmp/tmp.o5U09UapXE ++ rm /tmp/tmp.YDz2Oo9hBh /tmp/tmp.o5U09UapXE ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dbnKANDzlJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.yqsaT5YaW7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dbnKANDzlJ ++ cat /tmp/tmp.yqsaT5YaW7 ++ rm /tmp/tmp.dbnKANDzlJ /tmp/tmp.yqsaT5YaW7 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cNRQ2ZzA2J +++ mktemp ++ local LAST_ERR=/tmp/tmp.UNXxJ1ICsB ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cNRQ2ZzA2J ++ cat /tmp/tmp.UNXxJ1ICsB ++ rm /tmp/tmp.cNRQ2ZzA2J /tmp/tmp.UNXxJ1ICsB ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ucHdq78Uy1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.YwU31BShGl ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ucHdq78Uy1 ++ cat /tmp/tmp.YwU31BShGl ++ rm /tmp/tmp.ucHdq78Uy1 /tmp/tmp.YwU31BShGl ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xWI7BlWsU2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.jw919FSYiU ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.xWI7BlWsU2 ++ cat /tmp/tmp.jw919FSYiU ++ rm /tmp/tmp.xWI7BlWsU2 /tmp/tmp.jw919FSYiU ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kyhJYhwxwv +++ mktemp ++ local LAST_ERR=/tmp/tmp.4WsFqXJxJ2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kyhJYhwxwv ++ cat /tmp/tmp.4WsFqXJxJ2 ++ rm /tmp/tmp.kyhJYhwxwv /tmp/tmp.4WsFqXJxJ2 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.S8680EKHLI +++ mktemp ++ local LAST_ERR=/tmp/tmp.M5FfNrBeo3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.S8680EKHLI ++ cat /tmp/tmp.M5FfNrBeo3 ++ rm /tmp/tmp.S8680EKHLI /tmp/tmp.M5FfNrBeo3 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.abgXo51wHT +++ mktemp ++ local LAST_ERR=/tmp/tmp.qdcC3obqBR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.abgXo51wHT ++ cat /tmp/tmp.qdcC3obqBR ++ rm /tmp/tmp.abgXo51wHT /tmp/tmp.qdcC3obqBR ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7qoac9wMQD +++ mktemp ++ local LAST_ERR=/tmp/tmp.A8tfsBNPfI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.7qoac9wMQD ++ cat /tmp/tmp.A8tfsBNPfI ++ rm /tmp/tmp.7qoac9wMQD /tmp/tmp.A8tfsBNPfI ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6f2ahEqFg0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.3tsrBJTXq3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6f2ahEqFg0 ++ cat /tmp/tmp.3tsrBJTXq3 ++ rm /tmp/tmp.6f2ahEqFg0 /tmp/tmp.3tsrBJTXq3 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OKUy0Z7PBp +++ mktemp ++ local LAST_ERR=/tmp/tmp.f7u0DGyR90 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.OKUy0Z7PBp ++ cat /tmp/tmp.f7u0DGyR90 ++ rm /tmp/tmp.OKUy0Z7PBp /tmp/tmp.f7u0DGyR90 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1gRAIC4x5A +++ mktemp ++ local LAST_ERR=/tmp/tmp.1kZLSdfcon ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1gRAIC4x5A ++ cat /tmp/tmp.1kZLSdfcon ++ rm /tmp/tmp.1gRAIC4x5A /tmp/tmp.1kZLSdfcon ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JODcKqcZoe +++ mktemp ++ local LAST_ERR=/tmp/tmp.cd5fkGRTQr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.JODcKqcZoe ++ cat /tmp/tmp.cd5fkGRTQr ++ rm /tmp/tmp.JODcKqcZoe /tmp/tmp.cd5fkGRTQr ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lWobKKuvB0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.yqgX0T27ml ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.lWobKKuvB0 ++ cat /tmp/tmp.yqgX0T27ml ++ rm /tmp/tmp.lWobKKuvB0 /tmp/tmp.yqgX0T27ml ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NNuPdBQEkr +++ mktemp ++ local LAST_ERR=/tmp/tmp.o46GWqv2Wy ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NNuPdBQEkr ++ cat /tmp/tmp.o46GWqv2Wy ++ rm /tmp/tmp.NNuPdBQEkr /tmp/tmp.o46GWqv2Wy ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.CgCDRqhg4b +++ mktemp ++ local LAST_ERR=/tmp/tmp.jwhYaQ9d9u ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.CgCDRqhg4b ++ cat /tmp/tmp.jwhYaQ9d9u ++ rm /tmp/tmp.CgCDRqhg4b /tmp/tmp.jwhYaQ9d9u ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Kg9MWudLoP +++ mktemp ++ local LAST_ERR=/tmp/tmp.kEJmFIAwMc ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Kg9MWudLoP ++ cat /tmp/tmp.kEJmFIAwMc ++ rm /tmp/tmp.Kg9MWudLoP /tmp/tmp.kEJmFIAwMc ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ozmv29Q4Zg +++ mktemp ++ local LAST_ERR=/tmp/tmp.ibryjaFL82 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ozmv29Q4Zg ++ cat /tmp/tmp.ibryjaFL82 ++ rm /tmp/tmp.ozmv29Q4Zg /tmp/tmp.ibryjaFL82 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.X1m13R7Ij6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.mI81vIxHru ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.X1m13R7Ij6 ++ cat /tmp/tmp.mI81vIxHru ++ rm /tmp/tmp.X1m13R7Ij6 /tmp/tmp.mI81vIxHru ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.h600IVKIpk +++ mktemp ++ local LAST_ERR=/tmp/tmp.AzvjkpTkWI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.h600IVKIpk ++ cat /tmp/tmp.AzvjkpTkWI ++ rm /tmp/tmp.h600IVKIpk /tmp/tmp.AzvjkpTkWI ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.g4aMMyALm7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.zvX1V9ziBi ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.g4aMMyALm7 ++ cat /tmp/tmp.zvX1V9ziBi ++ rm /tmp/tmp.g4aMMyALm7 /tmp/tmp.zvX1V9ziBi ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 24 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AVYYGu4deq +++ mktemp ++ local LAST_ERR=/tmp/tmp.hPiHhez71Y ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.AVYYGu4deq ++ cat /tmp/tmp.hPiHhez71Y ++ rm /tmp/tmp.AVYYGu4deq /tmp/tmp.hPiHhez71Y ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FU7O0D0Rz6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.pHfXFQpT59 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.FU7O0D0Rz6 ++ cat /tmp/tmp.pHfXFQpT59 ++ rm /tmp/tmp.FU7O0D0Rz6 /tmp/tmp.pHfXFQpT59 ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine pitr +++ local cluster_name=pitr ++++ get_proxy pitr ++++ local target_cluster=pitr +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.8smXRAvzUY ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.xT875aZ0Kz +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.8smXRAvzUY +++++ cat /tmp/tmp.xT875aZ0Kz +++++ rm /tmp/tmp.8smXRAvzUY /tmp/tmp.xT875aZ0Kz +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.PwPXPsmiKc ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.dsBKefzDd3 +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.PwPXPsmiKc +++++ cat /tmp/tmp.dsBKefzDd3 +++++ rm /tmp/tmp.PwPXPsmiKc /tmp/tmp.dsBKefzDd3 +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo pitr-proxysql ++++ return +++ local cluster_proxy=pitr-proxysql +++ echo proxysql ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Rm4fFCmRiN +++ mktemp ++ local LAST_ERR=/tmp/tmp.Qe8WNnzEhj ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Rm4fFCmRiN ++ cat /tmp/tmp.Qe8WNnzEhj ++ rm /tmp/tmp.Rm4fFCmRiN /tmp/tmp.Qe8WNnzEhj ++ return 0 + [[ 2 == \2 ]] + echo ++ echo -n pass3 ++ base64 + patch_secret my-cluster-secrets xtrabackup cGFzczM= + local secret=my-cluster-secrets + local key=xtrabackup + local value=cGFzczM= + kubectl_bin patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczM="}}' ++ mktemp + local LAST_OUT=/tmp/tmp.90Rl9ulEaz ++ mktemp + local LAST_ERR=/tmp/tmp.dZxbXykdn8 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch secret my-cluster-secrets '-p={"data":{"xtrabackup": "cGFzczM="}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.90Rl9ulEaz secret/my-cluster-secrets patched + cat /tmp/tmp.dZxbXykdn8 + rm /tmp/tmp.90Rl9ulEaz /tmp/tmp.dZxbXykdn8 + return 0 + wait_cluster_consistency pitr 3 2 + local cluster_name=pitr + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/pitr to be ready' waiting for pxc/pitr to be ready++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ayhGjhjTEP +++ mktemp ++ local LAST_ERR=/tmp/tmp.hxBR4AmiL0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ayhGjhjTEP ++ cat /tmp/tmp.hxBR4AmiL0 ++ rm /tmp/tmp.ayhGjhjTEP /tmp/tmp.hxBR4AmiL0 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EcCfb4OKgr +++ mktemp ++ local LAST_ERR=/tmp/tmp.1rkNxWYtxN ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.EcCfb4OKgr ++ cat /tmp/tmp.1rkNxWYtxN ++ rm /tmp/tmp.EcCfb4OKgr /tmp/tmp.1rkNxWYtxN ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2CFqMfx6YG +++ mktemp ++ local LAST_ERR=/tmp/tmp.0Irf73oXvD ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2CFqMfx6YG ++ cat /tmp/tmp.0Irf73oXvD ++ rm /tmp/tmp.2CFqMfx6YG /tmp/tmp.0Irf73oXvD ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PVupZIu8uR +++ mktemp ++ local LAST_ERR=/tmp/tmp.Nv0cJTMIk1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PVupZIu8uR ++ cat /tmp/tmp.Nv0cJTMIk1 ++ rm /tmp/tmp.PVupZIu8uR /tmp/tmp.Nv0cJTMIk1 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kDHgC1oX23 +++ mktemp ++ local LAST_ERR=/tmp/tmp.uPgGZx4NDF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kDHgC1oX23 ++ cat /tmp/tmp.uPgGZx4NDF ++ rm /tmp/tmp.kDHgC1oX23 /tmp/tmp.uPgGZx4NDF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vP5ZNtSuoB +++ mktemp ++ local LAST_ERR=/tmp/tmp.DmIovuGOXy ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.vP5ZNtSuoB ++ cat /tmp/tmp.DmIovuGOXy ++ rm /tmp/tmp.vP5ZNtSuoB /tmp/tmp.DmIovuGOXy ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zzRv5fAKvQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.YYJlFCCKSp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.zzRv5fAKvQ ++ cat /tmp/tmp.YYJlFCCKSp ++ rm /tmp/tmp.zzRv5fAKvQ /tmp/tmp.YYJlFCCKSp ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pkPypGCjz7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.BHV484aSCk ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.pkPypGCjz7 ++ cat /tmp/tmp.BHV484aSCk ++ rm /tmp/tmp.pkPypGCjz7 /tmp/tmp.BHV484aSCk ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.r3etwy58VR +++ mktemp ++ local LAST_ERR=/tmp/tmp.cGTFglzXWr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.r3etwy58VR ++ cat /tmp/tmp.cGTFglzXWr ++ rm /tmp/tmp.r3etwy58VR /tmp/tmp.cGTFglzXWr ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kFXnOpgs2d +++ mktemp ++ local LAST_ERR=/tmp/tmp.q2x7EKLDOv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kFXnOpgs2d ++ cat /tmp/tmp.q2x7EKLDOv ++ rm /tmp/tmp.kFXnOpgs2d /tmp/tmp.q2x7EKLDOv ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3sqKaeG7Jc +++ mktemp ++ local LAST_ERR=/tmp/tmp.NLsVpklXCT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.3sqKaeG7Jc ++ cat /tmp/tmp.NLsVpklXCT ++ rm /tmp/tmp.3sqKaeG7Jc /tmp/tmp.NLsVpklXCT ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bf07rNepeI +++ mktemp ++ local LAST_ERR=/tmp/tmp.WlIyHny31R ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bf07rNepeI ++ cat /tmp/tmp.WlIyHny31R ++ rm /tmp/tmp.bf07rNepeI /tmp/tmp.WlIyHny31R ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KwMzck0UQA +++ mktemp ++ local LAST_ERR=/tmp/tmp.6Li7Y87nNM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KwMzck0UQA ++ cat /tmp/tmp.6Li7Y87nNM ++ rm /tmp/tmp.KwMzck0UQA /tmp/tmp.6Li7Y87nNM ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WPJ4ZlwlvK +++ mktemp ++ local LAST_ERR=/tmp/tmp.fquVMU6Gcc ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.WPJ4ZlwlvK ++ cat /tmp/tmp.fquVMU6Gcc ++ rm /tmp/tmp.WPJ4ZlwlvK /tmp/tmp.fquVMU6Gcc ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.czeQ5UCztZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.tteC8OhPOe ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.czeQ5UCztZ ++ cat /tmp/tmp.tteC8OhPOe ++ rm /tmp/tmp.czeQ5UCztZ /tmp/tmp.tteC8OhPOe ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Xv1u2Rgb7B +++ mktemp ++ local LAST_ERR=/tmp/tmp.a63VHfeu5J ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Xv1u2Rgb7B ++ cat /tmp/tmp.a63VHfeu5J ++ rm /tmp/tmp.Xv1u2Rgb7B /tmp/tmp.a63VHfeu5J ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Nzc4Unalz2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.FrZJ88d8pz ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Nzc4Unalz2 ++ cat /tmp/tmp.FrZJ88d8pz ++ rm /tmp/tmp.Nzc4Unalz2 /tmp/tmp.FrZJ88d8pz ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZjRmK3zZ0U +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZOI2HQ6Su3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ZjRmK3zZ0U ++ cat /tmp/tmp.ZOI2HQ6Su3 ++ rm /tmp/tmp.ZjRmK3zZ0U /tmp/tmp.ZOI2HQ6Su3 ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tjbxFsmwaA +++ mktemp ++ local LAST_ERR=/tmp/tmp.5BfCCt6IjD ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.tjbxFsmwaA ++ cat /tmp/tmp.5BfCCt6IjD ++ rm /tmp/tmp.tjbxFsmwaA /tmp/tmp.5BfCCt6IjD ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.CNwmoJaSDn +++ mktemp ++ local LAST_ERR=/tmp/tmp.CChGUdCxmc ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.CNwmoJaSDn ++ cat /tmp/tmp.CChGUdCxmc ++ rm /tmp/tmp.CNwmoJaSDn /tmp/tmp.CChGUdCxmc ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0HMnFSoc5x +++ mktemp ++ local LAST_ERR=/tmp/tmp.fPh90nPDhd ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.0HMnFSoc5x ++ cat /tmp/tmp.fPh90nPDhd ++ rm /tmp/tmp.0HMnFSoc5x /tmp/tmp.fPh90nPDhd ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wWTQVuQ27M +++ mktemp ++ local LAST_ERR=/tmp/tmp.9fz0EuYKZ5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.wWTQVuQ27M ++ cat /tmp/tmp.9fz0EuYKZ5 ++ rm /tmp/tmp.wWTQVuQ27M /tmp/tmp.9fz0EuYKZ5 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SnQiCmt4cO +++ mktemp ++ local LAST_ERR=/tmp/tmp.SRghv6hH2H ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.SnQiCmt4cO ++ cat /tmp/tmp.SRghv6hH2H ++ rm /tmp/tmp.SnQiCmt4cO /tmp/tmp.SRghv6hH2H ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Vs4Cw50GAS +++ mktemp ++ local LAST_ERR=/tmp/tmp.8TBcezxCBN ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Vs4Cw50GAS ++ cat /tmp/tmp.8TBcezxCBN ++ rm /tmp/tmp.Vs4Cw50GAS /tmp/tmp.8TBcezxCBN ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qmUPtUN8UR +++ mktemp ++ local LAST_ERR=/tmp/tmp.ahzXbmxpeF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qmUPtUN8UR ++ cat /tmp/tmp.ahzXbmxpeF ++ rm /tmp/tmp.qmUPtUN8UR /tmp/tmp.ahzXbmxpeF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 24 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Yc2oRub17y +++ mktemp ++ local LAST_ERR=/tmp/tmp.hYhtZDuZzK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Yc2oRub17y ++ cat /tmp/tmp.hYhtZDuZzK ++ rm /tmp/tmp.Yc2oRub17y /tmp/tmp.hYhtZDuZzK ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BR5oQaPvve +++ mktemp ++ local LAST_ERR=/tmp/tmp.A7YPngEj4a ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.BR5oQaPvve ++ cat /tmp/tmp.A7YPngEj4a ++ rm /tmp/tmp.BR5oQaPvve /tmp/tmp.A7YPngEj4a ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine pitr +++ local cluster_name=pitr ++++ get_proxy pitr ++++ local target_cluster=pitr +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.gRt5thCLJs ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.M7MkbA6K6G +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.gRt5thCLJs +++++ cat /tmp/tmp.M7MkbA6K6G +++++ rm /tmp/tmp.gRt5thCLJs /tmp/tmp.M7MkbA6K6G +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.H3dhZcUWv0 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.Mt0drlHaa1 +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.H3dhZcUWv0 +++++ cat /tmp/tmp.Mt0drlHaa1 +++++ rm /tmp/tmp.H3dhZcUWv0 /tmp/tmp.Mt0drlHaa1 +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo pitr-proxysql ++++ return +++ local cluster_proxy=pitr-proxysql +++ echo proxysql ++ kubectl_bin get pxc pitr -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AInl4sRjuZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.rWbrAxWHfT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc pitr -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.AInl4sRjuZ ++ cat /tmp/tmp.rWbrAxWHfT ++ rm /tmp/tmp.AInl4sRjuZ /tmp/tmp.rWbrAxWHfT ++ return 0 + [[ 2 == \2 ]] + echo + compare_kubectl secret/pitr-mysql-init + local resource=secret/pitr-mysql-init + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/secret_pitr-mysql-init.yml + local new_result=/tmp/tmp.I8e0A0R49H/secret_pitr-mysql-init.yml + desc 'compare secret/pitr-mysql-init-' + set +o xtrace ----------------------------------------------------------------------------------- compare secret/pitr-mysql-init- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/secret_pitr-mysql-init-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/secret_pitr-mysql-init-80.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + version_gt 1.33 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.31 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/secret_pitr-mysql-init-k129.yml ']' + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.31 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/secret_pitr-mysql-init-k127.yml ']' + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.31 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/secret_pitr-mysql-init-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.31 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/secret_pitr-mysql-init-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.31 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/secret_pitr-mysql-init-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/secret_pitr-mysql-init-oc.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/secret_pitr-mysql-init-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/secret_pitr-mysql-init-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/secret_pitr-mysql-init-aks.yml ']' + kubectl_bin get -o yaml secret/pitr-mysql-init ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("pitr-4490", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.HuiyR2ZLW9 ++ mktemp + local LAST_ERR=/tmp/tmp.FUCWLJREyy + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml secret/pitr-mysql-init + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.HuiyR2ZLW9 + cat /tmp/tmp.FUCWLJREyy + rm /tmp/tmp.HuiyR2ZLW9 /tmp/tmp.FUCWLJREyy + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/compare/secret_pitr-mysql-init.yml /tmp/tmp.I8e0A0R49H/secret_pitr-mysql-init.yml + desc 'show binlog events' + set +o xtrace ----------------------------------------------------------------------------------- show binlog events ----------------------------------------------------------------------------------- ++ get_proxy pitr ++ local target_cluster=pitr +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.SfPUuHYGqP ++++ mktemp +++ local LAST_ERR=/tmp/tmp.nZLXuiIEq6 +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.SfPUuHYGqP +++ cat /tmp/tmp.nZLXuiIEq6 +++ rm /tmp/tmp.SfPUuHYGqP /tmp/tmp.nZLXuiIEq6 +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.qFyq1FM71Q ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Mig2RpVtnZ +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.qFyq1FM71Q +++ cat /tmp/tmp.Mig2RpVtnZ +++ rm /tmp/tmp.qFyq1FM71Q /tmp/tmp.Mig2RpVtnZ +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo pitr-proxysql ++ return + proxy=pitr-proxysql + run_mysql 'SHOW BINLOG EVENTS IN '\''binlog.000005'\'';' '-h pitr-proxysql -uroot -proot_password' + local 'command=SHOW BINLOG EVENTS IN '\''binlog.000005'\'';' + local 'uri=-h pitr-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yi9R4ii0yN +++ mktemp ++ local LAST_ERR=/tmp/tmp.eKltC7RvBc ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.yi9R4ii0yN ++ cat /tmp/tmp.eKltC7RvBc ++ rm /tmp/tmp.yi9R4ii0yN /tmp/tmp.eKltC7RvBc ++ return 0 + client_pod=pxc-client-59944c5bbf-c955b + wait_pod pxc-client-59944c5bbf-c955b + local pod=pxc-client-59944c5bbf-c955b + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-c955b ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-c955b condition met waiting for pod/pxc-client-59944c5bbf-c955b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace binlog.000005 4 Format_desc 11128770 126 Server ver: 8.0.43-34.1, Binlog ver: 4 binlog.000005 126 Previous_gtids 11128770 197 da4ab97b-bfbb-11f0-a3dd-8f0b60e73819:1-18 binlog.000005 197 Gtid 11128770 274 SET @@SESSION.GTID_NEXT= 'da4ab97b-bfbb-11f0-a3dd-8f0b60e73819:19' binlog.000005 274 Query 11128770 396 CREATE DATABASE IF NOT EXISTS test /* xid=44 */ binlog.000005 396 Gtid 11128770 473 SET @@SESSION.GTID_NEXT= 'da4ab97b-bfbb-11f0-a3dd-8f0b60e73819:20' binlog.000005 473 Query 11128770 613 use `test`; CREATE TABLE IF NOT EXISTS test (id int PRIMARY KEY) /* xid=45 */ binlog.000005 613 Gtid 11128770 692 SET @@SESSION.GTID_NEXT= 'da4ab97b-bfbb-11f0-a3dd-8f0b60e73819:21' binlog.000005 692 Query 11128770 786 BEGIN binlog.000005 786 Table_map 11128770 836 table_id: 117 (test.test) binlog.000005 836 Write_rows 11128770 876 table_id: 117 flags: STMT_END_F binlog.000005 876 Xid 11128770 907 COMMIT /* xid=46 */ binlog.000005 907 Gtid 11128770 986 SET @@SESSION.GTID_NEXT= 'da4ab97b-bfbb-11f0-a3dd-8f0b60e73819:22' binlog.000005 986 Query 11128770 1080 BEGIN binlog.000005 1080 Table_map 11128770 1130 table_id: 117 (test.test) binlog.000005 1130 Write_rows 11128770 1170 table_id: 117 flags: STMT_END_F binlog.000005 1170 Xid 11128770 1201 COMMIT /* xid=47 */ binlog.000005 1201 Gtid 11128770 1280 SET @@SESSION.GTID_NEXT= 'da4ab97b-bfbb-11f0-a3dd-8f0b60e73819:23' binlog.000005 1280 Query 11128770 1374 BEGIN binlog.000005 1374 Table_map 11128770 1424 table_id: 117 (test.test) binlog.000005 1424 Write_rows 11128770 1464 table_id: 117 flags: STMT_END_F binlog.000005 1464 Xid 11128770 1495 COMMIT /* xid=48 */ binlog.000005 1495 Rotate 11128770 1539 binlog.000006;pos=4 + run_mysql 'SHOW BINLOG EVENTS IN '\''binlog.000006'\'';' '-h pitr-proxysql -uroot -proot_password' + local 'command=SHOW BINLOG EVENTS IN '\''binlog.000006'\'';' + local 'uri=-h pitr-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pdnKvjKaPL +++ mktemp ++ local LAST_ERR=/tmp/tmp.M0hkY4OhZv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.pdnKvjKaPL ++ cat /tmp/tmp.M0hkY4OhZv ++ rm /tmp/tmp.pdnKvjKaPL /tmp/tmp.M0hkY4OhZv ++ return 0 + client_pod=pxc-client-59944c5bbf-c955b + wait_pod pxc-client-59944c5bbf-c955b + local pod=pxc-client-59944c5bbf-c955b + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-c955b ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-c955b condition met waiting for pod/pxc-client-59944c5bbf-c955b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace binlog.000006 4 Format_desc 11128770 126 Server ver: 8.0.43-34.1, Binlog ver: 4 binlog.000006 126 Previous_gtids 11128770 197 da4ab97b-bfbb-11f0-a3dd-8f0b60e73819:1-23 binlog.000006 197 Gtid 11128771 276 SET @@SESSION.GTID_NEXT= 'da4ab97b-bfbb-11f0-a3dd-8f0b60e73819:24' binlog.000006 276 Query 11128771 516 use `mysql`; ALTER USER 'xtrabackup'@'%' IDENTIFIED WITH 'mysql_native_password' AS '*22A99BA288DB55E8E230679259740873101CD636' RETAIN CURRENT PASSWORD /* xid=49 */ binlog.000006 516 Gtid 11128772 593 SET @@SESSION.GTID_NEXT= 'da4ab97b-bfbb-11f0-a3dd-8f0b60e73819:25' binlog.000006 593 Query 11128772 731 use `mysql`; ALTER USER 'xtrabackup'@'%' DISCARD OLD PASSWORD /* xid=50 */ binlog.000006 731 Gtid 11128770 810 SET @@SESSION.GTID_NEXT= 'da4ab97b-bfbb-11f0-a3dd-8f0b60e73819:26' binlog.000006 810 Query 11128770 991 use `mysql`; CREATE FUNCTION IF NOT EXISTS get_last_record_timestamp_by_binlog RETURNS INTEGER SONAME 'binlog_utils_udf.so' binlog.000006 991 Gtid 11128770 1068 SET @@SESSION.GTID_NEXT= 'da4ab97b-bfbb-11f0-a3dd-8f0b60e73819:27' binlog.000006 1068 Query 11128770 1235 use `mysql`; CREATE FUNCTION IF NOT EXISTS get_gtid_set_by_binlog RETURNS STRING SONAME 'binlog_utils_udf.so' binlog.000006 1235 Gtid 11128770 1314 SET @@SESSION.GTID_NEXT= 'da4ab97b-bfbb-11f0-a3dd-8f0b60e73819:28' binlog.000006 1314 Query 11128770 1496 use `mysql`; CREATE FUNCTION IF NOT EXISTS get_first_record_timestamp_by_binlog RETURNS INTEGER SONAME 'binlog_utils_udf.so' binlog.000006 1496 Rotate 11128770 1540 binlog.000007;pos=4 ++ run_mysql 'SELECT now();' '-h pitr-proxysql -uroot -proot_password' ++ local 'command=SELECT now();' ++ local 'uri=-h pitr-proxysql -uroot -proot_password' +++ get_client_pod +++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.cfZgFdb5VL ++++ mktemp +++ local LAST_ERR=/tmp/tmp.BAiUOk5s9b +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.cfZgFdb5VL +++ cat /tmp/tmp.BAiUOk5s9b +++ rm /tmp/tmp.cfZgFdb5VL /tmp/tmp.BAiUOk5s9b +++ return 0 ++ client_pod=pxc-client-59944c5bbf-c955b ++ wait_pod pxc-client-59944c5bbf-c955b ++ local pod=pxc-client-59944c5bbf-c955b ++ local max_retry=480 ++ local ns= +++ echo pxc-client-59944c5bbf-c955b +++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' +++ egrep '^(pxc|proxysql)$' ++ local container= ++ set +o xtrace pod/pxc-client-59944c5bbf-c955b condition met waiting for pod/pxc-client-59944c5bbf-c955b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ++ set +o xtrace + time_now='2025-11-12 11:51:11' ++ run_mysql 'SELECT @@gtid_executed;' '-h pitr-proxysql -uroot -proot_password' ++ local 'command=SELECT @@gtid_executed;' ++ /usr/bin/sed 's/\([a-f0-9-]\{36\}\):[0-9]*-\([0-9]*\).*/\1:\2/' ++ local 'uri=-h pitr-proxysql -uroot -proot_password' +++ get_client_pod +++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.vrFM4LHl3Y ++++ mktemp +++ local LAST_ERR=/tmp/tmp.PMOETVHlBt +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.vrFM4LHl3Y +++ cat /tmp/tmp.PMOETVHlBt +++ rm /tmp/tmp.vrFM4LHl3Y /tmp/tmp.PMOETVHlBt +++ return 0 ++ client_pod=pxc-client-59944c5bbf-c955b ++ wait_pod pxc-client-59944c5bbf-c955b ++ local pod=pxc-client-59944c5bbf-c955b ++ local max_retry=480 ++ local ns= +++ echo pxc-client-59944c5bbf-c955b +++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' +++ egrep '^(pxc|proxysql)$' ++ local container= ++ set +o xtrace pod/pxc-client-59944c5bbf-c955b condition met waiting for pod/pxc-client-59944c5bbf-c955b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ++ set +o xtrace + gtid=da4ab97b-bfbb-11f0-a3dd-8f0b60e73819:38 + [[ ! da4ab97b-bfbb-11f0-a3dd-8f0b60e73819:38 =~ [A-F0-9a-f]{8}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{4}-[A-F0-9a-f]{12}:[0-9]+ ]] + write_data_for_pitr pitr + local cluster=pitr ++ get_proxy pitr ++ local target_cluster=pitr +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.OnAMGqPQ9J ++++ mktemp +++ local LAST_ERR=/tmp/tmp.XUWgVjxLmT +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.OnAMGqPQ9J +++ cat /tmp/tmp.XUWgVjxLmT +++ rm /tmp/tmp.OnAMGqPQ9J /tmp/tmp.XUWgVjxLmT +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.LY689qYgBp ++++ mktemp +++ local LAST_ERR=/tmp/tmp.8e3aYbRQ61 +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc pitr -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.LY689qYgBp +++ cat /tmp/tmp.8e3aYbRQ61 +++ rm /tmp/tmp.LY689qYgBp /tmp/tmp.8e3aYbRQ61 +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo pitr-proxysql ++ return + local proxy=pitr-proxysql + desc 'write data for pitr' + set +o xtrace ----------------------------------------------------------------------------------- write data for pitr ----------------------------------------------------------------------------------- + run_mysql 'INSERT test.test (id) VALUES (100503); INSERT test.test (id) VALUES (100504); INSERT test.test (id) VALUES (100505);' '-h pitr-proxysql -uroot -proot_password' + local 'command=INSERT test.test (id) VALUES (100503); INSERT test.test (id) VALUES (100504); INSERT test.test (id) VALUES (100505);' + local 'uri=-h pitr-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2ux22yaGUH +++ mktemp ++ local LAST_ERR=/tmp/tmp.xReWHYzhQ2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2ux22yaGUH ++ cat /tmp/tmp.xReWHYzhQ2 ++ rm /tmp/tmp.2ux22yaGUH /tmp/tmp.xReWHYzhQ2 ++ return 0 + client_pod=pxc-client-59944c5bbf-c955b + wait_pod pxc-client-59944c5bbf-c955b + local pod=pxc-client-59944c5bbf-c955b + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-c955b ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-c955b condition met waiting for pod/pxc-client-59944c5bbf-c955b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 120 + check_latest_restorable_time on-pitr-minio + local backup=on-pitr-minio + desc 'check latest restorable time' + set +o xtrace ----------------------------------------------------------------------------------- check latest restorable time ----------------------------------------------------------------------------------- ++ kubectl_bin get pxc-backup on-pitr-minio -o 'jsonpath={.status.latestRestorableTime}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.azPnJZHSLo +++ mktemp ++ local LAST_ERR=/tmp/tmp.E1JfLP9Spy ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc-backup on-pitr-minio -o 'jsonpath={.status.latestRestorableTime}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.azPnJZHSLo ++ cat /tmp/tmp.E1JfLP9Spy ++ rm /tmp/tmp.azPnJZHSLo /tmp/tmp.E1JfLP9Spy ++ return 0 + latest_restorable_time=2025-11-12T11:51:59Z + [[ 2025-11-12T11:51:59Z =~ ^[0-9]{4}-[0-9]{2}-[0-9]{2}T ]] + echo 'Latest restorable time is 2025-11-12T11:51:59Z' Latest restorable time is 2025-11-12T11:51:59Z + timeout=60 + binlogs_exist=0 ++ seq 1 5 + for i in '$(seq 1 5)' + echo 'Checking if binlogs exist in bucket (attempt 1)...' Checking if binlogs exist in bucket (attempt 1)... ++ kubectl_bin run -n '' -i --rm aws-cli --image=perconalab/awscli --restart=Never -- /usr/bin/env AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url https://minio-service:9000 --no-verify-ssl s3 ls operator-testing/binlogs/ ++ grep -c binlog ++ cat +++ mktemp ++ local LAST_OUT=/tmp/tmp.tWpB2PrBmS +++ mktemp ++ local LAST_ERR=/tmp/tmp.o3Ostjncs9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl run -n '' -i --rm aws-cli --image=perconalab/awscli --restart=Never -- /usr/bin/env AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url https://minio-service:9000 --no-verify-ssl s3 ls operator-testing/binlogs/ ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.tWpB2PrBmS ++ cat /tmp/tmp.o3Ostjncs9 ++ rm /tmp/tmp.tWpB2PrBmS /tmp/tmp.o3Ostjncs9 ++ return 0 ++ exit 0 + binlogs_exist=22 + '[' 22 -gt 0 ']' + echo '22 binlogs found in bucket' 22 binlogs found in bucket + break + '[' 22 -eq 0 ']' + run_recovery_check_pitr pitr restore-on-pitr-minio-gtid on-pitr-minio select-2 '' '' da4ab97b-bfbb-11f0-a3dd-8f0b60e73819:38 + local cluster=pitr + local restore=restore-on-pitr-minio-gtid + local backup=on-pitr-minio + local compare=select-2 + local time_now= + local dest= + local gtid=da4ab97b-bfbb-11f0-a3dd-8f0b60e73819:38 + desc 'recover backup' restore-on-pitr-minio-gtid + set +o xtrace ----------------------------------------------------------------------------------- recover backup restore-on-pitr-minio-gtid ----------------------------------------------------------------------------------- + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/pitr/conf/restore-on-pitr-minio-gtid.yaml + /usr/bin/sed -e 's///g' + /usr/bin/sed -e 's///g' + kubectl_bin apply -f - ++ mktemp + /usr/bin/sed -e 's//da4ab97b-bfbb-11f0-a3dd-8f0b60e73819:38/g' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.pitr-4490~ + local LAST_OUT=/tmp/tmp.CDNz8mv3GP ++ mktemp + local LAST_ERR=/tmp/tmp.B6BxyY8Kva + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.CDNz8mv3GP perconaxtradbclusterrestore.pxc.percona.com/restore-on-pitr-minio-gtid created + cat /tmp/tmp.B6BxyY8Kva + rm /tmp/tmp.CDNz8mv3GP /tmp/tmp.B6BxyY8Kva + return 0 + wait_backup_restore restore-on-pitr-minio-gtid 'Stopping Cluster' + local backup_name=restore-on-pitr-minio-gtid + local 'target_state=Stopping Cluster' + local wait_time=720 + set +o xtrace waiting for pxc-restore/restore-on-pitr-minio-gtid to reach Stopping Cluster state 2025-11-12T11:53:32 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:53:34 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:53:36 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:53:38 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:53:40 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:53:42 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:53:44 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:53:46 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:53:48 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:53:49 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:53:51 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:53:53 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:53:55 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:53:57 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:53:59 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:01 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:03 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:05 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:07 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:09 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:11 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:13 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:15 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:17 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:19 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:21 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:23 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:25 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:27 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:29 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:31 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:33 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:35 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:37 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:40 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:42 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:44 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:46 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:48 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:50 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:52 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:54 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:56 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:54:58 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:00 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:02 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:04 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:06 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:08 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:10 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:12 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:13 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:16 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:17 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:19 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:21 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:23 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:25 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:27 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:29 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:31 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:33 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:35 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:37 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:39 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:41 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:43 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:45 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:47 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:49 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:51 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:53 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:55 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:57 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:55:59 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:01 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:03 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:05 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:06 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:08 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:10 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:12 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:14 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:16 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:18 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:20 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:22 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:24 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:26 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:28 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:30 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:32 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:34 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:36 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:38 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:40 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:41 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:43 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:45 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:47 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:49 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:51 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:53 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:55 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:57 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:56:59 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:01 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:03 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:05 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:07 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:09 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:11 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:13 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:15 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:16 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:18 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:20 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:22 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:24 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:26 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:28 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:30 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:32 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:34 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:36 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:38 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:40 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:42 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:44 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:46 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:48 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:50 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:52 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:53 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:55 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:58 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:57:59 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:01 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:03 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:05 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:07 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:09 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:11 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:13 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:15 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:17 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:19 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:20 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:22 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:24 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:26 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:28 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:30 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:32 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:34 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:36 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:38 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:40 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:42 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:44 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:46 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:47 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:49 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:51 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:53 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:54 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:56 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:58:58 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:00 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:02 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:04 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:06 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:08 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:10 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:12 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:14 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:15 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:17 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:19 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:21 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:23 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:25 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:27 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:29 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:31 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:33 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:35 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:36 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:38 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:40 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:42 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:44 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:46 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:48 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:50 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:52 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:54 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:56 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T11:59:58 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:00 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:02 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:04 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:06 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:08 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:10 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:12 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:14 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:15 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:17 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:19 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:21 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:23 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:25 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:27 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:29 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:30 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:32 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:34 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:36 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:37 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:39 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:41 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:43 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:45 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:47 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:49 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:51 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:52 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:54 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:56 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:00:58 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:00 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:02 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:04 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:06 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:07 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:09 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:11 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:13 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:15 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:17 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:19 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:21 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:23 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:25 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:27 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:29 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:31 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:33 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:35 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:37 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:39 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:41 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:43 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:45 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:47 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:49 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:50 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:52 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:54 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:56 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:01:58 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:00 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:02 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:04 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:06 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:08 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:10 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:12 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:14 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:16 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:18 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:20 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:22 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:24 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:26 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:28 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:30 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:32 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:34 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:36 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:38 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:40 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:42 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:44 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:46 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:47 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:49 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:51 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:53 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:55 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:57 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:02:59 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:01 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:03 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:05 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:07 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:09 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:11 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:13 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:15 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:17 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:19 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:21 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:23 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:25 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:27 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:29 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:31 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:33 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:35 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:36 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:38 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:40 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:42 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:44 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:46 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:48 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:50 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:52 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:54 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:56 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:57 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:03:59 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:01 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:03 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:05 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:07 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:09 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:11 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:13 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:15 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:17 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:19 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:21 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:23 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:25 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:27 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:29 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:30 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:32 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:34 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:36 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:38 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:40 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:42 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:44 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:46 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:48 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:50 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:52 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:54 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:56 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:04:58 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:00 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:02 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:03 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:05 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:07 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:09 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:11 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:13 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:15 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:17 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:19 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:21 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:23 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:25 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:27 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:29 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:31 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:33 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:35 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:37 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:39 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:41 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:42 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:44 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:46 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:48 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:50 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:52 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:54 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:56 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:05:58 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:00 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:02 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:04 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:06 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:08 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:10 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:12 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:13 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:15 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:17 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:19 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:20 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:23 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:24 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:26 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:28 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:30 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:32 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:34 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:36 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:38 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:40 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:42 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:44 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:46 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:48 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:49 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:51 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:53 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:55 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:56 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:06:58 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:00 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:02 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:04 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:06 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:08 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:10 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:12 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:13 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:15 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:16 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:18 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:20 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:22 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:24 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:26 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:28 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:29 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:31 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:32 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:34 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:36 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:38 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:39 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:41 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:43 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:45 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:47 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:49 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:51 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:52 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:54 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:56 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:07:58 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:00 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:02 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:04 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:06 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:08 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:09 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:11 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:12 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:14 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:16 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:18 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:20 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:22 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:24 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:26 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:28 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:29 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:31 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:33 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:35 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:37 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:39 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:41 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:42 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:44 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:45 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:47 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:49 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:51 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:53 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:55 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:57 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:08:58 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:00 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:02 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:04 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:05 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:07 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:09 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:11 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:13 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:15 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:17 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:18 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:20 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:21 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:23 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:25 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:27 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:29 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:30 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:32 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:34 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:36 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:37 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:39 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:41 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:43 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:45 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:46 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:48 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:50 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:52 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:54 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:56 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:57 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:09:59 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:01 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:03 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:05 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:07 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:09 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:10 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:12 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:13 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:15 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:17 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:19 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:21 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:23 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:25 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:27 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:29 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:30 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:32 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:34 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:36 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:37 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:39 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:41 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:42 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:44 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:46 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:48 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:49 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:51 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:53 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:55 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:57 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:10:58 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:00 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:02 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:04 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:06 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:08 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:09 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:11 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:13 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:15 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:17 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:18 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:20 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:21 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:23 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:25 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:27 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:29 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:31 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:33 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:34 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:36 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:38 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:40 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:42 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:44 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:46 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:47 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:49 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:50 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:52 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:54 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:56 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:58 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:11:59 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:01 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:03 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:05 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:06 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:08 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:10 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:11 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:13 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:15 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:17 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:19 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:21 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:22 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:24 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:26 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:28 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:30 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:32 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:34 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:36 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:38 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:39 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:41 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:42 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:44 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:46 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:48 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:50 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:52 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:54 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:55 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:57 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:12:59 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:01 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:03 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:05 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:07 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:08 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:10 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:12 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:14 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:15 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:17 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:19 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:21 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:23 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:25 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:27 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:29 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:31 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:32 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:34 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:35 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:37 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:39 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:41 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:43 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:45 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:46 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:48 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:50 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:52 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:54 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:55 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:57 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:13:59 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:01 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:03 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:04 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:06 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:08 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:10 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:12 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:13 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:15 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:17 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:19 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:21 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:23 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:24 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:26 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:28 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:30 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:32 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:34 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:35 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:38 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:39 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:41 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:43 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:44 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:46 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:48 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:50 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:52 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:54 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:56 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:14:58 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:00 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:02 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:04 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:06 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:07 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:09 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:11 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:13 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:15 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:16 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:18 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:20 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:22 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:23 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:25 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:26 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:28 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:30 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:32 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:34 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:36 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:38 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:40 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:42 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:43 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:45 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:47 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:49 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:50 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:52 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:54 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:56 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:15:58 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:16:00 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:16:02 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:16:03 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:16:05 pxc-restore/restore-on-pitr-minio-gtid state: Starting 2025-11-12T12:16:06 pxc-restore/restore-on-pitr-minio-gtid state: Starting apiVersion: pxc.percona.com/v1 kind: PerconaXtraDBClusterRestore metadata: annotations: kubectl.kubernetes.io/last-applied-configuration: | {"apiVersion":"pxc.percona.com/v1","kind":"PerconaXtraDBClusterRestore","metadata":{"annotations":{},"name":"restore-on-pitr-minio-gtid","namespace":"pitr-4490"},"spec":{"backupName":"on-pitr-minio","pitr":{"backupSource":{"storageName":"minio-binlogs"},"gtid":"da4ab97b-bfbb-11f0-a3dd-8f0b60e73819:38","type":"transaction"},"pxcCluster":"pitr"}} creationTimestamp: "2025-11-12T11:53:31Z" generation: 1 name: restore-on-pitr-minio-gtid namespace: pitr-4490 resourceVersion: "1762948411859839020" uid: 3c31a74a-e8cd-4ccf-840a-4ff2dd44754c spec: backupName: on-pitr-minio pitr: backupSource: storageName: minio-binlogs gtid: da4ab97b-bfbb-11f0-a3dd-8f0b60e73819:38 type: transaction pxcCluster: pitr status: state: Starting unsafeFlags: {} Restore object pxc-restore/restore-on-pitr-minio-gtid is in Starting state. something went wrong with operator or kubernetes cluster