=== RUN   kuttl
    harness.go:464: starting setup
    harness.go:255: running tests using configured kubeconfig.
    harness.go:278: Successful connection to cluster at: https://34.136.112.110
    harness.go:363: running tests
    harness.go:75: going to run test suite with timeout of 180 seconds for each step
    harness.go:375: testsuite: e2e-tests/tests has 34 tests
=== RUN   kuttl/harness
=== RUN   kuttl/harness/gr-demand-backup
=== PAUSE kuttl/harness/gr-demand-backup
=== CONT  kuttl/harness/gr-demand-backup
    logger.go:42: 01:59:49 | gr-demand-backup | Creating namespace: kuttl-test-ready-deer
    logger.go:42: 01:59:49 | gr-demand-backup/0-minio-secret | starting test step 0-minio-secret
    logger.go:42: 01:59:50 | gr-demand-backup/0-minio-secret | Secret:kuttl-test-ready-deer/minio-secret created
    logger.go:42: 01:59:50 | gr-demand-backup/0-minio-secret | test step completed 0-minio-secret
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | starting test step 1-deploy-operator
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        init_temp_dir # do this only in the first TestStep
        
        apply_s3_storage_secrets
        deploy_operator
        deploy_non_tls_cluster_secrets
        deploy_tls_cluster_secrets
        deploy_client
        deploy_minio]
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | + source ../../functions
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ realpath ../../..
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | ++++ pwd
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/tests/gr-demand-backup
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | ++ test_name=gr-demand-backup
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/vars.sh
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ export GIT_BRANCH=PR-875
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ GIT_BRANCH=PR-875
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ export VERSION=PR-875-8c227d86
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ VERSION=PR-875-8c227d86
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ export PMM_SERVER_VERSION=1.4.0
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ PMM_SERVER_VERSION=1.4.0
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ export CERT_MANAGER_VER=1.16.3
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ CERT_MANAGER_VER=1.16.3
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | ++++ which gdate
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-875/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | ++++ which date
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ date=/usr/bin/date
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ oc get projects
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ :
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ kubectl get nodes
    logger.go:42: 01:59:50 | gr-demand-backup/1-deploy-operator | +++ grep '^minikube'
    logger.go:42: 01:59:51 | gr-demand-backup/1-deploy-operator | + init_temp_dir
    logger.go:42: 01:59:51 | gr-demand-backup/1-deploy-operator | + rm -rf /tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 01:59:51 | gr-demand-backup/1-deploy-operator | + mkdir -p /tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 01:59:51 | gr-demand-backup/1-deploy-operator | + apply_s3_storage_secrets
    logger.go:42: 01:59:51 | gr-demand-backup/1-deploy-operator | + kubectl -n kuttl-test-ready-deer apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf/minio-secret.yml
    logger.go:42: 01:59:53 | gr-demand-backup/1-deploy-operator | Warning: resource secrets/minio-secret is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically.
    logger.go:42: 01:59:53 | gr-demand-backup/1-deploy-operator | secret/minio-secret configured
    logger.go:42: 01:59:53 | gr-demand-backup/1-deploy-operator | + kubectl -n kuttl-test-ready-deer apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf/cloud-secret.yml
    logger.go:42: 01:59:54 | gr-demand-backup/1-deploy-operator | secret/aws-s3-secret created
    logger.go:42: 01:59:54 | gr-demand-backup/1-deploy-operator | secret/gcp-cs-secret created
    logger.go:42: 01:59:54 | gr-demand-backup/1-deploy-operator | secret/azure-secret created
    logger.go:42: 01:59:54 | gr-demand-backup/1-deploy-operator | + deploy_operator
    logger.go:42: 01:59:54 | gr-demand-backup/1-deploy-operator | + destroy_operator
    logger.go:42: 01:59:54 | gr-demand-backup/1-deploy-operator | + kubectl -n ps-operator delete deployment percona-server-mysql-operator --force --grace-period=0
    logger.go:42: 01:59:54 | gr-demand-backup/1-deploy-operator | Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.
    logger.go:42: 01:59:55 | gr-demand-backup/1-deploy-operator | Error from server (NotFound): deployments.apps "percona-server-mysql-operator" not found
    logger.go:42: 01:59:55 | gr-demand-backup/1-deploy-operator | + true
    logger.go:42: 01:59:55 | gr-demand-backup/1-deploy-operator | + [[ -n ps-operator ]]
    logger.go:42: 01:59:55 | gr-demand-backup/1-deploy-operator | + kubectl delete namespace ps-operator --force --grace-period=0
    logger.go:42: 01:59:55 | gr-demand-backup/1-deploy-operator | Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.
    logger.go:42: 01:59:55 | gr-demand-backup/1-deploy-operator | Error from server (NotFound): namespaces "ps-operator" not found
    logger.go:42: 01:59:55 | gr-demand-backup/1-deploy-operator | + true
    logger.go:42: 01:59:55 | gr-demand-backup/1-deploy-operator | + [[ -n ps-operator ]]
    logger.go:42: 01:59:55 | gr-demand-backup/1-deploy-operator | + create_namespace ps-operator
    logger.go:42: 01:59:55 | gr-demand-backup/1-deploy-operator | + local namespace=ps-operator
    logger.go:42: 01:59:55 | gr-demand-backup/1-deploy-operator | + [[ -n '' ]]
    logger.go:42: 01:59:55 | gr-demand-backup/1-deploy-operator | + kubectl delete namespace ps-operator --ignore-not-found
    logger.go:42: 01:59:56 | gr-demand-backup/1-deploy-operator | + kubectl wait --for=delete namespace ps-operator
    logger.go:42: 01:59:56 | gr-demand-backup/1-deploy-operator | + kubectl create namespace ps-operator
    logger.go:42: 01:59:56 | gr-demand-backup/1-deploy-operator | namespace/ps-operator created
    logger.go:42: 01:59:56 | gr-demand-backup/1-deploy-operator | + kubectl -n ps-operator apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy/crd.yaml
    logger.go:42: 01:59:57 | gr-demand-backup/1-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlbackups.ps.percona.com serverside-applied
    logger.go:42: 01:59:58 | gr-demand-backup/1-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlrestores.ps.percona.com serverside-applied
    logger.go:42: 01:59:59 | gr-demand-backup/1-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqls.ps.percona.com serverside-applied
    logger.go:42: 01:59:59 | gr-demand-backup/1-deploy-operator | + '[' -n ps-operator ']'
    logger.go:42: 01:59:59 | gr-demand-backup/1-deploy-operator | + kubectl -n ps-operator apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy/cw-rbac.yaml
    logger.go:42: 02:00:00 | gr-demand-backup/1-deploy-operator | serviceaccount/percona-server-mysql-operator created
    logger.go:42: 02:00:00 | gr-demand-backup/1-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created
    logger.go:42: 02:00:01 | gr-demand-backup/1-deploy-operator | clusterrole.rbac.authorization.k8s.io/percona-server-mysql-operator created
    logger.go:42: 02:00:01 | gr-demand-backup/1-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created
    logger.go:42: 02:00:01 | gr-demand-backup/1-deploy-operator | clusterrolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator created
    logger.go:42: 02:00:01 | gr-demand-backup/1-deploy-operator | + yq eval '(select(documentIndex==1).spec.template.spec.containers[] | select(.name=="manager").env[] | select(.name=="DISABLE_TELEMETRY").value) = "true"'
    logger.go:42: 02:00:01 | gr-demand-backup/1-deploy-operator | + yq eval '(select(documentIndex==1).spec.template.spec.containers[] | select(.name=="manager").env[] | select(.name=="LOG_LEVEL").value) = "DEBUG"'
    logger.go:42: 02:00:01 | gr-demand-backup/1-deploy-operator | + kubectl -n ps-operator apply -f -
    logger.go:42: 02:00:01 | gr-demand-backup/1-deploy-operator | ++ printf 'select(documentIndex==1).spec.template.spec.containers[0].image="%s"' perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 02:00:01 | gr-demand-backup/1-deploy-operator | + yq eval 'select(documentIndex==1).spec.template.spec.containers[0].image="perconalab/percona-server-mysql-operator:PR-875-8c227d86"' /mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy/cw-operator.yaml
    logger.go:42: 02:00:03 | gr-demand-backup/1-deploy-operator | configmap/percona-server-mysql-operator-config created
    logger.go:42: 02:00:03 | gr-demand-backup/1-deploy-operator | deployment.apps/percona-server-mysql-operator created
    logger.go:42: 02:00:03 | gr-demand-backup/1-deploy-operator | + deploy_non_tls_cluster_secrets
    logger.go:42: 02:00:03 | gr-demand-backup/1-deploy-operator | + kubectl -n kuttl-test-ready-deer apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf/secrets.yaml
    logger.go:42: 02:00:04 | gr-demand-backup/1-deploy-operator | secret/test-secrets created
    logger.go:42: 02:00:04 | gr-demand-backup/1-deploy-operator | + deploy_tls_cluster_secrets
    logger.go:42: 02:00:04 | gr-demand-backup/1-deploy-operator | + kubectl -n kuttl-test-ready-deer apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf/ssl-secret.yaml
    logger.go:42: 02:00:05 | gr-demand-backup/1-deploy-operator | secret/test-ssl created
    logger.go:42: 02:00:05 | gr-demand-backup/1-deploy-operator | + deploy_client
    logger.go:42: 02:00:05 | gr-demand-backup/1-deploy-operator | + kubectl -n kuttl-test-ready-deer apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf/client.yaml
    logger.go:42: 02:00:06 | gr-demand-backup/1-deploy-operator | pod/mysql-client created
    logger.go:42: 02:00:06 | gr-demand-backup/1-deploy-operator | + deploy_minio
    logger.go:42: 02:00:06 | gr-demand-backup/1-deploy-operator | + local access_key
    logger.go:42: 02:00:06 | gr-demand-backup/1-deploy-operator | + local secret_key
    logger.go:42: 02:00:06 | gr-demand-backup/1-deploy-operator | ++ kubectl -n kuttl-test-ready-deer get secret minio-secret -o 'jsonpath={.data.AWS_ACCESS_KEY_ID}'
    logger.go:42: 02:00:06 | gr-demand-backup/1-deploy-operator | ++ base64 -d
    logger.go:42: 02:00:06 | gr-demand-backup/1-deploy-operator | + access_key=some-access-key
    logger.go:42: 02:00:06 | gr-demand-backup/1-deploy-operator | ++ kubectl -n kuttl-test-ready-deer get secret minio-secret -o 'jsonpath={.data.AWS_SECRET_ACCESS_KEY}'
    logger.go:42: 02:00:06 | gr-demand-backup/1-deploy-operator | ++ base64 -d
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | + secret_key=some-secret-key
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | + helm uninstall -n kuttl-test-ready-deer minio-service
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-875/kubeconfig
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-875/kubeconfig
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | Error: uninstall: Release not loaded: minio-service: release: not found
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | + :
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | + helm repo remove minio
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-875/kubeconfig
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-875/kubeconfig
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | Error: no repositories configured
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | + :
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | + helm repo add minio https://charts.min.io/
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-875/kubeconfig
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-875/kubeconfig
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | "minio" has been added to your repositories
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | +++ printf %q some-access-key
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | ++ printf %q some-access-key
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | +++ printf %q some-secret-key
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | ++ printf %q some-secret-key
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | + retry 10 60 helm install minio-service -n kuttl-test-ready-deer --version 5.0.14 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/.minio/ --set persistence.size=2G --set securityContext.enabled=false minio/minio
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | + local max=10
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | + local delay=60
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | + shift 2
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | + local n=1
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | + helm install minio-service -n kuttl-test-ready-deer --version 5.0.14 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/.minio/ --set persistence.size=2G --set securityContext.enabled=false minio/minio
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-875/kubeconfig
    logger.go:42: 02:00:07 | gr-demand-backup/1-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-875/kubeconfig
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | NAME: minio-service
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | LAST DEPLOYED: Wed Mar 26 02:00:08 2025
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | NAMESPACE: kuttl-test-ready-deer
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | STATUS: deployed
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | REVISION: 1
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | TEST SUITE: None
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | NOTES:
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | MinIO can be accessed via port 9000 on the following DNS name from within your cluster:
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | minio-service.kuttl-test-ready-deer.svc.cluster.local
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | 
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | To access MinIO from localhost, run the below commands:
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | 
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator |   1. export POD_NAME=$(kubectl get pods --namespace kuttl-test-ready-deer -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}")
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | 
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator |   2. kubectl port-forward $POD_NAME 9000 --namespace kuttl-test-ready-deer
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | 
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | 
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | You can now access MinIO server on http://localhost:9000. Follow the below steps to connect to MinIO server with mc client:
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | 
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator |   1. Download the MinIO mc client - https://min.io/docs/minio/linux/reference/minio-mc.html#quickstart
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | 
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator |   2. export MC_HOST_minio-service-local=http://$(kubectl get secret --namespace kuttl-test-ready-deer minio-service -o jsonpath="{.data.rootUser}" | base64 --decode):$(kubectl get secret --namespace kuttl-test-ready-deer minio-service -o jsonpath="{.data.rootPassword}" | base64 --decode)@localhost:9000
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | 
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator |   3. mc ls minio-service-local
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | ++ kubectl -n kuttl-test-ready-deer get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | + MINIO_POD=minio-service-847fc8bb8d-5b565
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | + wait_pod minio-service-847fc8bb8d-5b565
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | + local pod=minio-service-847fc8bb8d-5b565
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | + set +o xtrace
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | minio-service-847fc8bb8d-5b565true
    logger.go:42: 02:00:46 | gr-demand-backup/1-deploy-operator | + kubectl -n kuttl-test-ready-deer run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- bash -c 'AWS_ACCESS_KEY_ID='\''some-access-key'\'' AWS_SECRET_ACCESS_KEY='\''some-secret-key'\'' AWS_DEFAULT_REGION=us-east-1         /usr/bin/aws --endpoint-url http://minio-service:9000 s3 mb s3://operator-testing'
    logger.go:42: 02:00:51 | gr-demand-backup/1-deploy-operator | If you don't see a command prompt, try pressing enter.
    logger.go:42: 02:00:52 | gr-demand-backup/1-deploy-operator | warning: couldn't attach to pod/aws-cli, falling back to streaming logs: Internal error occurred: error attaching to container: container is in CONTAINER_EXITED state
    logger.go:42: 02:00:52 | gr-demand-backup/1-deploy-operator | make_bucket: operator-testing
    logger.go:42: 02:00:54 | gr-demand-backup/1-deploy-operator | pod "aws-cli" deleted
[controller-runtime] log.SetLogger(...) was never called; logs will not be displayed.
Detected at:
	>  goroutine 14 [running]:
	>  runtime/debug.Stack()
	>  	/nix/store/wkbckbd30nlhq4dxzg64q6y4vm1xx4fk-go-1.22.1/share/go/src/runtime/debug/stack.go:24 +0x5e
	>  sigs.k8s.io/controller-runtime/pkg/log.eventuallyFulfillRoot()
	>  	/home/mowsiany/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.3/pkg/log/log.go:60 +0xcd
	>  sigs.k8s.io/controller-runtime/pkg/log.(*delegatingLogSink).WithName(0xc0002a9c00, {0x184a055, 0x14})
	>  	/home/mowsiany/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.3/pkg/log/deleg.go:147 +0x3e
	>  github.com/go-logr/logr.Logger.WithName({{0x1acb7d8, 0xc0002a9c00}, 0x0}, {0x184a055?, 0xc00035bf80?})
	>  	/home/mowsiany/go/pkg/mod/github.com/go-logr/logr@v1.2.4/logr.go:336 +0x36
	>  sigs.k8s.io/controller-runtime/pkg/client.newClient(0x131ead3?, {0x0, 0xc00043ca10, {0x1accd90, 0xc000368e40}, 0x0, {0x0, 0x0}, 0x0})
	>  	/home/mowsiany/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.3/pkg/client/client.go:122 +0xf1
	>  sigs.k8s.io/controller-runtime/pkg/client.New(0xc0000ec6c8?, {0x0, 0xc00043ca10, {0x1accd90, 0xc000368e40}, 0x0, {0x0, 0x0}, 0x0})
	>  	/home/mowsiany/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.3/pkg/client/client.go:103 +0x7d
	>  github.com/kudobuilder/kuttl/pkg/test/utils.NewRetryClient(0xc0000ec6c8, {0x0, 0xc00043ca10, {0x1accd90, 0xc000368e40}, 0x0, {0x0, 0x0}, 0x0})
	>  	/home/mowsiany/go/src/github.com/kudobuilder/kuttl/pkg/test/utils/kubernetes.go:177 +0x127
	>  github.com/kudobuilder/kuttl/pkg/test.(*Harness).Client(0xc0003a6908, 0xe6?)
	>  	/home/mowsiany/go/src/github.com/kudobuilder/kuttl/pkg/test/harness.go:323 +0x18e
	>  github.com/kudobuilder/kuttl/pkg/test.(*Step).Create(0xc000686f70, 0xc0004b04e0, {0xc00047d998, 0x15})
	>  	/home/mowsiany/go/src/github.com/kudobuilder/kuttl/pkg/test/step.go:177 +0x63
	>  github.com/kudobuilder/kuttl/pkg/test.(*Step).Run(0xc000686f70, 0xc0004b04e0, {0xc00047d998, 0x15})
	>  	/home/mowsiany/go/src/github.com/kudobuilder/kuttl/pkg/test/step.go:457 +0x24a
	>  github.com/kudobuilder/kuttl/pkg/test.(*Case).Run(0xc0003ad7c0, 0xc0004b04e0, 0xc000542e10)
	>  	/home/mowsiany/go/src/github.com/kudobuilder/kuttl/pkg/test/case.go:373 +0xaeb
	>  github.com/kudobuilder/kuttl/pkg/test.(*Harness).RunTests.func1.1(0xc0004b04e0)
	>  	/home/mowsiany/go/src/github.com/kudobuilder/kuttl/pkg/test/harness.go:401 +0x12e
	>  testing.tRunner(0xc0004b04e0, 0xc0004785b8)
	>  	/nix/store/wkbckbd30nlhq4dxzg64q6y4vm1xx4fk-go-1.22.1/share/go/src/testing/testing.go:1689 +0xfb
	>  created by testing.(*T).Run in goroutine 13
	>  	/nix/store/wkbckbd30nlhq4dxzg64q6y4vm1xx4fk-go-1.22.1/share/go/src/testing/testing.go:1742 +0x390
    logger.go:42: 02:00:55 | gr-demand-backup/1-deploy-operator | running command: [sh -c kubectl assert exist-enhanced deployment percona-server-mysql-operator -n ${OPERATOR_NS:-$NAMESPACE} --field-selector status.readyReplicas=1]
    logger.go:42: 02:00:55 | gr-demand-backup/1-deploy-operator | ASSERT deployment percona-server-mysql-operator matching field criteria 'status.readyReplicas=1' should exist.
    logger.go:42: 02:00:55 | gr-demand-backup/1-deploy-operator | INFO   Found 1 resource(s).
    logger.go:42: 02:00:55 | gr-demand-backup/1-deploy-operator | NAME                            NAMESPACE     COL0
    logger.go:42: 02:00:55 | gr-demand-backup/1-deploy-operator | percona-server-mysql-operator   ps-operator   1
    logger.go:42: 02:00:55 | gr-demand-backup/1-deploy-operator | ASSERT PASS
    logger.go:42: 02:00:55 | gr-demand-backup/1-deploy-operator | test step completed 1-deploy-operator
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | starting test step 2-create-cluster
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        get_cr \
            | yq eval ".spec.backup.backoffLimit=3" - \
            | yq eval '.spec.backup.storages.minio.type="s3"' - \
            | yq eval '.spec.backup.storages.minio.s3.bucket="operator-testing"' - \
            | yq eval '.spec.backup.storages.minio.s3.credentialsSecret="minio-secret"' - \
            | yq eval ".spec.backup.storages.minio.s3.endpointUrl=\"http://minio-service.${NAMESPACE}:9000\"" - \
            | yq eval '.spec.backup.storages.minio.s3.region="us-east-1"' - \
            | yq eval '.spec.backup.storages.aws-s3.type="s3"' - \
            | yq eval ".spec.backup.storages.aws-s3.verifyTLS=true" - \
            | yq eval '.spec.backup.storages.aws-s3.s3.bucket="operator-testing"' - \
            | yq eval '.spec.backup.storages.aws-s3.s3.credentialsSecret="aws-s3-secret"' - \
            | yq eval '.spec.backup.storages.aws-s3.s3.region="us-east-1"' - \
            | yq eval '.spec.backup.storages.aws-s3.s3.prefix="ps"' - \
            | yq eval '.spec.backup.storages.gcp-cs.type="gcs"' - \
            | yq eval ".spec.backup.storages.gcp-cs.verifyTLS=true" - \
            | yq eval '.spec.backup.storages.gcp-cs.gcs.bucket="operator-testing"' - \
            | yq eval '.spec.backup.storages.gcp-cs.gcs.credentialsSecret="gcp-cs-secret"' - \
            | yq eval '.spec.backup.storages.gcp-cs.gcs.endpointUrl="https://storage.googleapis.com"' - \
            | yq eval '.spec.backup.storages.gcp-cs.gcs.prefix="ps"' - \
            | yq eval '.spec.backup.storages.azure-blob.type="azure"' - \
            | yq eval ".spec.backup.storages.azure-blob.verifyTLS=true" - \
            | yq eval '.spec.backup.storages.azure-blob.azure.containerName="operator-testing"' - \
            | yq eval '.spec.backup.storages.azure-blob.azure.credentialsSecret="azure-secret"' - \
            | yq eval '.spec.backup.storages.azure-blob.azure.prefix="ps"' - \
            | yq eval '.spec.mysql.clusterType="group-replication"' - \
            | yq eval ".spec.proxy.router.enabled=true" - \
            | yq eval ".spec.proxy.haproxy.enabled=false" - \
            | kubectl -n "${NAMESPACE}" apply -f -]
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | + source ../../functions
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ realpath ../../..
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | ++++ pwd
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/tests/gr-demand-backup
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | ++ test_name=gr-demand-backup
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/vars.sh
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ export GIT_BRANCH=PR-875
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ GIT_BRANCH=PR-875
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ export VERSION=PR-875-8c227d86
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ VERSION=PR-875-8c227d86
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ export PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ export CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | ++++ which gdate
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-875/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | ++++ which date
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ date=/usr/bin/date
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ oc get projects
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ :
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ kubectl get nodes
    logger.go:42: 02:00:55 | gr-demand-backup/2-create-cluster | +++ grep '^minikube'
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + get_cr
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + local name_suffix=
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval .spec.backup.backoffLimit=3 -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.minio.s3.bucket="operator-testing"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.minio.type="s3"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval .spec.orchestrator.enabled=true -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.minio.s3.region="us-east-1"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.gcp-cs.gcs.prefix="ps"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.aws-s3.type="s3"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.gcp-cs.gcs.credentialsSecret="gcp-cs-secret"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.gcp-cs.gcs.endpointUrl="https://storage.googleapis.com"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval .spec.backup.storages.gcp-cs.verifyTLS=true -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.gcp-cs.gcs.bucket="operator-testing"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.aws-s3.s3.region="us-east-1"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.aws-s3.s3.credentialsSecret="aws-s3-secret"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.gcp-cs.type="gcs"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.aws-s3.s3.prefix="ps"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.aws-s3.s3.bucket="operator-testing"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval .spec.backup.storages.aws-s3.verifyTLS=true -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.minio.s3.endpointUrl="http://minio-service.kuttl-test-ready-deer:9000"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.azure-blob.type="azure"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval .spec.proxy.haproxy.enabled=false -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + kubectl -n kuttl-test-ready-deer apply -f -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.minio.s3.credentialsSecret="minio-secret"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval .spec.backup.storages.azure-blob.verifyTLS=true -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.azure-blob.azure.containerName="operator-testing"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | ++ printf '.spec.mysql.image="%s"' perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.mysql.image="perconalab/percona-server-mysql-operator:main-psmysql"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.azure-blob.azure.credentialsSecret="azure-secret"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.azure-blob.azure.prefix="ps"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.mysql.clusterType="group-replication"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval .spec.proxy.router.enabled=true -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | ++ printf '.spec.backup.image="%s"' perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.image="perconalab/percona-server-mysql-operator:main-backup"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | ++ printf '.spec.toolkit.image="%s"' perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.toolkit.image="perconalab/percona-server-mysql-operator:main-toolkit"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.mysql.clusterType="async"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | ++ printf '.metadata.name="%s"' gr-demand-backup
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | ++ printf '.spec.orchestrator.image="%s"' perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.orchestrator.image="perconalab/percona-server-mysql-operator:main-orchestrator"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.secretsName="test-secrets"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.metadata.name="gr-demand-backup"' /mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy/cr.yaml
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | ++ printf '.spec.initImage="%s"' perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.initImage="perconalab/percona-server-mysql-operator:PR-875-8c227d86"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.sslSecretName="test-ssl"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.upgradeOptions.apply="disabled"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + '[' -n '' ']'
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | ++ printf '.spec.proxy.router.image="%s"' perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.proxy.router.image="perconalab/percona-server-mysql-operator:main-router"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | ++ printf '.spec.pmm.image="%s"' perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | ++ printf '.spec.proxy.haproxy.image="%s"' perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.proxy.haproxy.image="perconalab/percona-server-mysql-operator:main-haproxy"' -
    logger.go:42: 02:00:56 | gr-demand-backup/2-create-cluster | + yq eval '.spec.pmm.image="perconalab/pmm-client:3-dev-latest"' -
    logger.go:42: 02:00:57 | gr-demand-backup/2-create-cluster | perconaservermysql.ps.percona.com/gr-demand-backup created
    logger.go:42: 02:06:43 | gr-demand-backup/2-create-cluster | test step completed 2-create-cluster
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | starting test step 3-write-data
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | running command: [sh -c set -o errexit
        set -o pipefail
        set -o xtrace
        
        source ../../functions
        
        run_mysql \
            "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" \
            "-h $(get_mysql_router_service $(get_cluster_name)) -uroot -proot_password"
        
        run_mysql \
            "INSERT myDB.myTable (id) VALUES (100500)" \
            "-h $(get_mysql_router_service $(get_cluster_name)) -uroot -proot_password"]
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | + source ../../functions
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ realpath ../../..
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | ++++ pwd
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/tests/gr-demand-backup
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | ++ test_name=gr-demand-backup
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/vars.sh
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ export GIT_BRANCH=PR-875
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ GIT_BRANCH=PR-875
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ export VERSION=PR-875-8c227d86
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ VERSION=PR-875-8c227d86
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ export PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ export CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | ++++ which gdate
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-875/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | ++++ which date
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ date=/usr/bin/date
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ oc get projects
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ :
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ kubectl get nodes
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ grep '^minikube'
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ get_cluster_name
    logger.go:42: 02:06:43 | gr-demand-backup/3-write-data | +++ kubectl -n kuttl-test-ready-deer get ps -o 'jsonpath={.items[0].metadata.name}'
    logger.go:42: 02:06:44 | gr-demand-backup/3-write-data | ++ get_mysql_router_service gr-demand-backup
    logger.go:42: 02:06:44 | gr-demand-backup/3-write-data | ++ local cluster=gr-demand-backup
    logger.go:42: 02:06:44 | gr-demand-backup/3-write-data | ++ echo gr-demand-backup-router
    logger.go:42: 02:06:44 | gr-demand-backup/3-write-data | + run_mysql 'CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' '-h gr-demand-backup-router -uroot -proot_password'
    logger.go:42: 02:06:44 | gr-demand-backup/3-write-data | + local 'command=CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)'
    logger.go:42: 02:06:44 | gr-demand-backup/3-write-data | + local 'uri=-h gr-demand-backup-router -uroot -proot_password'
    logger.go:42: 02:06:44 | gr-demand-backup/3-write-data | + local pod=
    logger.go:42: 02:06:44 | gr-demand-backup/3-write-data | ++ get_client_pod
    logger.go:42: 02:06:44 | gr-demand-backup/3-write-data | ++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:06:44 | gr-demand-backup/3-write-data | + client_pod=mysql-client
    logger.go:42: 02:06:44 | gr-demand-backup/3-write-data | + wait_pod mysql-client
    logger.go:42: 02:06:44 | gr-demand-backup/3-write-data | + local pod=mysql-client
    logger.go:42: 02:06:44 | gr-demand-backup/3-write-data | + set +o xtrace
    logger.go:42: 02:06:44 | gr-demand-backup/3-write-data | mysql-clienttrue
    logger.go:42: 02:06:44 | gr-demand-backup/3-write-data | + sed -e 's/mysql: //'
    logger.go:42: 02:06:44 | gr-demand-backup/3-write-data | + grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:06:44 | gr-demand-backup/3-write-data | + kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" | mysql -sN -h gr-demand-backup-router -uroot -proot_password'
    logger.go:42: 02:06:46 | gr-demand-backup/3-write-data | + :
    logger.go:42: 02:06:46 | gr-demand-backup/3-write-data | +++ get_cluster_name
    logger.go:42: 02:06:46 | gr-demand-backup/3-write-data | +++ kubectl -n kuttl-test-ready-deer get ps -o 'jsonpath={.items[0].metadata.name}'
    logger.go:42: 02:06:46 | gr-demand-backup/3-write-data | ++ get_mysql_router_service gr-demand-backup
    logger.go:42: 02:06:46 | gr-demand-backup/3-write-data | ++ local cluster=gr-demand-backup
    logger.go:42: 02:06:46 | gr-demand-backup/3-write-data | ++ echo gr-demand-backup-router
    logger.go:42: 02:06:46 | gr-demand-backup/3-write-data | + run_mysql 'INSERT myDB.myTable (id) VALUES (100500)' '-h gr-demand-backup-router -uroot -proot_password'
    logger.go:42: 02:06:46 | gr-demand-backup/3-write-data | + local 'command=INSERT myDB.myTable (id) VALUES (100500)'
    logger.go:42: 02:06:46 | gr-demand-backup/3-write-data | + local 'uri=-h gr-demand-backup-router -uroot -proot_password'
    logger.go:42: 02:06:46 | gr-demand-backup/3-write-data | + local pod=
    logger.go:42: 02:06:46 | gr-demand-backup/3-write-data | ++ get_client_pod
    logger.go:42: 02:06:46 | gr-demand-backup/3-write-data | ++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:06:47 | gr-demand-backup/3-write-data | + client_pod=mysql-client
    logger.go:42: 02:06:47 | gr-demand-backup/3-write-data | + wait_pod mysql-client
    logger.go:42: 02:06:47 | gr-demand-backup/3-write-data | + local pod=mysql-client
    logger.go:42: 02:06:47 | gr-demand-backup/3-write-data | + set +o xtrace
    logger.go:42: 02:06:47 | gr-demand-backup/3-write-data | mysql-clienttrue
    logger.go:42: 02:06:47 | gr-demand-backup/3-write-data | + kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "INSERT myDB.myTable (id) VALUES (100500)" | mysql -sN -h gr-demand-backup-router -uroot -proot_password'
    logger.go:42: 02:06:47 | gr-demand-backup/3-write-data | + sed -e 's/mysql: //'
    logger.go:42: 02:06:47 | gr-demand-backup/3-write-data | + grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:06:49 | gr-demand-backup/3-write-data | + :
    logger.go:42: 02:06:49 | gr-demand-backup/3-write-data | test step completed 3-write-data
    logger.go:42: 02:06:49 | gr-demand-backup/4-create-backup-minio | starting test step 4-create-backup-minio
    logger.go:42: 02:06:49 | gr-demand-backup/4-create-backup-minio | PerconaServerMySQLBackup:kuttl-test-ready-deer/gr-demand-backup-minio created
    logger.go:42: 02:07:01 | gr-demand-backup/4-create-backup-minio | test step completed 4-create-backup-minio
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | starting test step 5-delete-data
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        run_mysql \
            "TRUNCATE TABLE myDB.myTable" \
            "-h $(get_mysql_router_service $(get_cluster_name)) -uroot -proot_password"
        
        cluster_name=$(get_cluster_name)
        for i in 0 1 2; do
            data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password")
            kubectl create configmap -n "${NAMESPACE}" 04-delete-data-minio-${i} --from-literal=data="${data}"
        done]
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | + source ../../functions
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ realpath ../../..
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | ++++ pwd
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/tests/gr-demand-backup
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | ++ test_name=gr-demand-backup
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/vars.sh
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ export GIT_BRANCH=PR-875
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ GIT_BRANCH=PR-875
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ export VERSION=PR-875-8c227d86
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ VERSION=PR-875-8c227d86
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ export PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ export CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | ++++ which gdate
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-875/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | ++++ which date
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ date=/usr/bin/date
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ oc get projects
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ :
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ kubectl get nodes
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ grep '^minikube'
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ get_cluster_name
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | +++ kubectl -n kuttl-test-ready-deer get ps -o 'jsonpath={.items[0].metadata.name}'
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | ++ get_mysql_router_service gr-demand-backup
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | ++ local cluster=gr-demand-backup
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | ++ echo gr-demand-backup-router
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | + run_mysql 'TRUNCATE TABLE myDB.myTable' '-h gr-demand-backup-router -uroot -proot_password'
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | + local 'command=TRUNCATE TABLE myDB.myTable'
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | + local 'uri=-h gr-demand-backup-router -uroot -proot_password'
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | + local pod=
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | ++ get_client_pod
    logger.go:42: 02:07:01 | gr-demand-backup/5-delete-data | ++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:07:02 | gr-demand-backup/5-delete-data | + client_pod=mysql-client
    logger.go:42: 02:07:02 | gr-demand-backup/5-delete-data | + wait_pod mysql-client
    logger.go:42: 02:07:02 | gr-demand-backup/5-delete-data | + local pod=mysql-client
    logger.go:42: 02:07:02 | gr-demand-backup/5-delete-data | + set +o xtrace
    logger.go:42: 02:07:02 | gr-demand-backup/5-delete-data | mysql-clienttrue
    logger.go:42: 02:07:02 | gr-demand-backup/5-delete-data | + kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "TRUNCATE TABLE myDB.myTable" | mysql -sN -h gr-demand-backup-router -uroot -proot_password'
    logger.go:42: 02:07:02 | gr-demand-backup/5-delete-data | + sed -e 's/mysql: //'
    logger.go:42: 02:07:02 | gr-demand-backup/5-delete-data | + grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:07:04 | gr-demand-backup/5-delete-data | + :
    logger.go:42: 02:07:04 | gr-demand-backup/5-delete-data | ++ get_cluster_name
    logger.go:42: 02:07:04 | gr-demand-backup/5-delete-data | ++ kubectl -n kuttl-test-ready-deer get ps -o 'jsonpath={.items[0].metadata.name}'
    logger.go:42: 02:07:04 | gr-demand-backup/5-delete-data | + cluster_name=gr-demand-backup
    logger.go:42: 02:07:04 | gr-demand-backup/5-delete-data | + for i in 0 1 2
    logger.go:42: 02:07:04 | gr-demand-backup/5-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:07:04 | gr-demand-backup/5-delete-data | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 02:07:04 | gr-demand-backup/5-delete-data | ++ local 'uri=-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:07:04 | gr-demand-backup/5-delete-data | ++ local pod=
    logger.go:42: 02:07:04 | gr-demand-backup/5-delete-data | +++ get_client_pod
    logger.go:42: 02:07:04 | gr-demand-backup/5-delete-data | +++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:07:05 | gr-demand-backup/5-delete-data | ++ client_pod=mysql-client
    logger.go:42: 02:07:05 | gr-demand-backup/5-delete-data | ++ wait_pod mysql-client
    logger.go:42: 02:07:05 | gr-demand-backup/5-delete-data | ++ local pod=mysql-client
    logger.go:42: 02:07:05 | gr-demand-backup/5-delete-data | ++ set +o xtrace
    logger.go:42: 02:07:05 | gr-demand-backup/5-delete-data | mysql-clienttrue
    logger.go:42: 02:07:05 | gr-demand-backup/5-delete-data | ++ kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:07:05 | gr-demand-backup/5-delete-data | ++ sed -e 's/mysql: //'
    logger.go:42: 02:07:05 | gr-demand-backup/5-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:07:07 | gr-demand-backup/5-delete-data | ++ :
    logger.go:42: 02:07:07 | gr-demand-backup/5-delete-data | + data=
    logger.go:42: 02:07:07 | gr-demand-backup/5-delete-data | + kubectl create configmap -n kuttl-test-ready-deer 04-delete-data-minio-0 --from-literal=data=
    logger.go:42: 02:07:07 | gr-demand-backup/5-delete-data | configmap/04-delete-data-minio-0 created
    logger.go:42: 02:07:07 | gr-demand-backup/5-delete-data | + for i in 0 1 2
    logger.go:42: 02:07:07 | gr-demand-backup/5-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:07:07 | gr-demand-backup/5-delete-data | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 02:07:07 | gr-demand-backup/5-delete-data | ++ local 'uri=-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:07:07 | gr-demand-backup/5-delete-data | ++ local pod=
    logger.go:42: 02:07:07 | gr-demand-backup/5-delete-data | +++ get_client_pod
    logger.go:42: 02:07:07 | gr-demand-backup/5-delete-data | +++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:07:07 | gr-demand-backup/5-delete-data | ++ client_pod=mysql-client
    logger.go:42: 02:07:07 | gr-demand-backup/5-delete-data | ++ wait_pod mysql-client
    logger.go:42: 02:07:07 | gr-demand-backup/5-delete-data | ++ local pod=mysql-client
    logger.go:42: 02:07:07 | gr-demand-backup/5-delete-data | ++ set +o xtrace
    logger.go:42: 02:07:08 | gr-demand-backup/5-delete-data | mysql-clienttrue
    logger.go:42: 02:07:08 | gr-demand-backup/5-delete-data | ++ kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:07:08 | gr-demand-backup/5-delete-data | ++ sed -e 's/mysql: //'
    logger.go:42: 02:07:08 | gr-demand-backup/5-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:07:09 | gr-demand-backup/5-delete-data | ++ :
    logger.go:42: 02:07:09 | gr-demand-backup/5-delete-data | + data=
    logger.go:42: 02:07:09 | gr-demand-backup/5-delete-data | + kubectl create configmap -n kuttl-test-ready-deer 04-delete-data-minio-1 --from-literal=data=
    logger.go:42: 02:07:10 | gr-demand-backup/5-delete-data | configmap/04-delete-data-minio-1 created
    logger.go:42: 02:07:10 | gr-demand-backup/5-delete-data | + for i in 0 1 2
    logger.go:42: 02:07:10 | gr-demand-backup/5-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:07:10 | gr-demand-backup/5-delete-data | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 02:07:10 | gr-demand-backup/5-delete-data | ++ local 'uri=-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:07:10 | gr-demand-backup/5-delete-data | ++ local pod=
    logger.go:42: 02:07:10 | gr-demand-backup/5-delete-data | +++ get_client_pod
    logger.go:42: 02:07:10 | gr-demand-backup/5-delete-data | +++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:07:10 | gr-demand-backup/5-delete-data | ++ client_pod=mysql-client
    logger.go:42: 02:07:10 | gr-demand-backup/5-delete-data | ++ wait_pod mysql-client
    logger.go:42: 02:07:10 | gr-demand-backup/5-delete-data | ++ local pod=mysql-client
    logger.go:42: 02:07:10 | gr-demand-backup/5-delete-data | ++ set +o xtrace
    logger.go:42: 02:07:10 | gr-demand-backup/5-delete-data | mysql-clienttrue
    logger.go:42: 02:07:10 | gr-demand-backup/5-delete-data | ++ sed -e 's/mysql: //'
    logger.go:42: 02:07:10 | gr-demand-backup/5-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:07:10 | gr-demand-backup/5-delete-data | ++ kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:07:12 | gr-demand-backup/5-delete-data | ++ :
    logger.go:42: 02:07:12 | gr-demand-backup/5-delete-data | + data=
    logger.go:42: 02:07:12 | gr-demand-backup/5-delete-data | + kubectl create configmap -n kuttl-test-ready-deer 04-delete-data-minio-2 --from-literal=data=
    logger.go:42: 02:07:12 | gr-demand-backup/5-delete-data | configmap/04-delete-data-minio-2 created
    logger.go:42: 02:07:13 | gr-demand-backup/5-delete-data | test step completed 5-delete-data
    logger.go:42: 02:07:13 | gr-demand-backup/6-restore-from-minio | starting test step 6-restore-from-minio
    logger.go:42: 02:07:13 | gr-demand-backup/6-restore-from-minio | PerconaServerMySQLRestore:kuttl-test-ready-deer/gr-demand-backup-restore-minio created
    logger.go:42: 02:12:05 | gr-demand-backup/6-restore-from-minio | test step completed 6-restore-from-minio
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | starting test step 7-read-data
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        cluster_name=$(get_cluster_name)
        for i in 0 1 2; do
            data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password")
            kubectl create configmap -n "${NAMESPACE}" 06-read-data-minio-${i} --from-literal=data="${data}"
        done]
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | + source ../../functions
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ realpath ../../..
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | ++++ pwd
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/tests/gr-demand-backup
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | ++ test_name=gr-demand-backup
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/vars.sh
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ export GIT_BRANCH=PR-875
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ GIT_BRANCH=PR-875
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ export VERSION=PR-875-8c227d86
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ VERSION=PR-875-8c227d86
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ export PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ export CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | ++++ which gdate
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-875/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | ++++ which date
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ date=/usr/bin/date
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ oc get projects
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ :
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ kubectl get nodes
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | +++ grep '^minikube'
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | ++ get_cluster_name
    logger.go:42: 02:12:05 | gr-demand-backup/7-read-data | ++ kubectl -n kuttl-test-ready-deer get ps -o 'jsonpath={.items[0].metadata.name}'
    logger.go:42: 02:12:06 | gr-demand-backup/7-read-data | + cluster_name=gr-demand-backup
    logger.go:42: 02:12:06 | gr-demand-backup/7-read-data | + for i in 0 1 2
    logger.go:42: 02:12:06 | gr-demand-backup/7-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:12:06 | gr-demand-backup/7-read-data | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 02:12:06 | gr-demand-backup/7-read-data | ++ local 'uri=-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:12:06 | gr-demand-backup/7-read-data | ++ local pod=
    logger.go:42: 02:12:06 | gr-demand-backup/7-read-data | +++ get_client_pod
    logger.go:42: 02:12:06 | gr-demand-backup/7-read-data | +++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:12:06 | gr-demand-backup/7-read-data | ++ client_pod=mysql-client
    logger.go:42: 02:12:06 | gr-demand-backup/7-read-data | ++ wait_pod mysql-client
    logger.go:42: 02:12:06 | gr-demand-backup/7-read-data | ++ local pod=mysql-client
    logger.go:42: 02:12:06 | gr-demand-backup/7-read-data | ++ set +o xtrace
    logger.go:42: 02:12:07 | gr-demand-backup/7-read-data | mysql-clienttrue
    logger.go:42: 02:12:07 | gr-demand-backup/7-read-data | ++ kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:12:07 | gr-demand-backup/7-read-data | ++ sed -e 's/mysql: //'
    logger.go:42: 02:12:07 | gr-demand-backup/7-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:12:08 | gr-demand-backup/7-read-data | + data=100500
    logger.go:42: 02:12:08 | gr-demand-backup/7-read-data | + kubectl create configmap -n kuttl-test-ready-deer 06-read-data-minio-0 --from-literal=data=100500
    logger.go:42: 02:12:08 | gr-demand-backup/7-read-data | configmap/06-read-data-minio-0 created
    logger.go:42: 02:12:08 | gr-demand-backup/7-read-data | + for i in 0 1 2
    logger.go:42: 02:12:08 | gr-demand-backup/7-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:12:08 | gr-demand-backup/7-read-data | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 02:12:08 | gr-demand-backup/7-read-data | ++ local 'uri=-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:12:08 | gr-demand-backup/7-read-data | ++ local pod=
    logger.go:42: 02:12:08 | gr-demand-backup/7-read-data | +++ get_client_pod
    logger.go:42: 02:12:08 | gr-demand-backup/7-read-data | +++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:12:09 | gr-demand-backup/7-read-data | ++ client_pod=mysql-client
    logger.go:42: 02:12:09 | gr-demand-backup/7-read-data | ++ wait_pod mysql-client
    logger.go:42: 02:12:09 | gr-demand-backup/7-read-data | ++ local pod=mysql-client
    logger.go:42: 02:12:09 | gr-demand-backup/7-read-data | ++ set +o xtrace
    logger.go:42: 02:12:09 | gr-demand-backup/7-read-data | mysql-clienttrue
    logger.go:42: 02:12:09 | gr-demand-backup/7-read-data | ++ kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:12:09 | gr-demand-backup/7-read-data | ++ sed -e 's/mysql: //'
    logger.go:42: 02:12:09 | gr-demand-backup/7-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:12:11 | gr-demand-backup/7-read-data | + data=100500
    logger.go:42: 02:12:11 | gr-demand-backup/7-read-data | + kubectl create configmap -n kuttl-test-ready-deer 06-read-data-minio-1 --from-literal=data=100500
    logger.go:42: 02:12:11 | gr-demand-backup/7-read-data | configmap/06-read-data-minio-1 created
    logger.go:42: 02:12:11 | gr-demand-backup/7-read-data | + for i in 0 1 2
    logger.go:42: 02:12:11 | gr-demand-backup/7-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:12:11 | gr-demand-backup/7-read-data | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 02:12:11 | gr-demand-backup/7-read-data | ++ local 'uri=-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:12:11 | gr-demand-backup/7-read-data | ++ local pod=
    logger.go:42: 02:12:11 | gr-demand-backup/7-read-data | +++ get_client_pod
    logger.go:42: 02:12:11 | gr-demand-backup/7-read-data | +++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:12:11 | gr-demand-backup/7-read-data | ++ client_pod=mysql-client
    logger.go:42: 02:12:11 | gr-demand-backup/7-read-data | ++ wait_pod mysql-client
    logger.go:42: 02:12:11 | gr-demand-backup/7-read-data | ++ local pod=mysql-client
    logger.go:42: 02:12:11 | gr-demand-backup/7-read-data | ++ set +o xtrace
    logger.go:42: 02:12:12 | gr-demand-backup/7-read-data | mysql-clienttrue
    logger.go:42: 02:12:12 | gr-demand-backup/7-read-data | ++ kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:12:12 | gr-demand-backup/7-read-data | ++ sed -e 's/mysql: //'
    logger.go:42: 02:12:12 | gr-demand-backup/7-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:12:13 | gr-demand-backup/7-read-data | + data=100500
    logger.go:42: 02:12:13 | gr-demand-backup/7-read-data | + kubectl create configmap -n kuttl-test-ready-deer 06-read-data-minio-2 --from-literal=data=100500
    logger.go:42: 02:12:14 | gr-demand-backup/7-read-data | configmap/06-read-data-minio-2 created
    logger.go:42: 02:12:14 | gr-demand-backup/7-read-data | test step completed 7-read-data
    logger.go:42: 02:12:14 | gr-demand-backup/8-create-backup-s3 | starting test step 8-create-backup-s3
    logger.go:42: 02:12:15 | gr-demand-backup/8-create-backup-s3 | PerconaServerMySQLBackup:kuttl-test-ready-deer/gr-demand-backup-s3 created
    logger.go:42: 02:12:26 | gr-demand-backup/8-create-backup-s3 | test step completed 8-create-backup-s3
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | starting test step 9-delete-data
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        run_mysql \
            "TRUNCATE TABLE myDB.myTable" \
            "-h $(get_mysql_router_service $(get_cluster_name)) -uroot -proot_password"
        
        cluster_name=$(get_cluster_name)
        for i in 0 1 2; do
            data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password")
            kubectl create configmap -n "${NAMESPACE}" 08-delete-data-s3-${i} --from-literal=data="${data}"
        done]
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | + source ../../functions
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ realpath ../../..
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | ++++ pwd
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/tests/gr-demand-backup
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | ++ test_name=gr-demand-backup
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/vars.sh
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ export GIT_BRANCH=PR-875
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ GIT_BRANCH=PR-875
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ export VERSION=PR-875-8c227d86
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ VERSION=PR-875-8c227d86
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ export PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ export CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | ++++ which gdate
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-875/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | ++++ which date
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ date=/usr/bin/date
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ oc get projects
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ :
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ kubectl get nodes
    logger.go:42: 02:12:26 | gr-demand-backup/9-delete-data | +++ grep '^minikube'
    logger.go:42: 02:12:27 | gr-demand-backup/9-delete-data | +++ get_cluster_name
    logger.go:42: 02:12:27 | gr-demand-backup/9-delete-data | +++ kubectl -n kuttl-test-ready-deer get ps -o 'jsonpath={.items[0].metadata.name}'
    logger.go:42: 02:12:27 | gr-demand-backup/9-delete-data | ++ get_mysql_router_service gr-demand-backup
    logger.go:42: 02:12:27 | gr-demand-backup/9-delete-data | ++ local cluster=gr-demand-backup
    logger.go:42: 02:12:27 | gr-demand-backup/9-delete-data | ++ echo gr-demand-backup-router
    logger.go:42: 02:12:27 | gr-demand-backup/9-delete-data | + run_mysql 'TRUNCATE TABLE myDB.myTable' '-h gr-demand-backup-router -uroot -proot_password'
    logger.go:42: 02:12:27 | gr-demand-backup/9-delete-data | + local 'command=TRUNCATE TABLE myDB.myTable'
    logger.go:42: 02:12:27 | gr-demand-backup/9-delete-data | + local 'uri=-h gr-demand-backup-router -uroot -proot_password'
    logger.go:42: 02:12:27 | gr-demand-backup/9-delete-data | + local pod=
    logger.go:42: 02:12:27 | gr-demand-backup/9-delete-data | ++ get_client_pod
    logger.go:42: 02:12:27 | gr-demand-backup/9-delete-data | ++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:12:28 | gr-demand-backup/9-delete-data | + client_pod=mysql-client
    logger.go:42: 02:12:28 | gr-demand-backup/9-delete-data | + wait_pod mysql-client
    logger.go:42: 02:12:28 | gr-demand-backup/9-delete-data | + local pod=mysql-client
    logger.go:42: 02:12:28 | gr-demand-backup/9-delete-data | + set +o xtrace
    logger.go:42: 02:12:28 | gr-demand-backup/9-delete-data | mysql-clienttrue
    logger.go:42: 02:12:28 | gr-demand-backup/9-delete-data | + kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "TRUNCATE TABLE myDB.myTable" | mysql -sN -h gr-demand-backup-router -uroot -proot_password'
    logger.go:42: 02:12:28 | gr-demand-backup/9-delete-data | + sed -e 's/mysql: //'
    logger.go:42: 02:12:28 | gr-demand-backup/9-delete-data | + grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:12:29 | gr-demand-backup/9-delete-data | + :
    logger.go:42: 02:12:29 | gr-demand-backup/9-delete-data | ++ get_cluster_name
    logger.go:42: 02:12:29 | gr-demand-backup/9-delete-data | ++ kubectl -n kuttl-test-ready-deer get ps -o 'jsonpath={.items[0].metadata.name}'
    logger.go:42: 02:12:30 | gr-demand-backup/9-delete-data | + cluster_name=gr-demand-backup
    logger.go:42: 02:12:30 | gr-demand-backup/9-delete-data | + for i in 0 1 2
    logger.go:42: 02:12:30 | gr-demand-backup/9-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:12:30 | gr-demand-backup/9-delete-data | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 02:12:30 | gr-demand-backup/9-delete-data | ++ local 'uri=-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:12:30 | gr-demand-backup/9-delete-data | ++ local pod=
    logger.go:42: 02:12:30 | gr-demand-backup/9-delete-data | +++ get_client_pod
    logger.go:42: 02:12:30 | gr-demand-backup/9-delete-data | +++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:12:30 | gr-demand-backup/9-delete-data | ++ client_pod=mysql-client
    logger.go:42: 02:12:30 | gr-demand-backup/9-delete-data | ++ wait_pod mysql-client
    logger.go:42: 02:12:30 | gr-demand-backup/9-delete-data | ++ local pod=mysql-client
    logger.go:42: 02:12:30 | gr-demand-backup/9-delete-data | ++ set +o xtrace
    logger.go:42: 02:12:31 | gr-demand-backup/9-delete-data | mysql-clienttrue
    logger.go:42: 02:12:31 | gr-demand-backup/9-delete-data | ++ sed -e 's/mysql: //'
    logger.go:42: 02:12:31 | gr-demand-backup/9-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:12:31 | gr-demand-backup/9-delete-data | ++ kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:12:32 | gr-demand-backup/9-delete-data | ++ :
    logger.go:42: 02:12:32 | gr-demand-backup/9-delete-data | + data=
    logger.go:42: 02:12:32 | gr-demand-backup/9-delete-data | + kubectl create configmap -n kuttl-test-ready-deer 08-delete-data-s3-0 --from-literal=data=
    logger.go:42: 02:12:33 | gr-demand-backup/9-delete-data | configmap/08-delete-data-s3-0 created
    logger.go:42: 02:12:33 | gr-demand-backup/9-delete-data | + for i in 0 1 2
    logger.go:42: 02:12:33 | gr-demand-backup/9-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:12:33 | gr-demand-backup/9-delete-data | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 02:12:33 | gr-demand-backup/9-delete-data | ++ local 'uri=-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:12:33 | gr-demand-backup/9-delete-data | ++ local pod=
    logger.go:42: 02:12:33 | gr-demand-backup/9-delete-data | +++ get_client_pod
    logger.go:42: 02:12:33 | gr-demand-backup/9-delete-data | +++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:12:33 | gr-demand-backup/9-delete-data | ++ client_pod=mysql-client
    logger.go:42: 02:12:33 | gr-demand-backup/9-delete-data | ++ wait_pod mysql-client
    logger.go:42: 02:12:33 | gr-demand-backup/9-delete-data | ++ local pod=mysql-client
    logger.go:42: 02:12:33 | gr-demand-backup/9-delete-data | ++ set +o xtrace
    logger.go:42: 02:12:33 | gr-demand-backup/9-delete-data | mysql-clienttrue
    logger.go:42: 02:12:33 | gr-demand-backup/9-delete-data | ++ kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:12:33 | gr-demand-backup/9-delete-data | ++ sed -e 's/mysql: //'
    logger.go:42: 02:12:33 | gr-demand-backup/9-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:12:35 | gr-demand-backup/9-delete-data | ++ :
    logger.go:42: 02:12:35 | gr-demand-backup/9-delete-data | + data=
    logger.go:42: 02:12:35 | gr-demand-backup/9-delete-data | + kubectl create configmap -n kuttl-test-ready-deer 08-delete-data-s3-1 --from-literal=data=
    logger.go:42: 02:12:36 | gr-demand-backup/9-delete-data | configmap/08-delete-data-s3-1 created
    logger.go:42: 02:12:36 | gr-demand-backup/9-delete-data | + for i in 0 1 2
    logger.go:42: 02:12:36 | gr-demand-backup/9-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:12:36 | gr-demand-backup/9-delete-data | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 02:12:36 | gr-demand-backup/9-delete-data | ++ local 'uri=-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:12:36 | gr-demand-backup/9-delete-data | ++ local pod=
    logger.go:42: 02:12:36 | gr-demand-backup/9-delete-data | +++ get_client_pod
    logger.go:42: 02:12:36 | gr-demand-backup/9-delete-data | +++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:12:36 | gr-demand-backup/9-delete-data | ++ client_pod=mysql-client
    logger.go:42: 02:12:36 | gr-demand-backup/9-delete-data | ++ wait_pod mysql-client
    logger.go:42: 02:12:36 | gr-demand-backup/9-delete-data | ++ local pod=mysql-client
    logger.go:42: 02:12:36 | gr-demand-backup/9-delete-data | ++ set +o xtrace
    logger.go:42: 02:12:36 | gr-demand-backup/9-delete-data | mysql-clienttrue
    logger.go:42: 02:12:36 | gr-demand-backup/9-delete-data | ++ kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:12:36 | gr-demand-backup/9-delete-data | ++ sed -e 's/mysql: //'
    logger.go:42: 02:12:36 | gr-demand-backup/9-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:12:38 | gr-demand-backup/9-delete-data | ++ :
    logger.go:42: 02:12:38 | gr-demand-backup/9-delete-data | + data=
    logger.go:42: 02:12:38 | gr-demand-backup/9-delete-data | + kubectl create configmap -n kuttl-test-ready-deer 08-delete-data-s3-2 --from-literal=data=
    logger.go:42: 02:12:38 | gr-demand-backup/9-delete-data | configmap/08-delete-data-s3-2 created
    logger.go:42: 02:12:39 | gr-demand-backup/9-delete-data | test step completed 9-delete-data
    logger.go:42: 02:12:39 | gr-demand-backup/10-restore-from-s3 | starting test step 10-restore-from-s3
    logger.go:42: 02:12:39 | gr-demand-backup/10-restore-from-s3 | PerconaServerMySQLRestore:kuttl-test-ready-deer/gr-demand-backup-restore-s3 created
    logger.go:42: 02:17:39 | gr-demand-backup/10-restore-from-s3 | test step completed 10-restore-from-s3
    logger.go:42: 02:17:39 | gr-demand-backup/11-read-data | starting test step 11-read-data
    logger.go:42: 02:17:39 | gr-demand-backup/11-read-data | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        cluster_name=$(get_cluster_name)
        for i in 0 1 2; do
            data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password")
            kubectl create configmap -n "${NAMESPACE}" 06-read-data-s3-${i} --from-literal=data="${data}"
        done]
    logger.go:42: 02:17:39 | gr-demand-backup/11-read-data | + source ../../functions
    logger.go:42: 02:17:39 | gr-demand-backup/11-read-data | +++ realpath ../../..
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | ++++ pwd
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/tests/gr-demand-backup
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | ++ test_name=gr-demand-backup
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/vars.sh
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ export GIT_BRANCH=PR-875
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ GIT_BRANCH=PR-875
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ export VERSION=PR-875-8c227d86
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ VERSION=PR-875-8c227d86
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ export PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ export CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | ++++ which gdate
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-875/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | ++++ which date
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ date=/usr/bin/date
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ oc get projects
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ :
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ kubectl get nodes
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ grep '^minikube'
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | ++ get_cluster_name
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | ++ kubectl -n kuttl-test-ready-deer get ps -o 'jsonpath={.items[0].metadata.name}'
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | + cluster_name=gr-demand-backup
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | + for i in 0 1 2
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | ++ local 'uri=-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | ++ local pod=
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ get_client_pod
    logger.go:42: 02:17:40 | gr-demand-backup/11-read-data | +++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:17:41 | gr-demand-backup/11-read-data | ++ client_pod=mysql-client
    logger.go:42: 02:17:41 | gr-demand-backup/11-read-data | ++ wait_pod mysql-client
    logger.go:42: 02:17:41 | gr-demand-backup/11-read-data | ++ local pod=mysql-client
    logger.go:42: 02:17:41 | gr-demand-backup/11-read-data | ++ set +o xtrace
    logger.go:42: 02:17:41 | gr-demand-backup/11-read-data | mysql-clienttrue
    logger.go:42: 02:17:41 | gr-demand-backup/11-read-data | ++ kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:17:41 | gr-demand-backup/11-read-data | ++ sed -e 's/mysql: //'
    logger.go:42: 02:17:41 | gr-demand-backup/11-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:17:43 | gr-demand-backup/11-read-data | + data=100500
    logger.go:42: 02:17:43 | gr-demand-backup/11-read-data | + kubectl create configmap -n kuttl-test-ready-deer 06-read-data-s3-0 --from-literal=data=100500
    logger.go:42: 02:17:43 | gr-demand-backup/11-read-data | configmap/06-read-data-s3-0 created
    logger.go:42: 02:17:43 | gr-demand-backup/11-read-data | + for i in 0 1 2
    logger.go:42: 02:17:43 | gr-demand-backup/11-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:17:43 | gr-demand-backup/11-read-data | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 02:17:43 | gr-demand-backup/11-read-data | ++ local 'uri=-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:17:43 | gr-demand-backup/11-read-data | ++ local pod=
    logger.go:42: 02:17:43 | gr-demand-backup/11-read-data | +++ get_client_pod
    logger.go:42: 02:17:43 | gr-demand-backup/11-read-data | +++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:17:43 | gr-demand-backup/11-read-data | ++ client_pod=mysql-client
    logger.go:42: 02:17:43 | gr-demand-backup/11-read-data | ++ wait_pod mysql-client
    logger.go:42: 02:17:43 | gr-demand-backup/11-read-data | ++ local pod=mysql-client
    logger.go:42: 02:17:43 | gr-demand-backup/11-read-data | ++ set +o xtrace
    logger.go:42: 02:17:44 | gr-demand-backup/11-read-data | mysql-clienttrue
    logger.go:42: 02:17:44 | gr-demand-backup/11-read-data | ++ kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:17:44 | gr-demand-backup/11-read-data | ++ sed -e 's/mysql: //'
    logger.go:42: 02:17:44 | gr-demand-backup/11-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:17:45 | gr-demand-backup/11-read-data | + data=100500
    logger.go:42: 02:17:45 | gr-demand-backup/11-read-data | + kubectl create configmap -n kuttl-test-ready-deer 06-read-data-s3-1 --from-literal=data=100500
    logger.go:42: 02:17:46 | gr-demand-backup/11-read-data | configmap/06-read-data-s3-1 created
    logger.go:42: 02:17:46 | gr-demand-backup/11-read-data | + for i in 0 1 2
    logger.go:42: 02:17:46 | gr-demand-backup/11-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:17:46 | gr-demand-backup/11-read-data | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 02:17:46 | gr-demand-backup/11-read-data | ++ local 'uri=-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:17:46 | gr-demand-backup/11-read-data | ++ local pod=
    logger.go:42: 02:17:46 | gr-demand-backup/11-read-data | +++ get_client_pod
    logger.go:42: 02:17:46 | gr-demand-backup/11-read-data | +++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:17:46 | gr-demand-backup/11-read-data | ++ client_pod=mysql-client
    logger.go:42: 02:17:46 | gr-demand-backup/11-read-data | ++ wait_pod mysql-client
    logger.go:42: 02:17:46 | gr-demand-backup/11-read-data | ++ local pod=mysql-client
    logger.go:42: 02:17:46 | gr-demand-backup/11-read-data | ++ set +o xtrace
    logger.go:42: 02:17:46 | gr-demand-backup/11-read-data | mysql-clienttrue
    logger.go:42: 02:17:46 | gr-demand-backup/11-read-data | ++ sed -e 's/mysql: //'
    logger.go:42: 02:17:46 | gr-demand-backup/11-read-data | ++ kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:17:46 | gr-demand-backup/11-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:17:48 | gr-demand-backup/11-read-data | + data=100500
    logger.go:42: 02:17:48 | gr-demand-backup/11-read-data | + kubectl create configmap -n kuttl-test-ready-deer 06-read-data-s3-2 --from-literal=data=100500
    logger.go:42: 02:17:48 | gr-demand-backup/11-read-data | configmap/06-read-data-s3-2 created
    logger.go:42: 02:17:49 | gr-demand-backup/11-read-data | test step completed 11-read-data
    logger.go:42: 02:17:49 | gr-demand-backup/12-create-backup-gcp | starting test step 12-create-backup-gcp
    logger.go:42: 02:17:50 | gr-demand-backup/12-create-backup-gcp | PerconaServerMySQLBackup:kuttl-test-ready-deer/gr-demand-backup-gcp created
    logger.go:42: 02:18:01 | gr-demand-backup/12-create-backup-gcp | test step completed 12-create-backup-gcp
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | starting test step 13-delete-data
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        run_mysql \
            "TRUNCATE TABLE myDB.myTable" \
            "-h $(get_mysql_router_service $(get_cluster_name)) -uroot -proot_password"
        
        cluster_name=$(get_cluster_name)
        for i in 0 1 2; do
            data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password")
            kubectl create configmap -n "${NAMESPACE}" 12-delete-data-gcp-${i} --from-literal=data="${data}"
        done]
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | + source ../../functions
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ realpath ../../..
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | ++++ pwd
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/tests/gr-demand-backup
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | ++ test_name=gr-demand-backup
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/vars.sh
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ export GIT_BRANCH=PR-875
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ GIT_BRANCH=PR-875
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ export VERSION=PR-875-8c227d86
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ VERSION=PR-875-8c227d86
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ export PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ export CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | ++++ which gdate
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-875/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | ++++ which date
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ date=/usr/bin/date
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ oc get projects
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ :
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ kubectl get nodes
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ grep '^minikube'
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ get_cluster_name
    logger.go:42: 02:18:01 | gr-demand-backup/13-delete-data | +++ kubectl -n kuttl-test-ready-deer get ps -o 'jsonpath={.items[0].metadata.name}'
    logger.go:42: 02:18:02 | gr-demand-backup/13-delete-data | ++ get_mysql_router_service gr-demand-backup
    logger.go:42: 02:18:02 | gr-demand-backup/13-delete-data | ++ local cluster=gr-demand-backup
    logger.go:42: 02:18:02 | gr-demand-backup/13-delete-data | ++ echo gr-demand-backup-router
    logger.go:42: 02:18:02 | gr-demand-backup/13-delete-data | + run_mysql 'TRUNCATE TABLE myDB.myTable' '-h gr-demand-backup-router -uroot -proot_password'
    logger.go:42: 02:18:02 | gr-demand-backup/13-delete-data | + local 'command=TRUNCATE TABLE myDB.myTable'
    logger.go:42: 02:18:02 | gr-demand-backup/13-delete-data | + local 'uri=-h gr-demand-backup-router -uroot -proot_password'
    logger.go:42: 02:18:02 | gr-demand-backup/13-delete-data | + local pod=
    logger.go:42: 02:18:02 | gr-demand-backup/13-delete-data | ++ get_client_pod
    logger.go:42: 02:18:02 | gr-demand-backup/13-delete-data | ++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:18:02 | gr-demand-backup/13-delete-data | + client_pod=mysql-client
    logger.go:42: 02:18:02 | gr-demand-backup/13-delete-data | + wait_pod mysql-client
    logger.go:42: 02:18:02 | gr-demand-backup/13-delete-data | + local pod=mysql-client
    logger.go:42: 02:18:02 | gr-demand-backup/13-delete-data | + set +o xtrace
    logger.go:42: 02:18:03 | gr-demand-backup/13-delete-data | mysql-clienttrue
    logger.go:42: 02:18:03 | gr-demand-backup/13-delete-data | + sed -e 's/mysql: //'
    logger.go:42: 02:18:03 | gr-demand-backup/13-delete-data | + grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:18:03 | gr-demand-backup/13-delete-data | + kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "TRUNCATE TABLE myDB.myTable" | mysql -sN -h gr-demand-backup-router -uroot -proot_password'
    logger.go:42: 02:18:05 | gr-demand-backup/13-delete-data | + :
    logger.go:42: 02:18:05 | gr-demand-backup/13-delete-data | ++ get_cluster_name
    logger.go:42: 02:18:05 | gr-demand-backup/13-delete-data | ++ kubectl -n kuttl-test-ready-deer get ps -o 'jsonpath={.items[0].metadata.name}'
    logger.go:42: 02:18:05 | gr-demand-backup/13-delete-data | + cluster_name=gr-demand-backup
    logger.go:42: 02:18:05 | gr-demand-backup/13-delete-data | + for i in 0 1 2
    logger.go:42: 02:18:05 | gr-demand-backup/13-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:18:05 | gr-demand-backup/13-delete-data | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 02:18:05 | gr-demand-backup/13-delete-data | ++ local 'uri=-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:18:05 | gr-demand-backup/13-delete-data | ++ local pod=
    logger.go:42: 02:18:05 | gr-demand-backup/13-delete-data | +++ get_client_pod
    logger.go:42: 02:18:05 | gr-demand-backup/13-delete-data | +++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:18:05 | gr-demand-backup/13-delete-data | ++ client_pod=mysql-client
    logger.go:42: 02:18:05 | gr-demand-backup/13-delete-data | ++ wait_pod mysql-client
    logger.go:42: 02:18:05 | gr-demand-backup/13-delete-data | ++ local pod=mysql-client
    logger.go:42: 02:18:05 | gr-demand-backup/13-delete-data | ++ set +o xtrace
    logger.go:42: 02:18:06 | gr-demand-backup/13-delete-data | mysql-clienttrue
    logger.go:42: 02:18:06 | gr-demand-backup/13-delete-data | ++ kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:18:06 | gr-demand-backup/13-delete-data | ++ sed -e 's/mysql: //'
    logger.go:42: 02:18:06 | gr-demand-backup/13-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:18:07 | gr-demand-backup/13-delete-data | ++ :
    logger.go:42: 02:18:07 | gr-demand-backup/13-delete-data | + data=
    logger.go:42: 02:18:07 | gr-demand-backup/13-delete-data | + kubectl create configmap -n kuttl-test-ready-deer 12-delete-data-gcp-0 --from-literal=data=
    logger.go:42: 02:18:08 | gr-demand-backup/13-delete-data | configmap/12-delete-data-gcp-0 created
    logger.go:42: 02:18:08 | gr-demand-backup/13-delete-data | + for i in 0 1 2
    logger.go:42: 02:18:08 | gr-demand-backup/13-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:18:08 | gr-demand-backup/13-delete-data | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 02:18:08 | gr-demand-backup/13-delete-data | ++ local 'uri=-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:18:08 | gr-demand-backup/13-delete-data | ++ local pod=
    logger.go:42: 02:18:08 | gr-demand-backup/13-delete-data | +++ get_client_pod
    logger.go:42: 02:18:08 | gr-demand-backup/13-delete-data | +++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:18:08 | gr-demand-backup/13-delete-data | ++ client_pod=mysql-client
    logger.go:42: 02:18:08 | gr-demand-backup/13-delete-data | ++ wait_pod mysql-client
    logger.go:42: 02:18:08 | gr-demand-backup/13-delete-data | ++ local pod=mysql-client
    logger.go:42: 02:18:08 | gr-demand-backup/13-delete-data | ++ set +o xtrace
    logger.go:42: 02:18:08 | gr-demand-backup/13-delete-data | mysql-clienttrue
    logger.go:42: 02:18:08 | gr-demand-backup/13-delete-data | ++ sed -e 's/mysql: //'
    logger.go:42: 02:18:08 | gr-demand-backup/13-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:18:08 | gr-demand-backup/13-delete-data | ++ kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:18:10 | gr-demand-backup/13-delete-data | ++ :
    logger.go:42: 02:18:10 | gr-demand-backup/13-delete-data | + data=
    logger.go:42: 02:18:10 | gr-demand-backup/13-delete-data | + kubectl create configmap -n kuttl-test-ready-deer 12-delete-data-gcp-1 --from-literal=data=
    logger.go:42: 02:18:10 | gr-demand-backup/13-delete-data | configmap/12-delete-data-gcp-1 created
    logger.go:42: 02:18:10 | gr-demand-backup/13-delete-data | + for i in 0 1 2
    logger.go:42: 02:18:10 | gr-demand-backup/13-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:18:10 | gr-demand-backup/13-delete-data | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 02:18:10 | gr-demand-backup/13-delete-data | ++ local 'uri=-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:18:10 | gr-demand-backup/13-delete-data | ++ local pod=
    logger.go:42: 02:18:10 | gr-demand-backup/13-delete-data | +++ get_client_pod
    logger.go:42: 02:18:10 | gr-demand-backup/13-delete-data | +++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:18:11 | gr-demand-backup/13-delete-data | ++ client_pod=mysql-client
    logger.go:42: 02:18:11 | gr-demand-backup/13-delete-data | ++ wait_pod mysql-client
    logger.go:42: 02:18:11 | gr-demand-backup/13-delete-data | ++ local pod=mysql-client
    logger.go:42: 02:18:11 | gr-demand-backup/13-delete-data | ++ set +o xtrace
    logger.go:42: 02:18:11 | gr-demand-backup/13-delete-data | mysql-clienttrue
    logger.go:42: 02:18:11 | gr-demand-backup/13-delete-data | ++ kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:18:11 | gr-demand-backup/13-delete-data | ++ sed -e 's/mysql: //'
    logger.go:42: 02:18:11 | gr-demand-backup/13-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:18:13 | gr-demand-backup/13-delete-data | ++ :
    logger.go:42: 02:18:13 | gr-demand-backup/13-delete-data | + data=
    logger.go:42: 02:18:13 | gr-demand-backup/13-delete-data | + kubectl create configmap -n kuttl-test-ready-deer 12-delete-data-gcp-2 --from-literal=data=
    logger.go:42: 02:18:13 | gr-demand-backup/13-delete-data | configmap/12-delete-data-gcp-2 created
    logger.go:42: 02:18:14 | gr-demand-backup/13-delete-data | test step completed 13-delete-data
    logger.go:42: 02:18:14 | gr-demand-backup/14-restore-from-gcp | starting test step 14-restore-from-gcp
    logger.go:42: 02:18:14 | gr-demand-backup/14-restore-from-gcp | PerconaServerMySQLRestore:kuttl-test-ready-deer/gr-demand-backup-restore-gcp created
    logger.go:42: 02:23:10 | gr-demand-backup/14-restore-from-gcp | test step completed 14-restore-from-gcp
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | starting test step 15-read-data
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        cluster_name=$(get_cluster_name)
        for i in 0 1 2; do
            data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password")
            kubectl create configmap -n "${NAMESPACE}" 06-read-data-gcp-${i} --from-literal=data="${data}"
        done]
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | + source ../../functions
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ realpath ../../..
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | ++++ pwd
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/tests/gr-demand-backup
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | ++ test_name=gr-demand-backup
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/vars.sh
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ export GIT_BRANCH=PR-875
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ GIT_BRANCH=PR-875
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ export VERSION=PR-875-8c227d86
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ VERSION=PR-875-8c227d86
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ export PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ export CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | ++++ which gdate
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-875/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | ++++ which date
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ date=/usr/bin/date
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ oc get projects
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ :
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ kubectl get nodes
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | +++ grep '^minikube'
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | ++ get_cluster_name
    logger.go:42: 02:23:10 | gr-demand-backup/15-read-data | ++ kubectl -n kuttl-test-ready-deer get ps -o 'jsonpath={.items[0].metadata.name}'
    logger.go:42: 02:23:11 | gr-demand-backup/15-read-data | + cluster_name=gr-demand-backup
    logger.go:42: 02:23:11 | gr-demand-backup/15-read-data | + for i in 0 1 2
    logger.go:42: 02:23:11 | gr-demand-backup/15-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:23:11 | gr-demand-backup/15-read-data | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 02:23:11 | gr-demand-backup/15-read-data | ++ local 'uri=-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:23:11 | gr-demand-backup/15-read-data | ++ local pod=
    logger.go:42: 02:23:11 | gr-demand-backup/15-read-data | +++ get_client_pod
    logger.go:42: 02:23:11 | gr-demand-backup/15-read-data | +++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:23:11 | gr-demand-backup/15-read-data | ++ client_pod=mysql-client
    logger.go:42: 02:23:11 | gr-demand-backup/15-read-data | ++ wait_pod mysql-client
    logger.go:42: 02:23:11 | gr-demand-backup/15-read-data | ++ local pod=mysql-client
    logger.go:42: 02:23:11 | gr-demand-backup/15-read-data | ++ set +o xtrace
    logger.go:42: 02:23:11 | gr-demand-backup/15-read-data | mysql-clienttrue
    logger.go:42: 02:23:11 | gr-demand-backup/15-read-data | ++ kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:23:11 | gr-demand-backup/15-read-data | ++ sed -e 's/mysql: //'
    logger.go:42: 02:23:11 | gr-demand-backup/15-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:23:13 | gr-demand-backup/15-read-data | + data=100500
    logger.go:42: 02:23:13 | gr-demand-backup/15-read-data | + kubectl create configmap -n kuttl-test-ready-deer 06-read-data-gcp-0 --from-literal=data=100500
    logger.go:42: 02:23:13 | gr-demand-backup/15-read-data | configmap/06-read-data-gcp-0 created
    logger.go:42: 02:23:13 | gr-demand-backup/15-read-data | + for i in 0 1 2
    logger.go:42: 02:23:13 | gr-demand-backup/15-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:23:13 | gr-demand-backup/15-read-data | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 02:23:13 | gr-demand-backup/15-read-data | ++ local 'uri=-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:23:13 | gr-demand-backup/15-read-data | ++ local pod=
    logger.go:42: 02:23:13 | gr-demand-backup/15-read-data | +++ get_client_pod
    logger.go:42: 02:23:13 | gr-demand-backup/15-read-data | +++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:23:14 | gr-demand-backup/15-read-data | ++ client_pod=mysql-client
    logger.go:42: 02:23:14 | gr-demand-backup/15-read-data | ++ wait_pod mysql-client
    logger.go:42: 02:23:14 | gr-demand-backup/15-read-data | ++ local pod=mysql-client
    logger.go:42: 02:23:14 | gr-demand-backup/15-read-data | ++ set +o xtrace
    logger.go:42: 02:23:14 | gr-demand-backup/15-read-data | mysql-clienttrue
    logger.go:42: 02:23:14 | gr-demand-backup/15-read-data | ++ kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:23:14 | gr-demand-backup/15-read-data | ++ sed -e 's/mysql: //'
    logger.go:42: 02:23:14 | gr-demand-backup/15-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:23:16 | gr-demand-backup/15-read-data | + data=100500
    logger.go:42: 02:23:16 | gr-demand-backup/15-read-data | + kubectl create configmap -n kuttl-test-ready-deer 06-read-data-gcp-1 --from-literal=data=100500
    logger.go:42: 02:23:16 | gr-demand-backup/15-read-data | configmap/06-read-data-gcp-1 created
    logger.go:42: 02:23:16 | gr-demand-backup/15-read-data | + for i in 0 1 2
    logger.go:42: 02:23:16 | gr-demand-backup/15-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:23:16 | gr-demand-backup/15-read-data | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 02:23:16 | gr-demand-backup/15-read-data | ++ local 'uri=-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:23:16 | gr-demand-backup/15-read-data | ++ local pod=
    logger.go:42: 02:23:16 | gr-demand-backup/15-read-data | +++ get_client_pod
    logger.go:42: 02:23:16 | gr-demand-backup/15-read-data | +++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:23:16 | gr-demand-backup/15-read-data | ++ client_pod=mysql-client
    logger.go:42: 02:23:16 | gr-demand-backup/15-read-data | ++ wait_pod mysql-client
    logger.go:42: 02:23:16 | gr-demand-backup/15-read-data | ++ local pod=mysql-client
    logger.go:42: 02:23:16 | gr-demand-backup/15-read-data | ++ set +o xtrace
    logger.go:42: 02:23:17 | gr-demand-backup/15-read-data | mysql-clienttrue
    logger.go:42: 02:23:17 | gr-demand-backup/15-read-data | ++ kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:23:17 | gr-demand-backup/15-read-data | ++ sed -e 's/mysql: //'
    logger.go:42: 02:23:17 | gr-demand-backup/15-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:23:18 | gr-demand-backup/15-read-data | + data=100500
    logger.go:42: 02:23:18 | gr-demand-backup/15-read-data | + kubectl create configmap -n kuttl-test-ready-deer 06-read-data-gcp-2 --from-literal=data=100500
    logger.go:42: 02:23:19 | gr-demand-backup/15-read-data | configmap/06-read-data-gcp-2 created
    logger.go:42: 02:23:20 | gr-demand-backup/15-read-data | test step completed 15-read-data
    logger.go:42: 02:23:20 | gr-demand-backup/16-create-backup-azure | starting test step 16-create-backup-azure
    logger.go:42: 02:23:20 | gr-demand-backup/16-create-backup-azure | PerconaServerMySQLBackup:kuttl-test-ready-deer/gr-demand-backup-azure created
    logger.go:42: 02:23:31 | gr-demand-backup/16-create-backup-azure | test step completed 16-create-backup-azure
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | starting test step 17-delete-data
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        run_mysql \
            "TRUNCATE TABLE myDB.myTable" \
            "-h $(get_mysql_router_service $(get_cluster_name)) -uroot -proot_password"
        
        cluster_name=$(get_cluster_name)
        for i in 0 1 2; do
            data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password")
            kubectl create configmap -n "${NAMESPACE}" 16-delete-data-azure-${i} --from-literal=data="${data}"
        done]
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | + source ../../functions
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ realpath ../../..
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | ++++ pwd
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/tests/gr-demand-backup
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | ++ test_name=gr-demand-backup
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/vars.sh
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ export GIT_BRANCH=PR-875
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ GIT_BRANCH=PR-875
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ export VERSION=PR-875-8c227d86
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ VERSION=PR-875-8c227d86
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ export PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ export CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | ++++ which gdate
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-875/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | ++++ which date
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ date=/usr/bin/date
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ oc get projects
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ :
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ kubectl get nodes
    logger.go:42: 02:23:31 | gr-demand-backup/17-delete-data | +++ grep '^minikube'
    logger.go:42: 02:23:32 | gr-demand-backup/17-delete-data | +++ get_cluster_name
    logger.go:42: 02:23:32 | gr-demand-backup/17-delete-data | +++ kubectl -n kuttl-test-ready-deer get ps -o 'jsonpath={.items[0].metadata.name}'
    logger.go:42: 02:23:32 | gr-demand-backup/17-delete-data | ++ get_mysql_router_service gr-demand-backup
    logger.go:42: 02:23:32 | gr-demand-backup/17-delete-data | ++ local cluster=gr-demand-backup
    logger.go:42: 02:23:32 | gr-demand-backup/17-delete-data | ++ echo gr-demand-backup-router
    logger.go:42: 02:23:32 | gr-demand-backup/17-delete-data | + run_mysql 'TRUNCATE TABLE myDB.myTable' '-h gr-demand-backup-router -uroot -proot_password'
    logger.go:42: 02:23:32 | gr-demand-backup/17-delete-data | + local 'command=TRUNCATE TABLE myDB.myTable'
    logger.go:42: 02:23:32 | gr-demand-backup/17-delete-data | + local 'uri=-h gr-demand-backup-router -uroot -proot_password'
    logger.go:42: 02:23:32 | gr-demand-backup/17-delete-data | + local pod=
    logger.go:42: 02:23:32 | gr-demand-backup/17-delete-data | ++ get_client_pod
    logger.go:42: 02:23:32 | gr-demand-backup/17-delete-data | ++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:23:33 | gr-demand-backup/17-delete-data | + client_pod=mysql-client
    logger.go:42: 02:23:33 | gr-demand-backup/17-delete-data | + wait_pod mysql-client
    logger.go:42: 02:23:33 | gr-demand-backup/17-delete-data | + local pod=mysql-client
    logger.go:42: 02:23:33 | gr-demand-backup/17-delete-data | + set +o xtrace
    logger.go:42: 02:23:33 | gr-demand-backup/17-delete-data | mysql-clienttrue
    logger.go:42: 02:23:33 | gr-demand-backup/17-delete-data | + kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "TRUNCATE TABLE myDB.myTable" | mysql -sN -h gr-demand-backup-router -uroot -proot_password'
    logger.go:42: 02:23:33 | gr-demand-backup/17-delete-data | + sed -e 's/mysql: //'
    logger.go:42: 02:23:33 | gr-demand-backup/17-delete-data | + grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:23:35 | gr-demand-backup/17-delete-data | + :
    logger.go:42: 02:23:35 | gr-demand-backup/17-delete-data | ++ get_cluster_name
    logger.go:42: 02:23:35 | gr-demand-backup/17-delete-data | ++ kubectl -n kuttl-test-ready-deer get ps -o 'jsonpath={.items[0].metadata.name}'
    logger.go:42: 02:23:36 | gr-demand-backup/17-delete-data | + cluster_name=gr-demand-backup
    logger.go:42: 02:23:36 | gr-demand-backup/17-delete-data | + for i in 0 1 2
    logger.go:42: 02:23:36 | gr-demand-backup/17-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:23:36 | gr-demand-backup/17-delete-data | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 02:23:36 | gr-demand-backup/17-delete-data | ++ local 'uri=-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:23:36 | gr-demand-backup/17-delete-data | ++ local pod=
    logger.go:42: 02:23:36 | gr-demand-backup/17-delete-data | +++ get_client_pod
    logger.go:42: 02:23:36 | gr-demand-backup/17-delete-data | +++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:23:36 | gr-demand-backup/17-delete-data | ++ client_pod=mysql-client
    logger.go:42: 02:23:36 | gr-demand-backup/17-delete-data | ++ wait_pod mysql-client
    logger.go:42: 02:23:36 | gr-demand-backup/17-delete-data | ++ local pod=mysql-client
    logger.go:42: 02:23:36 | gr-demand-backup/17-delete-data | ++ set +o xtrace
    logger.go:42: 02:23:36 | gr-demand-backup/17-delete-data | mysql-clienttrue
    logger.go:42: 02:23:36 | gr-demand-backup/17-delete-data | ++ kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:23:36 | gr-demand-backup/17-delete-data | ++ sed -e 's/mysql: //'
    logger.go:42: 02:23:36 | gr-demand-backup/17-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:23:38 | gr-demand-backup/17-delete-data | ++ :
    logger.go:42: 02:23:38 | gr-demand-backup/17-delete-data | + data=
    logger.go:42: 02:23:38 | gr-demand-backup/17-delete-data | + kubectl create configmap -n kuttl-test-ready-deer 16-delete-data-azure-0 --from-literal=data=
    logger.go:42: 02:23:38 | gr-demand-backup/17-delete-data | configmap/16-delete-data-azure-0 created
    logger.go:42: 02:23:38 | gr-demand-backup/17-delete-data | + for i in 0 1 2
    logger.go:42: 02:23:38 | gr-demand-backup/17-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:23:38 | gr-demand-backup/17-delete-data | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 02:23:38 | gr-demand-backup/17-delete-data | ++ local 'uri=-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:23:38 | gr-demand-backup/17-delete-data | ++ local pod=
    logger.go:42: 02:23:38 | gr-demand-backup/17-delete-data | +++ get_client_pod
    logger.go:42: 02:23:38 | gr-demand-backup/17-delete-data | +++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:23:39 | gr-demand-backup/17-delete-data | ++ client_pod=mysql-client
    logger.go:42: 02:23:39 | gr-demand-backup/17-delete-data | ++ wait_pod mysql-client
    logger.go:42: 02:23:39 | gr-demand-backup/17-delete-data | ++ local pod=mysql-client
    logger.go:42: 02:23:39 | gr-demand-backup/17-delete-data | ++ set +o xtrace
    logger.go:42: 02:23:39 | gr-demand-backup/17-delete-data | mysql-clienttrue
    logger.go:42: 02:23:39 | gr-demand-backup/17-delete-data | ++ sed -e 's/mysql: //'
    logger.go:42: 02:23:39 | gr-demand-backup/17-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:23:39 | gr-demand-backup/17-delete-data | ++ kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:23:40 | gr-demand-backup/17-delete-data | ++ :
    logger.go:42: 02:23:40 | gr-demand-backup/17-delete-data | + data=
    logger.go:42: 02:23:40 | gr-demand-backup/17-delete-data | + kubectl create configmap -n kuttl-test-ready-deer 16-delete-data-azure-1 --from-literal=data=
    logger.go:42: 02:23:41 | gr-demand-backup/17-delete-data | configmap/16-delete-data-azure-1 created
    logger.go:42: 02:23:41 | gr-demand-backup/17-delete-data | + for i in 0 1 2
    logger.go:42: 02:23:41 | gr-demand-backup/17-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:23:41 | gr-demand-backup/17-delete-data | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 02:23:41 | gr-demand-backup/17-delete-data | ++ local 'uri=-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:23:41 | gr-demand-backup/17-delete-data | ++ local pod=
    logger.go:42: 02:23:41 | gr-demand-backup/17-delete-data | +++ get_client_pod
    logger.go:42: 02:23:41 | gr-demand-backup/17-delete-data | +++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:23:41 | gr-demand-backup/17-delete-data | ++ client_pod=mysql-client
    logger.go:42: 02:23:41 | gr-demand-backup/17-delete-data | ++ wait_pod mysql-client
    logger.go:42: 02:23:41 | gr-demand-backup/17-delete-data | ++ local pod=mysql-client
    logger.go:42: 02:23:41 | gr-demand-backup/17-delete-data | ++ set +o xtrace
    logger.go:42: 02:23:42 | gr-demand-backup/17-delete-data | mysql-clienttrue
    logger.go:42: 02:23:42 | gr-demand-backup/17-delete-data | ++ kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:23:42 | gr-demand-backup/17-delete-data | ++ sed -e 's/mysql: //'
    logger.go:42: 02:23:42 | gr-demand-backup/17-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:23:43 | gr-demand-backup/17-delete-data | ++ :
    logger.go:42: 02:23:43 | gr-demand-backup/17-delete-data | + data=
    logger.go:42: 02:23:43 | gr-demand-backup/17-delete-data | + kubectl create configmap -n kuttl-test-ready-deer 16-delete-data-azure-2 --from-literal=data=
    logger.go:42: 02:23:44 | gr-demand-backup/17-delete-data | configmap/16-delete-data-azure-2 created
    logger.go:42: 02:23:44 | gr-demand-backup/17-delete-data | test step completed 17-delete-data
    logger.go:42: 02:23:44 | gr-demand-backup/18-restore-from-azure | starting test step 18-restore-from-azure
    logger.go:42: 02:23:45 | gr-demand-backup/18-restore-from-azure | PerconaServerMySQLRestore:kuttl-test-ready-deer/gr-demand-backup-restore-azure created
    logger.go:42: 02:28:39 | gr-demand-backup/18-restore-from-azure | test step completed 18-restore-from-azure
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | starting test step 19-read-data
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        cluster_name=$(get_cluster_name)
        for i in 0 1 2; do
            data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password")
            kubectl create configmap -n "${NAMESPACE}" 06-read-data-azure-${i} --from-literal=data="${data}"
        done]
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | + source ../../functions
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ realpath ../../..
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | ++++ pwd
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/tests/gr-demand-backup
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | ++ test_name=gr-demand-backup
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/vars.sh
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ export GIT_BRANCH=PR-875
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ GIT_BRANCH=PR-875
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ export VERSION=PR-875-8c227d86
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ VERSION=PR-875-8c227d86
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ export PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ export CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | ++++ which gdate
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-875/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | ++++ which date
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ date=/usr/bin/date
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ oc get projects
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ :
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ kubectl get nodes
    logger.go:42: 02:28:39 | gr-demand-backup/19-read-data | +++ grep '^minikube'
    logger.go:42: 02:28:40 | gr-demand-backup/19-read-data | ++ get_cluster_name
    logger.go:42: 02:28:40 | gr-demand-backup/19-read-data | ++ kubectl -n kuttl-test-ready-deer get ps -o 'jsonpath={.items[0].metadata.name}'
    logger.go:42: 02:28:40 | gr-demand-backup/19-read-data | + cluster_name=gr-demand-backup
    logger.go:42: 02:28:40 | gr-demand-backup/19-read-data | + for i in 0 1 2
    logger.go:42: 02:28:40 | gr-demand-backup/19-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:28:40 | gr-demand-backup/19-read-data | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 02:28:40 | gr-demand-backup/19-read-data | ++ local 'uri=-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:28:40 | gr-demand-backup/19-read-data | ++ local pod=
    logger.go:42: 02:28:40 | gr-demand-backup/19-read-data | +++ get_client_pod
    logger.go:42: 02:28:40 | gr-demand-backup/19-read-data | +++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:28:41 | gr-demand-backup/19-read-data | ++ client_pod=mysql-client
    logger.go:42: 02:28:41 | gr-demand-backup/19-read-data | ++ wait_pod mysql-client
    logger.go:42: 02:28:41 | gr-demand-backup/19-read-data | ++ local pod=mysql-client
    logger.go:42: 02:28:41 | gr-demand-backup/19-read-data | ++ set +o xtrace
    logger.go:42: 02:28:41 | gr-demand-backup/19-read-data | mysql-clienttrue
    logger.go:42: 02:28:41 | gr-demand-backup/19-read-data | ++ sed -e 's/mysql: //'
    logger.go:42: 02:28:41 | gr-demand-backup/19-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:28:41 | gr-demand-backup/19-read-data | ++ kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:28:43 | gr-demand-backup/19-read-data | + data=100500
    logger.go:42: 02:28:43 | gr-demand-backup/19-read-data | + kubectl create configmap -n kuttl-test-ready-deer 06-read-data-azure-0 --from-literal=data=100500
    logger.go:42: 02:28:43 | gr-demand-backup/19-read-data | configmap/06-read-data-azure-0 created
    logger.go:42: 02:28:43 | gr-demand-backup/19-read-data | + for i in 0 1 2
    logger.go:42: 02:28:43 | gr-demand-backup/19-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:28:43 | gr-demand-backup/19-read-data | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 02:28:43 | gr-demand-backup/19-read-data | ++ local 'uri=-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:28:43 | gr-demand-backup/19-read-data | ++ local pod=
    logger.go:42: 02:28:43 | gr-demand-backup/19-read-data | +++ get_client_pod
    logger.go:42: 02:28:43 | gr-demand-backup/19-read-data | +++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:28:43 | gr-demand-backup/19-read-data | ++ client_pod=mysql-client
    logger.go:42: 02:28:43 | gr-demand-backup/19-read-data | ++ wait_pod mysql-client
    logger.go:42: 02:28:43 | gr-demand-backup/19-read-data | ++ local pod=mysql-client
    logger.go:42: 02:28:43 | gr-demand-backup/19-read-data | ++ set +o xtrace
    logger.go:42: 02:28:44 | gr-demand-backup/19-read-data | mysql-clienttrue
    logger.go:42: 02:28:44 | gr-demand-backup/19-read-data | ++ kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:28:44 | gr-demand-backup/19-read-data | ++ sed -e 's/mysql: //'
    logger.go:42: 02:28:44 | gr-demand-backup/19-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:28:45 | gr-demand-backup/19-read-data | + data=100500
    logger.go:42: 02:28:45 | gr-demand-backup/19-read-data | + kubectl create configmap -n kuttl-test-ready-deer 06-read-data-azure-1 --from-literal=data=100500
    logger.go:42: 02:28:46 | gr-demand-backup/19-read-data | configmap/06-read-data-azure-1 created
    logger.go:42: 02:28:46 | gr-demand-backup/19-read-data | + for i in 0 1 2
    logger.go:42: 02:28:46 | gr-demand-backup/19-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:28:46 | gr-demand-backup/19-read-data | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 02:28:46 | gr-demand-backup/19-read-data | ++ local 'uri=-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:28:46 | gr-demand-backup/19-read-data | ++ local pod=
    logger.go:42: 02:28:46 | gr-demand-backup/19-read-data | +++ get_client_pod
    logger.go:42: 02:28:46 | gr-demand-backup/19-read-data | +++ kubectl -n kuttl-test-ready-deer get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:28:46 | gr-demand-backup/19-read-data | ++ client_pod=mysql-client
    logger.go:42: 02:28:46 | gr-demand-backup/19-read-data | ++ wait_pod mysql-client
    logger.go:42: 02:28:46 | gr-demand-backup/19-read-data | ++ local pod=mysql-client
    logger.go:42: 02:28:46 | gr-demand-backup/19-read-data | ++ set +o xtrace
    logger.go:42: 02:28:46 | gr-demand-backup/19-read-data | mysql-clienttrue
    logger.go:42: 02:28:46 | gr-demand-backup/19-read-data | ++ kubectl -n kuttl-test-ready-deer exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password'
    logger.go:42: 02:28:46 | gr-demand-backup/19-read-data | ++ sed -e 's/mysql: //'
    logger.go:42: 02:28:46 | gr-demand-backup/19-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:28:48 | gr-demand-backup/19-read-data | + data=100500
    logger.go:42: 02:28:48 | gr-demand-backup/19-read-data | + kubectl create configmap -n kuttl-test-ready-deer 06-read-data-azure-2 --from-literal=data=100500
    logger.go:42: 02:28:48 | gr-demand-backup/19-read-data | configmap/06-read-data-azure-2 created
    logger.go:42: 02:28:49 | gr-demand-backup/19-read-data | test step completed 19-read-data
    logger.go:42: 02:28:49 | gr-demand-backup/98-drop-finalizer | starting test step 98-drop-finalizer
    logger.go:42: 02:28:49 | gr-demand-backup/98-drop-finalizer | PerconaServerMySQL:kuttl-test-ready-deer/gr-demand-backup updated
    logger.go:42: 02:28:49 | gr-demand-backup/98-drop-finalizer | test step completed 98-drop-finalizer
    logger.go:42: 02:28:49 | gr-demand-backup/99-remove-cluster-gracefully | starting test step 99-remove-cluster-gracefully
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        destroy_operator]
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | + source ../../functions
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ realpath ../../..
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | ++++ pwd
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/tests/gr-demand-backup
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | ++ test_name=gr-demand-backup
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/vars.sh
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-875
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/deploy
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-875/e2e-tests/conf
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ export GIT_BRANCH=PR-875
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ GIT_BRANCH=PR-875
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ export VERSION=PR-875-8c227d86
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ VERSION=PR-875-8c227d86
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-875-8c227d86
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ export PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ export CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | ++++ which gdate
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-875/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | ++++ which date
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ date=/usr/bin/date
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ oc get projects
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ :
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ kubectl get nodes
    logger.go:42: 02:28:50 | gr-demand-backup/99-remove-cluster-gracefully | +++ grep '^minikube'
    logger.go:42: 02:28:51 | gr-demand-backup/99-remove-cluster-gracefully | + destroy_operator
    logger.go:42: 02:28:51 | gr-demand-backup/99-remove-cluster-gracefully | + kubectl -n ps-operator delete deployment percona-server-mysql-operator --force --grace-period=0
    logger.go:42: 02:28:51 | gr-demand-backup/99-remove-cluster-gracefully | Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.
    logger.go:42: 02:28:51 | gr-demand-backup/99-remove-cluster-gracefully | deployment.apps "percona-server-mysql-operator" force deleted
    logger.go:42: 02:28:51 | gr-demand-backup/99-remove-cluster-gracefully | + [[ -n ps-operator ]]
    logger.go:42: 02:28:51 | gr-demand-backup/99-remove-cluster-gracefully | + kubectl delete namespace ps-operator --force --grace-period=0
    logger.go:42: 02:28:51 | gr-demand-backup/99-remove-cluster-gracefully | Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.
    logger.go:42: 02:28:52 | gr-demand-backup/99-remove-cluster-gracefully | namespace "ps-operator" force deleted
    logger.go:42: 02:28:57 | gr-demand-backup/99-remove-cluster-gracefully | test step completed 99-remove-cluster-gracefully
    logger.go:42: 02:28:58 | gr-demand-backup | gr-demand-backup events from ns kuttl-test-ready-deer:
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:06 +0000 UTC	Normal	Pod mysql-client	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/mysql-client to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3wfl	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:06 +0000 UTC	Normal	Pod mysql-client.spec.containers{mysql-client}		Pulling	Pulling image "percona/percona-server:8.0.33"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:14 +0000 UTC	Normal	ReplicaSet.apps minio-service-847fc8bb8d		SuccessfulCreate	Created pod: minio-service-847fc8bb8d-5b565	replicaset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:14 +0000 UTC	Normal	PersistentVolumeClaim minio-service		WaitForFirstConsumer	waiting for first consumer to be created before binding	persistentvolume-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:14 +0000 UTC	Normal	Deployment.apps minio-service		ScalingReplicaSet	Scaled up replica set minio-service-847fc8bb8d to 1	deployment-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:14 +0000 UTC	Normal	PersistentVolumeClaim minio-service		ExternalProvisioning	Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.	persistentvolume-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:15 +0000 UTC	Normal	Pod minio-service-post-job-npbhc	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/minio-service-post-job-npbhc to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3wfl	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:15 +0000 UTC	Normal	Job.batch minio-service-post-job		SuccessfulCreate	Created pod: minio-service-post-job-npbhc	job-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:15 +0000 UTC	Normal	PersistentVolumeClaim minio-service		Provisioning	External provisioner is provisioning volume for claim "kuttl-test-ready-deer/minio-service"	pd.csi.storage.gke.io_gke-90e882eae5d44a1387ff-3819-ecb5-vm_9c1755ac-9c8b-4676-a238-b44d38aab72b	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:18 +0000 UTC	Normal	Pod minio-service-847fc8bb8d-5b565	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/minio-service-847fc8bb8d-5b565 to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3wfl	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:18 +0000 UTC	Normal	PersistentVolumeClaim minio-service		ProvisioningSucceeded	Successfully provisioned volume pvc-b02e9ad9-20bd-433a-92d5-3547fc227fff	pd.csi.storage.gke.io_gke-90e882eae5d44a1387ff-3819-ecb5-vm_9c1755ac-9c8b-4676-a238-b44d38aab72b	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:20 +0000 UTC	Normal	Pod minio-service-post-job-npbhc.spec.containers{minio-make-user}		Pulling	Pulling image "quay.io/minio/mc:RELEASE.2023-09-29T16-41-22Z"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:26 +0000 UTC	Normal	Pod minio-service-847fc8bb8d-5b565		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-b02e9ad9-20bd-433a-92d5-3547fc227fff" 	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:27 +0000 UTC	Normal	Pod minio-service-847fc8bb8d-5b565.spec.containers{minio}		Pulling	Pulling image "quay.io/minio/minio:RELEASE.2023-09-30T07-02-29Z"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:28 +0000 UTC	Normal	Pod minio-service-post-job-npbhc.spec.containers{minio-make-user}		Pulled	Successfully pulled image "quay.io/minio/mc:RELEASE.2023-09-29T16-41-22Z" in 8.791s (8.791s including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:29 +0000 UTC	Normal	Pod minio-service-post-job-npbhc.spec.containers{minio-make-user}		Created	Created container: minio-make-user	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:29 +0000 UTC	Normal	Pod minio-service-post-job-npbhc.spec.containers{minio-make-user}		Started	Started container minio-make-user	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:33 +0000 UTC	Normal	Pod minio-service-847fc8bb8d-5b565.spec.containers{minio}		Pulled	Successfully pulled image "quay.io/minio/minio:RELEASE.2023-09-30T07-02-29Z" in 5.914s (5.914s including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:33 +0000 UTC	Normal	Pod minio-service-847fc8bb8d-5b565.spec.containers{minio}		Created	Created container: minio	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:33 +0000 UTC	Normal	Pod mysql-client.spec.containers{mysql-client}		Pulled	Successfully pulled image "percona/percona-server:8.0.33" in 26.587s (26.587s including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:33 +0000 UTC	Normal	Pod mysql-client.spec.containers{mysql-client}		Created	Created container: mysql-client	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:33 +0000 UTC	Normal	Pod mysql-client.spec.containers{mysql-client}		Started	Started container mysql-client	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:34 +0000 UTC	Normal	Pod minio-service-847fc8bb8d-5b565.spec.containers{minio}		Started	Started container minio	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:45 +0000 UTC	Normal	Job.batch minio-service-post-job		Completed	Job completed	job-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:47 +0000 UTC	Normal	Pod aws-cli	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/aws-cli to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3wfl	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:47 +0000 UTC	Normal	Pod aws-cli.spec.containers{aws-cli}		Pulling	Pulling image "perconalab/awscli"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:51 +0000 UTC	Normal	Pod aws-cli.spec.containers{aws-cli}		Pulled	Successfully pulled image "perconalab/awscli" in 3.301s (3.301s including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:51 +0000 UTC	Normal	Pod aws-cli.spec.containers{aws-cli}		Created	Created container: aws-cli	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:51 +0000 UTC	Normal	Pod aws-cli.spec.containers{aws-cli}		Started	Started container aws-cli	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:58 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-0		WaitForFirstConsumer	waiting for first consumer to be created before binding	persistentvolume-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:58 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-0		ExternalProvisioning	Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.	persistentvolume-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:58 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-0		Provisioning	External provisioner is provisioning volume for claim "kuttl-test-ready-deer/datadir-gr-demand-backup-mysql-0"	pd.csi.storage.gke.io_gke-90e882eae5d44a1387ff-3819-ecb5-vm_9c1755ac-9c8b-4676-a238-b44d38aab72b	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:58 +0000 UTC	Normal	StatefulSet.apps gr-demand-backup-mysql		SuccessfulCreate	create Claim datadir-gr-demand-backup-mysql-0 Pod gr-demand-backup-mysql-0 in StatefulSet gr-demand-backup-mysql success	statefulset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:00:58 +0000 UTC	Normal	StatefulSet.apps gr-demand-backup-mysql		SuccessfulCreate	create Pod gr-demand-backup-mysql-0 in StatefulSet gr-demand-backup-mysql successful	statefulset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:01:02 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-0		ProvisioningSucceeded	Successfully provisioned volume pvc-fc9b3101-02dd-4334-af00-d8e38f4fe87c	pd.csi.storage.gke.io_gke-90e882eae5d44a1387ff-3819-ecb5-vm_9c1755ac-9c8b-4676-a238-b44d38aab72b	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:01:02 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-mysql-0 to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3z29	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:01:10 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-fc9b3101-02dd-4334-af00-d8e38f4fe87c" 	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:01:11 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:01:15 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 4.29s (4.29s including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:01:15 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init}		Created	Created container: mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:01:15 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init}		Started	Started container mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:01:19 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-psmysql"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:01:41 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 21.849s (21.849s including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:01:41 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Created	Created container: mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:01:41 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Started	Started container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:01:41 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:01:56 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 15.266s (15.266s including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:01:56 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:01:56 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:02:17 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-1		WaitForFirstConsumer	waiting for first consumer to be created before binding	persistentvolume-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:02:17 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-1		ExternalProvisioning	Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.	persistentvolume-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:02:17 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-1		Provisioning	External provisioner is provisioning volume for claim "kuttl-test-ready-deer/datadir-gr-demand-backup-mysql-1"	pd.csi.storage.gke.io_gke-90e882eae5d44a1387ff-3819-ecb5-vm_9c1755ac-9c8b-4676-a238-b44d38aab72b	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:02:17 +0000 UTC	Normal	StatefulSet.apps gr-demand-backup-mysql		SuccessfulCreate	create Claim datadir-gr-demand-backup-mysql-1 Pod gr-demand-backup-mysql-1 in StatefulSet gr-demand-backup-mysql success	statefulset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:02:17 +0000 UTC	Normal	StatefulSet.apps gr-demand-backup-mysql		SuccessfulCreate	create Pod gr-demand-backup-mysql-1 in StatefulSet gr-demand-backup-mysql successful	statefulset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:02:21 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-1		ProvisioningSucceeded	Successfully provisioned volume pvc-7722f153-021a-401e-88a9-70c6cbae3e29	pd.csi.storage.gke.io_gke-90e882eae5d44a1387ff-3819-ecb5-vm_9c1755ac-9c8b-4676-a238-b44d38aab72b	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:02:21 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-mysql-1 to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3wfl	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:02:29 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-7722f153-021a-401e-88a9-70c6cbae3e29" 	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:02:30 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:02:30 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 140ms (140ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:02:30 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init}		Created	Created container: mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:02:30 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init}		Started	Started container mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:02:32 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-psmysql"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:02:54 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 21.909s (21.909s including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:02:54 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Created	Created container: mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:02:54 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Started	Started container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:02:54 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:03:11 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 16.384s (16.384s including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:03:11 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:03:11 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:03:36 +0000 UTC	Warning	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Unhealthy	Startup probe failed: 2025/03/26 02:03:20 Waiting for MySQL ready state
        2025/03/26 02:03:20 MySQL is ready
        2025/03/26 02:03:20 Bootstrap starting...
        2025/03/26 02:03:20 Running dba.configureLocalInstance('operator:*****@gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer', {'clearReadOnly': true})
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        WARNING: The clearReadOnly option is deprecated and will be removed in a future release.
        
        WARNING: This function is deprecated and will be removed in a future release of MySQL Shell, use dba.configureInstance() instead.
        Configuring local MySQL instance listening at port 3306 for use in an InnoDB cluster...
        
        This instance reports its own address as gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        
        applierWorkerThreads will be set to the default value of 4.
        
        NOTE: Some configuration options need to be fixed:
        
        
        +----------------------------------------+---------------+----------------+----------------------------+
        
        +----------------------------------------+---------------+----------------+----------------------------+
        | Variable                               | Current Value | Required Value | Note                       |+----------------------------------------+---------------+----------------+----------------------------+
        | binlog_transaction_dependency_tracking | COMMIT_ORDER  | WRITESET       | Update the server variable |Disabled super_read_only on the instance 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'
        Enabling super_read_only on the instance 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'
        Configuring instance...
        
        WARNING:*****@binlog_transaction_dependency_tracking' is deprecated and will be removed in a future release. (Code 1287).
        The instance 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' was configured to be used in an InnoDB cluster.
        2025/03/26 02:03:20 Instance (gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer) configured to join to the InnoDB cluster
        2025/03/26 02:03:20 peers: [gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer]
        2025/03/26 02:03:20 Running dba.getCluster('grdemandbackup')
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        2025/03/26 02:03:21 Connected to peer gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer
        2025/03/26 02:03:21 Cluster status:
        
        ClusterName: grdemandbackup
        Status: OK_NO_TOLERANCE
        StatusText: Cluster is NOT tolerant to any failures.
        SSL: REQUIRED
        Primary: gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        Topology:
        	
        	Member 0
        	Address: gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        	State: ONLINE
        	Errors: []
        
        		
        2025/03/26 02:03:21 Adding instance (gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer) to InnoDB cluster
        2025/03/26 02:03:21 Running dba.getCluster('grdemandbackup').addInstance('operator:*****@gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer', {'recoveryMethod': 'clone', 'waitRecovery': 3})
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        WARNING: The waitRecovery option is deprecated. Please use the recoveryProgress option instead.
        
        
        NOTE: The target instance 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' has not been pre-provisioned (GTID set is empty). The Shell is unable to decide whether incremental state recovery can correctly provision it.
        
        Clone based recovery selected through the recoveryMethod option
        
        Validating instance configuration at gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306...
        
        This instance reports its own address as gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        
        Instance configuration is suitable.
        NOTE: Group Replication will communicate with other members using 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'. Use the localAddress option to override.
        
        * Checking connectivity and SSL configuration...
        A new instance will be added to the InnoDB Cluster. Depending on the amount of
        data on the cluster this might take from a few seconds to several hours.
        
        Adding instance to the cluster...
        
        Monitoring recovery process of the new cluster member. Press ^C to stop monitoring and let it continue in background.
        Clone based state recovery is now in progress.
        
        NOTE: A server restart is expected to happen as part of the clone process. If the
        server does not support the RESTART command or does not come back after a
        while, you may need to manually start it back.
        
        * Waiting for clone to finish...
        NOTE: gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306 is being cloned from gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        ** Stage DROP DATA: Completed
        ** Clone Transfer      FILE COPY    100%  Completed    PAGE COPY    100%  Completed    REDO COPY    100%  Completed
        NOTE: gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306 is shutting down...
        
        * Waiting for server restart... \ 
* Waiting for server restart... | 
* Waiting for server restart... / 
* Waiting for server restart... - 
	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:03:36 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Killing	Container mysql failed startup probe, will be restarted	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:03:36 +0000 UTC	Warning	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		FailedPreStopHook	PreStopHook failed	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:03:36 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 109ms (109ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:04:10 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-2		WaitForFirstConsumer	waiting for first consumer to be created before binding	persistentvolume-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:04:10 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-2		ExternalProvisioning	Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.	persistentvolume-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:04:10 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-2		Provisioning	External provisioner is provisioning volume for claim "kuttl-test-ready-deer/datadir-gr-demand-backup-mysql-2"	pd.csi.storage.gke.io_gke-90e882eae5d44a1387ff-3819-ecb5-vm_9c1755ac-9c8b-4676-a238-b44d38aab72b	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:04:10 +0000 UTC	Normal	StatefulSet.apps gr-demand-backup-mysql		SuccessfulCreate	create Claim datadir-gr-demand-backup-mysql-2 Pod gr-demand-backup-mysql-2 in StatefulSet gr-demand-backup-mysql success	statefulset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:04:10 +0000 UTC	Normal	StatefulSet.apps gr-demand-backup-mysql		SuccessfulCreate	create Pod gr-demand-backup-mysql-2 in StatefulSet gr-demand-backup-mysql successful	statefulset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:04:14 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-2		ProvisioningSucceeded	Successfully provisioned volume pvc-0e933d8f-98a6-4d4e-b228-a449e4a42ce6	pd.csi.storage.gke.io_gke-90e882eae5d44a1387ff-3819-ecb5-vm_9c1755ac-9c8b-4676-a238-b44d38aab72b	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:04:14 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-mysql-2 to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-gs77	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:04:22 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-0e933d8f-98a6-4d4e-b228-a449e4a42ce6" 	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:04:23 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:04:27 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 4.315s (4.315s including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:04:27 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init}		Created	Created container: mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:04:27 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init}		Started	Started container mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:04:31 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-psmysql"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:04:53 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 21.636s (21.636s including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:04:53 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Created	Created container: mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:04:53 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Started	Started container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:04:53 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:05:08 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 15.556s (15.556s including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:05:08 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:05:08 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:05:32 +0000 UTC	Warning	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Unhealthy	Startup probe failed: 2025/03/26 02:05:13 Waiting for bootstrap.lock to be deleted
        2025/03/26 02:05:17 Waiting for MySQL ready state
        2025/03/26 02:05:17 MySQL is ready
        2025/03/26 02:05:17 Bootstrap starting...
        2025/03/26 02:05:17 Running dba.configureLocalInstance('operator:*****@gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer', {'clearReadOnly': true})
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        WARNING: The clearReadOnly option is deprecated and will be removed in a future release.
        
        WARNING: This function is deprecated and will be removed in a future release of MySQL Shell, use dba.configureInstance() instead.
        Configuring local MySQL instance listening at port 3306 for use in an InnoDB cluster...
        
        This instance reports its own address as gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        
        applierWorkerThreads will be set to the default value of 4.
        
        NOTE: Some configuration options need to be fixed:
        
        
        +----------------------------------------+---------------+----------------+----------------------------+
        
        +----------------------------------------+---------------+----------------+----------------------------+
        | Variable                               | Current Value | Required Value | Note                       |+----------------------------------------+---------------+----------------+----------------------------+
        | binlog_transaction_dependency_tracking | COMMIT_ORDER  | WRITESET       | Update the server variable |Disabled super_read_only on the instance 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'
        Enabling super_read_only on the instance 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'
        Configuring instance...
        
        WARNING:*****@binlog_transaction_dependency_tracking' is deprecated and will be removed in a future release. (Code 1287).
        The instance 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' was configured to be used in an InnoDB cluster.
        2025/03/26 02:05:17 Instance (gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer) configured to join to the InnoDB cluster
        2025/03/26 02:05:17 peers: [gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer]
        2025/03/26 02:05:17 Running dba.getCluster('grdemandbackup')
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        2025/03/26 02:05:17 Connected to peer gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer
        2025/03/26 02:05:18 Cluster status:
        
        ClusterName: grdemandbackup
        Status: OK_NO_TOLERANCE
        StatusText: Cluster is NOT tolerant to any failures.
        SSL: REQUIRED
        Primary: gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        Topology:
        	
        	Member 0
        	Address: gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        	State: ONLINE
        	Errors: []
        
        		
        	Member 1
        	Address: gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        	State: ONLINE
        	Errors: []
        
        		
        2025/03/26 02:05:18 Adding instance (gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer) to InnoDB cluster
        2025/03/26 02:05:18 Running dba.getCluster('grdemandbackup').addInstance('operator:*****@gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer', {'recoveryMethod': 'clone', 'waitRecovery': 3})
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        WARNING: The waitRecovery option is deprecated. Please use the recoveryProgress option instead.
        
        
        NOTE: The target instance 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' has not been pre-provisioned (GTID set is empty). The Shell is unable to decide whether incremental state recovery can correctly provision it.
        
        Clone based recovery selected through the recoveryMethod option
        
        Validating instance configuration at gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306...
        
        This instance reports its own address as gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        
        Instance configuration is suitable.
        NOTE: Group Replication will communicate with other members using 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'. Use the localAddress option to override.
        
        * Checking connectivity and SSL configuration...
        A new instance will be added to the InnoDB Cluster. Depending on the amount of
        data on the cluster this might take from a few seconds to several hours.
        
        Adding instance to the cluster...
        
        Monitoring recovery process of the new cluster member. Press ^C to stop monitoring and let it continue in background.
        Clone based state recovery is now in progress.
        
        NOTE: A server restart is expected to happen as part of the clone process. If the
        server does not support the RESTART command or does not come back after a
        while, you may need to manually start it back.
        
        * Waiting for clone to finish...
        NOTE: gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306 is being cloned from gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        ** Stage DROP DATA: \ 
** Stage DROP DATA: Completed 
        ** Clone Transfer      FILE COPY      0%  In Progress    PAGE COPY      0%  Not Started    REDO COPY      0%  Not Started** Clone Transfer      FILE COPY      0%  In Progress    PAGE COPY      0%  Not Started    REDO COPY      0%  Not Started** Clone Transfer      FILE COPY      0%  In Progress    PAGE COPY      0%  Not Started    REDO COPY      0%  Not Started** Clone Transfer      FILE COPY    100%  Completed    PAGE COPY      0%  In Progress    REDO COPY      0%  Not Started** Clone Transfer      FILE COPY    100%  Completed    PAGE COPY    100%  Completed    REDO COPY    100%  Completed
        NOTE: gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306 is shutting down...
        
        * Waiting for server restart... \ 
* Waiting for server restart... | 
* Waiting for server restart... / 
* Waiting for server restart... - 
	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:05:32 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Killing	Container mysql failed startup probe, will be restarted	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:05:32 +0000 UTC	Warning	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		FailedPreStopHook	PreStopHook failed	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:05:32 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 116ms (116ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:07 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-5s9cd	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-router-557848cfc-5s9cd to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3wfl	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:07 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-r4499	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-router-557848cfc-r4499 to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3z29	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:07 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z7hrq	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-router-557848cfc-z7hrq to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-gs77	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:07 +0000 UTC	Normal	ReplicaSet.apps gr-demand-backup-router-557848cfc		SuccessfulCreate	Created pod: gr-demand-backup-router-557848cfc-r4499	replicaset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:07 +0000 UTC	Normal	ReplicaSet.apps gr-demand-backup-router-557848cfc		SuccessfulCreate	Created pod: gr-demand-backup-router-557848cfc-5s9cd	replicaset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:07 +0000 UTC	Normal	ReplicaSet.apps gr-demand-backup-router-557848cfc		SuccessfulCreate	Created pod: gr-demand-backup-router-557848cfc-z7hrq	replicaset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:07 +0000 UTC	Normal	Deployment.apps gr-demand-backup-router		ScalingReplicaSet	Scaled up replica set gr-demand-backup-router-557848cfc to 3	deployment-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:08 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-5s9cd.spec.initContainers{router-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:08 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-5s9cd.spec.initContainers{router-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 144ms (144ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:08 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-5s9cd.spec.initContainers{router-init}		Created	Created container: router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:08 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-5s9cd.spec.initContainers{router-init}		Started	Started container router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:08 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-r4499.spec.initContainers{router-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:08 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-r4499.spec.initContainers{router-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 143ms (143ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:08 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-r4499.spec.initContainers{router-init}		Created	Created container: router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:08 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-r4499.spec.initContainers{router-init}		Started	Started container router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:08 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z7hrq.spec.initContainers{router-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:08 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z7hrq.spec.initContainers{router-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 142ms (142ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:08 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z7hrq.spec.initContainers{router-init}		Created	Created container: router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:08 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z7hrq.spec.initContainers{router-init}		Started	Started container router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:10 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-5s9cd.spec.containers{router}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-router"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:10 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-r4499.spec.containers{router}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-router"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:10 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z7hrq.spec.containers{router}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-router"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:24 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-5s9cd.spec.containers{router}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 14.661s (14.661s including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:25 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-5s9cd.spec.containers{router}		Created	Created container: router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:25 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-5s9cd.spec.containers{router}		Started	Started container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:25 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-r4499.spec.containers{router}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 14.755s (14.755s including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:25 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-r4499.spec.containers{router}		Created	Created container: router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:25 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-r4499.spec.containers{router}		Started	Started container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:25 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z7hrq.spec.containers{router}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 14.384s (14.384s including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:25 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z7hrq.spec.containers{router}		Created	Created container: router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:25 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z7hrq.spec.containers{router}		Started	Started container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:50 +0000 UTC	Normal	Pod xb-gr-demand-backup-minio-minio-glh2t	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/xb-gr-demand-backup-minio-minio-glh2t to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3z29	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:50 +0000 UTC	Normal	Pod xb-gr-demand-backup-minio-minio-glh2t.spec.initContainers{xtrabackup-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:50 +0000 UTC	Normal	Job.batch xb-gr-demand-backup-minio-minio		SuccessfulCreate	Created pod: xb-gr-demand-backup-minio-minio-glh2t	job-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:51 +0000 UTC	Normal	Pod xb-gr-demand-backup-minio-minio-glh2t.spec.initContainers{xtrabackup-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 136ms (136ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:51 +0000 UTC	Normal	Pod xb-gr-demand-backup-minio-minio-glh2t.spec.initContainers{xtrabackup-init}		Created	Created container: xtrabackup-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:51 +0000 UTC	Normal	Pod xb-gr-demand-backup-minio-minio-glh2t.spec.initContainers{xtrabackup-init}		Started	Started container xtrabackup-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:52 +0000 UTC	Normal	Pod xb-gr-demand-backup-minio-minio-glh2t.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:52 +0000 UTC	Normal	Pod xb-gr-demand-backup-minio-minio-glh2t.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 101ms (101ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:52 +0000 UTC	Normal	Pod xb-gr-demand-backup-minio-minio-glh2t.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:52 +0000 UTC	Normal	Pod xb-gr-demand-backup-minio-minio-glh2t.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:06:58 +0000 UTC	Normal	Job.batch xb-gr-demand-backup-minio-minio		Completed	Job completed	job-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:07:14 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup}		Killing	Stopping container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:07:14 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Killing	Stopping container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:07:14 +0000 UTC	Normal	StatefulSet.apps gr-demand-backup-mysql		SuccessfulDelete	delete Pod gr-demand-backup-mysql-2 in StatefulSet gr-demand-backup-mysql successful	statefulset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:07:14 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-5s9cd.spec.containers{router}		Killing	Stopping container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:07:14 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-r4499.spec.containers{router}		Killing	Stopping container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:07:14 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z7hrq.spec.containers{router}		Killing	Stopping container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:07:14 +0000 UTC	Normal	ReplicaSet.apps gr-demand-backup-router-557848cfc		SuccessfulDelete	Deleted pod: gr-demand-backup-router-557848cfc-z7hrq	replicaset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:07:14 +0000 UTC	Normal	ReplicaSet.apps gr-demand-backup-router-557848cfc		SuccessfulDelete	Deleted pod: gr-demand-backup-router-557848cfc-r4499	replicaset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:07:14 +0000 UTC	Normal	ReplicaSet.apps gr-demand-backup-router-557848cfc		SuccessfulDelete	Deleted pod: gr-demand-backup-router-557848cfc-5s9cd	replicaset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:07:14 +0000 UTC	Normal	Deployment.apps gr-demand-backup-router		ScalingReplicaSet	Scaled down replica set gr-demand-backup-router-557848cfc to 0 from 3	deployment-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:07:22 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup}		Killing	Stopping container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:07:22 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Killing	Stopping container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:07:22 +0000 UTC	Normal	StatefulSet.apps gr-demand-backup-mysql		SuccessfulDelete	delete Pod gr-demand-backup-mysql-1 in StatefulSet gr-demand-backup-mysql successful	statefulset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:07:29 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Killing	Stopping container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:07:29 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup}		Killing	Stopping container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:07:29 +0000 UTC	Normal	StatefulSet.apps gr-demand-backup-mysql		SuccessfulDelete	delete Pod gr-demand-backup-mysql-0 in StatefulSet gr-demand-backup-mysql successful	statefulset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:07:31 +0000 UTC	Warning	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Unhealthy	Readiness probe failed: 2025/03/26 02:07:31 MySQL state is not ready...
        	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:07:36 +0000 UTC	Warning	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Unhealthy	Readiness probe failed: 2025/03/26 02:07:36 MySQL state is not ready...
        	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:07:39 +0000 UTC	Warning	Pod xb-restore-gr-demand-backup-restore-minio-z5wxd		FailedAttachVolume	Multi-Attach error for volume "pvc-fc9b3101-02dd-4334-af00-d8e38f4fe87c" Volume is already exclusively attached to one node and can't be attached to another	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:07:39 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-minio-z5wxd	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/xb-restore-gr-demand-backup-restore-minio-z5wxd to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3wfl	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:07:39 +0000 UTC	Normal	Job.batch xb-restore-gr-demand-backup-restore-minio		SuccessfulCreate	Created pod: xb-restore-gr-demand-backup-restore-minio-z5wxd	job-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:08:01 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-minio-z5wxd		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-fc9b3101-02dd-4334-af00-d8e38f4fe87c" 	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:08:02 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-minio-z5wxd.spec.initContainers{xtrabackup-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:08:02 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-minio-z5wxd.spec.initContainers{xtrabackup-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 145ms (145ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:08:02 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-minio-z5wxd.spec.initContainers{xtrabackup-init}		Created	Created container: xtrabackup-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:08:02 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-minio-z5wxd.spec.initContainers{xtrabackup-init}		Started	Started container xtrabackup-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:08:04 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-minio-z5wxd.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:08:04 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-minio-z5wxd.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 107ms (107ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:08:04 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-minio-z5wxd.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:08:04 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-minio-z5wxd.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:08:13 +0000 UTC	Normal	Job.batch xb-restore-gr-demand-backup-restore-minio		Completed	Job completed	job-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:08:14 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-mysql-0 to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3z29	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:08:14 +0000 UTC	Warning	Pod gr-demand-backup-mysql-0		FailedAttachVolume	Multi-Attach error for volume "pvc-fc9b3101-02dd-4334-af00-d8e38f4fe87c" Volume is already exclusively attached to one node and can't be attached to another	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:08:31 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-fc9b3101-02dd-4334-af00-d8e38f4fe87c" 	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:08:32 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:08:33 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 141ms (141ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:08:33 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init}		Created	Created container: mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:08:33 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init}		Started	Started container mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:08:34 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-psmysql"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:08:34 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 120ms (120ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:08:34 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Created	Created container: mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:08:35 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Started	Started container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:08:35 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:08:35 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 133ms (133ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:08:35 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:08:35 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:09:08 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-1		WaitForFirstConsumer	waiting for first consumer to be created before binding	persistentvolume-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:09:08 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-1		ExternalProvisioning	Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.	persistentvolume-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:09:08 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-1		Provisioning	External provisioner is provisioning volume for claim "kuttl-test-ready-deer/datadir-gr-demand-backup-mysql-1"	pd.csi.storage.gke.io_gke-90e882eae5d44a1387ff-3819-ecb5-vm_9c1755ac-9c8b-4676-a238-b44d38aab72b	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:09:11 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-1		ProvisioningSucceeded	Successfully provisioned volume pvc-ce51b7f0-1446-478a-bba1-144dbcf8383b	pd.csi.storage.gke.io_gke-90e882eae5d44a1387ff-3819-ecb5-vm_9c1755ac-9c8b-4676-a238-b44d38aab72b	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:09:12 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-mysql-1 to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3wfl	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:09:13 +0000 UTC	Warning	Pod gr-demand-backup-mysql-1		FailedMount	MountVolume.SetUp failed for volume "users" : failed to sync secret cache: timed out waiting for the condition	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:09:19 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-ce51b7f0-1446-478a-bba1-144dbcf8383b" 	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:09:20 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:09:21 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 144ms (144ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:09:21 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init}		Created	Created container: mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:09:21 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init}		Started	Started container mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:09:22 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-psmysql"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:09:22 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 102ms (102ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:09:22 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Created	Created container: mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:09:23 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Started	Started container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:09:23 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:09:23 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 93ms (93ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:09:23 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:09:23 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:09:53 +0000 UTC	Warning	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Unhealthy	Startup probe failed: 2025/03/26 02:09:40 Waiting for MySQL ready state
        2025/03/26 02:09:40 MySQL is ready
        2025/03/26 02:09:40 Bootstrap starting...
        2025/03/26 02:09:40 Running dba.configureLocalInstance('operator:*****@gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer', {'clearReadOnly': true})
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        WARNING: The clearReadOnly option is deprecated and will be removed in a future release.
        
        WARNING: This function is deprecated and will be removed in a future release of MySQL Shell, use dba.configureInstance() instead.
        Configuring local MySQL instance listening at port 3306 for use in an InnoDB cluster...
        
        This instance reports its own address as gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        
        applierWorkerThreads will be set to the default value of 4.
        
        NOTE: Some configuration options need to be fixed:
        
        
        +----------------------------------------+---------------+----------------+----------------------------+
        
        +----------------------------------------+---------------+----------------+----------------------------+
        | Variable                               | Current Value | Required Value | Note                       |+----------------------------------------+---------------+----------------+----------------------------+
        | binlog_transaction_dependency_tracking | COMMIT_ORDER  | WRITESET       | Update the server variable |Disabled super_read_only on the instance 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'
        Enabling super_read_only on the instance 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'
        Configuring instance...
        
        WARNING:*****@binlog_transaction_dependency_tracking' is deprecated and will be removed in a future release. (Code 1287).
        The instance 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' was configured to be used in an InnoDB cluster.
        2025/03/26 02:09:41 Instance (gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer) configured to join to the InnoDB cluster
        2025/03/26 02:09:41 peers: [gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer]
        2025/03/26 02:09:41 Running dba.getCluster('grdemandbackup')
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        2025/03/26 02:09:41 Connected to peer gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer
        2025/03/26 02:09:42 Cluster status:
        
        ClusterName: grdemandbackup
        Status: OK_NO_TOLERANCE
        StatusText: Cluster is NOT tolerant to any failures.
        SSL: REQUIRED
        Primary: gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        Topology:
        	
        	Member 0
        	Address: gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        	State: ONLINE
        	Errors: []
        
        		
        2025/03/26 02:09:42 Adding instance (gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer) to InnoDB cluster
        2025/03/26 02:09:42 Running dba.getCluster('grdemandbackup').addInstance('operator:*****@gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer', {'recoveryMethod': 'clone', 'waitRecovery': 3})
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        WARNING: The waitRecovery option is deprecated. Please use the recoveryProgress option instead.
        
        NOTE: A GTID set check of the MySQL instance at 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' determined that it is missing transactions that were purged from all cluster members.
        NOTE: The target instance 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' has not been pre-provisioned (GTID set is empty). The Shell is unable to determine whether the instance has pre-existing data that would be overwritten with clone based recovery.
        
        Clone based recovery selected through the recoveryMethod option
        
        Validating instance configuration at gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306...
        
        This instance reports its own address as gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        
        Instance configuration is suitable.
        NOTE: Group Replication will communicate with other members using 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'. Use the localAddress option to override.
        
        * Checking connectivity and SSL configuration...
        A new instance will be added to the InnoDB Cluster. Depending on the amount of
        data on the cluster this might take from a few seconds to several hours.
        
        Adding instance to the cluster...
        
        Monitoring recovery process of the new cluster member. Press ^C to stop monitoring and let it continue in background.
        Clone based state recovery is now in progress.
        
        NOTE: A server restart is expected to happen as part of the clone process. If the
        server does not support the RESTART command or does not come back after a
        while, you may need to manually start it back.
        
        * Waiting for clone to finish...
        NOTE: gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306 is being cloned from gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        ** Stage DROP DATA: Completed
        ** Clone Transfer      FILE COPY      0%  In Progress    PAGE COPY      0%  Not Started    REDO COPY      0%  Not Started** Clone Transfer      FILE COPY    100%  Completed    PAGE COPY    100%  Completed    REDO COPY    100%  Completed
        NOTE: gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306 is shutting down...
        
        * Waiting for server restart... \ 
* Waiting for server restart... | 
* Waiting for server restart... / 
	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:09:53 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Killing	Container mysql failed startup probe, will be restarted	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:09:53 +0000 UTC	Warning	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		FailedPreStopHook	PreStopHook failed	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:09:53 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 121ms (122ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:10:25 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-2		WaitForFirstConsumer	waiting for first consumer to be created before binding	persistentvolume-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:10:26 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-2		ExternalProvisioning	Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.	persistentvolume-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:10:26 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-2		Provisioning	External provisioner is provisioning volume for claim "kuttl-test-ready-deer/datadir-gr-demand-backup-mysql-2"	pd.csi.storage.gke.io_gke-90e882eae5d44a1387ff-3819-ecb5-vm_9c1755ac-9c8b-4676-a238-b44d38aab72b	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:10:29 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-2		ProvisioningSucceeded	Successfully provisioned volume pvc-b1a5a29a-eb81-4f98-90f6-d5980b2abe41	pd.csi.storage.gke.io_gke-90e882eae5d44a1387ff-3819-ecb5-vm_9c1755ac-9c8b-4676-a238-b44d38aab72b	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:10:30 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-mysql-2 to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-gs77	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:10:38 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-b1a5a29a-eb81-4f98-90f6-d5980b2abe41" 	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:10:39 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:10:39 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 162ms (162ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:10:39 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init}		Created	Created container: mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:10:39 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init}		Started	Started container mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:10:41 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-psmysql"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:10:41 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 114ms (114ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:10:41 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Created	Created container: mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:10:41 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Started	Started container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:10:41 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:10:41 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 97ms (97ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:10:41 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:10:41 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:12 +0000 UTC	Warning	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Unhealthy	Startup probe failed: 2025/03/26 02:10:59 Waiting for MySQL ready state
        2025/03/26 02:10:59 MySQL is ready
        2025/03/26 02:10:59 Bootstrap starting...
        2025/03/26 02:10:59 Running dba.configureLocalInstance('operator:*****@gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer', {'clearReadOnly': true})
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        WARNING: The clearReadOnly option is deprecated and will be removed in a future release.
        
        WARNING: This function is deprecated and will be removed in a future release of MySQL Shell, use dba.configureInstance() instead.
        Configuring local MySQL instance listening at port 3306 for use in an InnoDB cluster...
        
        This instance reports its own address as gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        
        applierWorkerThreads will be set to the default value of 4.
        
        NOTE: Some configuration options need to be fixed:
        
        
        +----------------------------------------+---------------+----------------+----------------------------+
        
        +----------------------------------------+---------------+----------------+----------------------------+
        | Variable                               | Current Value | Required Value | Note                       |+----------------------------------------+---------------+----------------+----------------------------+
        | binlog_transaction_dependency_tracking | COMMIT_ORDER  | WRITESET       | Update the server variable |Disabled super_read_only on the instance 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'
        Enabling super_read_only on the instance 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'
        Configuring instance...
        
        WARNING:*****@binlog_transaction_dependency_tracking' is deprecated and will be removed in a future release. (Code 1287).
        The instance 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' was configured to be used in an InnoDB cluster.
        2025/03/26 02:10:59 Instance (gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer) configured to join to the InnoDB cluster
        2025/03/26 02:10:59 peers: [gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer]
        2025/03/26 02:10:59 Running dba.getCluster('grdemandbackup')
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        2025/03/26 02:11:00 Connected to peer gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer
        2025/03/26 02:11:00 Cluster status:
        
        ClusterName: grdemandbackup
        Status: OK_NO_TOLERANCE
        StatusText: Cluster is NOT tolerant to any failures.
        SSL: REQUIRED
        Primary: gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        Topology:
        	
        	Member 0
        	Address: gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        	State: ONLINE
        	Errors: []
        
        		
        	Member 1
        	Address: gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        	State: ONLINE
        	Errors: []
        
        		
        2025/03/26 02:11:00 Adding instance (gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer) to InnoDB cluster
        2025/03/26 02:11:00 Running dba.getCluster('grdemandbackup').addInstance('operator:*****@gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer', {'recoveryMethod': 'clone', 'waitRecovery': 3})
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        WARNING: The waitRecovery option is deprecated. Please use the recoveryProgress option instead.
        
        NOTE: A GTID set check of the MySQL instance at 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' determined that it is missing transactions that were purged from all cluster members.
        NOTE: The target instance 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' has not been pre-provisioned (GTID set is empty). The Shell is unable to determine whether the instance has pre-existing data that would be overwritten with clone based recovery.
        
        Clone based recovery selected through the recoveryMethod option
        
        Validating instance configuration at gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306...
        
        This instance reports its own address as gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        
        Instance configuration is suitable.
        NOTE: Group Replication will communicate with other members using 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'. Use the localAddress option to override.
        
        * Checking connectivity and SSL configuration...
        A new instance will be added to the InnoDB Cluster. Depending on the amount of
        data on the cluster this might take from a few seconds to several hours.
        
        Adding instance to the cluster...
        
        Monitoring recovery process of the new cluster member. Press ^C to stop monitoring and let it continue in background.
        Clone based state recovery is now in progress.
        
        NOTE: A server restart is expected to happen as part of the clone process. If the
        server does not support the RESTART command or does not come back after a
        while, you may need to manually start it back.
        
        * Waiting for clone to finish...
        NOTE: gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306 is being cloned from gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        ** Stage DROP DATA: Completed
        ** Clone Transfer      FILE COPY    100%  Completed    PAGE COPY    100%  Completed    REDO COPY    100%  Completed
        NOTE: gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306 is shutting down...
        
        * Waiting for server restart... \ 
* Waiting for server restart... | 
* Waiting for server restart... / 
* Waiting for server restart... - 
	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:12 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Killing	Container mysql failed startup probe, will be restarted	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:12 +0000 UTC	Warning	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		FailedPreStopHook	PreStopHook failed	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:12 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 111ms (111ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:52 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-d5t58	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-router-557848cfc-d5t58 to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-gs77	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:52 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-hskj9	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-router-557848cfc-hskj9 to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3wfl	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:52 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-zfh2c	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-router-557848cfc-zfh2c to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3z29	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:52 +0000 UTC	Normal	ReplicaSet.apps gr-demand-backup-router-557848cfc		SuccessfulCreate	Created pod: gr-demand-backup-router-557848cfc-zfh2c	replicaset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:52 +0000 UTC	Normal	ReplicaSet.apps gr-demand-backup-router-557848cfc		SuccessfulCreate	Created pod: gr-demand-backup-router-557848cfc-hskj9	replicaset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:52 +0000 UTC	Normal	ReplicaSet.apps gr-demand-backup-router-557848cfc		SuccessfulCreate	Created pod: gr-demand-backup-router-557848cfc-d5t58	replicaset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:52 +0000 UTC	Normal	Deployment.apps gr-demand-backup-router		ScalingReplicaSet	Scaled up replica set gr-demand-backup-router-557848cfc to 3 from 0	deployment-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:53 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-d5t58.spec.initContainers{router-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:53 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-d5t58.spec.initContainers{router-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 147ms (147ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:53 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-d5t58.spec.initContainers{router-init}		Created	Created container: router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:53 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-d5t58.spec.initContainers{router-init}		Started	Started container router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:53 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-hskj9.spec.initContainers{router-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:53 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-hskj9.spec.initContainers{router-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 162ms (162ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:53 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-hskj9.spec.initContainers{router-init}		Created	Created container: router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:53 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-hskj9.spec.initContainers{router-init}		Started	Started container router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:53 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-zfh2c.spec.initContainers{router-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:53 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-zfh2c.spec.initContainers{router-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 158ms (158ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:53 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-zfh2c.spec.initContainers{router-init}		Created	Created container: router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:53 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-zfh2c.spec.initContainers{router-init}		Started	Started container router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:54 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-d5t58.spec.containers{router}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-router"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:54 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-d5t58.spec.containers{router}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 108ms (108ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:54 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-d5t58.spec.containers{router}		Created	Created container: router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:54 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-d5t58.spec.containers{router}		Started	Started container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:54 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-zfh2c.spec.containers{router}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-router"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:54 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-zfh2c.spec.containers{router}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 100ms (100ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:54 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-zfh2c.spec.containers{router}		Created	Created container: router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:54 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-zfh2c.spec.containers{router}		Started	Started container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:55 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-hskj9.spec.containers{router}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-router"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:55 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-hskj9.spec.containers{router}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 91ms (91ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:55 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-hskj9.spec.containers{router}		Created	Created container: router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:11:55 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-hskj9.spec.containers{router}		Started	Started container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:12:15 +0000 UTC	Normal	Pod xb-gr-demand-backup-s3-aws-s3-bjlgh	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/xb-gr-demand-backup-s3-aws-s3-bjlgh to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3z29	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:12:15 +0000 UTC	Normal	Job.batch xb-gr-demand-backup-s3-aws-s3		SuccessfulCreate	Created pod: xb-gr-demand-backup-s3-aws-s3-bjlgh	job-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:12:16 +0000 UTC	Normal	Pod xb-gr-demand-backup-s3-aws-s3-bjlgh.spec.initContainers{xtrabackup-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:12:16 +0000 UTC	Normal	Pod xb-gr-demand-backup-s3-aws-s3-bjlgh.spec.initContainers{xtrabackup-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 142ms (142ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:12:16 +0000 UTC	Normal	Pod xb-gr-demand-backup-s3-aws-s3-bjlgh.spec.initContainers{xtrabackup-init}		Created	Created container: xtrabackup-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:12:16 +0000 UTC	Normal	Pod xb-gr-demand-backup-s3-aws-s3-bjlgh.spec.initContainers{xtrabackup-init}		Started	Started container xtrabackup-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:12:18 +0000 UTC	Normal	Pod xb-gr-demand-backup-s3-aws-s3-bjlgh.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:12:18 +0000 UTC	Normal	Pod xb-gr-demand-backup-s3-aws-s3-bjlgh.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 105ms (105ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:12:18 +0000 UTC	Normal	Pod xb-gr-demand-backup-s3-aws-s3-bjlgh.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:12:18 +0000 UTC	Normal	Pod xb-gr-demand-backup-s3-aws-s3-bjlgh.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:12:24 +0000 UTC	Normal	Job.batch xb-gr-demand-backup-s3-aws-s3		Completed	Job completed	job-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:12:40 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup}		Killing	Stopping container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:12:41 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Killing	Stopping container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:12:41 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-d5t58.spec.containers{router}		Killing	Stopping container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:12:41 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-hskj9.spec.containers{router}		Killing	Stopping container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:12:41 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-zfh2c.spec.containers{router}		Killing	Stopping container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:12:41 +0000 UTC	Normal	ReplicaSet.apps gr-demand-backup-router-557848cfc		SuccessfulDelete	Deleted pod: gr-demand-backup-router-557848cfc-zfh2c	replicaset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:12:41 +0000 UTC	Normal	ReplicaSet.apps gr-demand-backup-router-557848cfc		SuccessfulDelete	Deleted pod: gr-demand-backup-router-557848cfc-d5t58	replicaset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:12:41 +0000 UTC	Normal	ReplicaSet.apps gr-demand-backup-router-557848cfc		SuccessfulDelete	Deleted pod: gr-demand-backup-router-557848cfc-hskj9	replicaset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:12:48 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup}		Killing	Stopping container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:12:48 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Killing	Stopping container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:12:55 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Killing	Stopping container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:12:55 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup}		Killing	Stopping container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:12:57 +0000 UTC	Warning	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Unhealthy	Readiness probe failed: 2025/03/26 02:12:57 MySQL state is not ready...
        	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:13:05 +0000 UTC	Normal	Job.batch xb-restore-gr-demand-backup-restore-s3		SuccessfulCreate	Created pod: xb-restore-gr-demand-backup-restore-s3-4zqcw	job-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:13:06 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-s3-4zqcw	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/xb-restore-gr-demand-backup-restore-s3-4zqcw to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3wfl	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:13:06 +0000 UTC	Warning	Pod xb-restore-gr-demand-backup-restore-s3-4zqcw		FailedAttachVolume	Multi-Attach error for volume "pvc-fc9b3101-02dd-4334-af00-d8e38f4fe87c" Volume is already exclusively attached to one node and can't be attached to another	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:13:28 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-s3-4zqcw		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-fc9b3101-02dd-4334-af00-d8e38f4fe87c" 	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:13:29 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-s3-4zqcw.spec.initContainers{xtrabackup-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:13:29 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-s3-4zqcw.spec.initContainers{xtrabackup-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 133ms (133ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:13:29 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-s3-4zqcw.spec.initContainers{xtrabackup-init}		Created	Created container: xtrabackup-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:13:29 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-s3-4zqcw.spec.initContainers{xtrabackup-init}		Started	Started container xtrabackup-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:13:31 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-s3-4zqcw.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:13:31 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-s3-4zqcw.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 111ms (111ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:13:31 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-s3-4zqcw.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:13:31 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-s3-4zqcw.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:13:44 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-mysql-0 to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3z29	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:13:44 +0000 UTC	Normal	Job.batch xb-restore-gr-demand-backup-restore-s3		Completed	Job completed	job-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:13:45 +0000 UTC	Warning	Pod gr-demand-backup-mysql-0		FailedAttachVolume	Multi-Attach error for volume "pvc-fc9b3101-02dd-4334-af00-d8e38f4fe87c" Volume is already exclusively attached to one node and can't be attached to another	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:08 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-fc9b3101-02dd-4334-af00-d8e38f4fe87c" 	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:09 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:09 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 132ms (132ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:09 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init}		Created	Created container: mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:09 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init}		Started	Started container mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:10 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-psmysql"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:10 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 108ms (108ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:10 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Created	Created container: mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:10 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Started	Started container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:10 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:10 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 126ms (126ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:11 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:11 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:44 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-1		WaitForFirstConsumer	waiting for first consumer to be created before binding	persistentvolume-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:44 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-1		ExternalProvisioning	Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.	persistentvolume-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:44 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-1		Provisioning	External provisioner is provisioning volume for claim "kuttl-test-ready-deer/datadir-gr-demand-backup-mysql-1"	pd.csi.storage.gke.io_gke-90e882eae5d44a1387ff-3819-ecb5-vm_9c1755ac-9c8b-4676-a238-b44d38aab72b	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:47 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-1		ProvisioningSucceeded	Successfully provisioned volume pvc-00890f67-4dac-4305-be6c-7675b4a5e499	pd.csi.storage.gke.io_gke-90e882eae5d44a1387ff-3819-ecb5-vm_9c1755ac-9c8b-4676-a238-b44d38aab72b	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:48 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-mysql-1 to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3wfl	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:55 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-00890f67-4dac-4305-be6c-7675b4a5e499" 	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:57 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:57 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 143ms (143ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:57 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init}		Created	Created container: mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:57 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init}		Started	Started container mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:58 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-psmysql"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:58 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 104ms (104ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:58 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Created	Created container: mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:58 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Started	Started container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:58 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:59 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 101ms (101ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:59 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:14:59 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:15:29 +0000 UTC	Warning	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Unhealthy	Startup probe failed: 2025/03/26 02:15:17 Waiting for MySQL ready state
        2025/03/26 02:15:17 MySQL is ready
        2025/03/26 02:15:17 Bootstrap starting...
        2025/03/26 02:15:17 Running dba.configureLocalInstance('operator:*****@gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer', {'clearReadOnly': true})
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        WARNING: The clearReadOnly option is deprecated and will be removed in a future release.
        
        WARNING: This function is deprecated and will be removed in a future release of MySQL Shell, use dba.configureInstance() instead.
        Configuring local MySQL instance listening at port 3306 for use in an InnoDB cluster...
        
        This instance reports its own address as gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        
        applierWorkerThreads will be set to the default value of 4.
        
        NOTE: Some configuration options need to be fixed:
        
        
        +----------------------------------------+---------------+----------------+----------------------------+
        
        +----------------------------------------+---------------+----------------+----------------------------+
        | Variable                               | Current Value | Required Value | Note                       |+----------------------------------------+---------------+----------------+----------------------------+
        | binlog_transaction_dependency_tracking | COMMIT_ORDER  | WRITESET       | Update the server variable |Disabled super_read_only on the instance 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'
        Enabling super_read_only on the instance 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'
        Configuring instance...
        
        WARNING:*****@binlog_transaction_dependency_tracking' is deprecated and will be removed in a future release. (Code 1287).
        The instance 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' was configured to be used in an InnoDB cluster.
        2025/03/26 02:15:17 Instance (gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer) configured to join to the InnoDB cluster
        2025/03/26 02:15:17 peers: [gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer]
        2025/03/26 02:15:17 Running dba.getCluster('grdemandbackup')
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        2025/03/26 02:15:17 Connected to peer gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer
        2025/03/26 02:15:18 Cluster status:
        
        ClusterName: grdemandbackup
        Status: OK_NO_TOLERANCE
        StatusText: Cluster is NOT tolerant to any failures.
        SSL: REQUIRED
        Primary: gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        Topology:
        	
        	Member 0
        	Address: gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        	State: ONLINE
        	Errors: []
        
        		
        2025/03/26 02:15:18 Adding instance (gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer) to InnoDB cluster
        2025/03/26 02:15:18 Running dba.getCluster('grdemandbackup').addInstance('operator:*****@gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer', {'recoveryMethod': 'clone', 'waitRecovery': 3})
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        WARNING: The waitRecovery option is deprecated. Please use the recoveryProgress option instead.
        
        NOTE: A GTID set check of the MySQL instance at 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' determined that it is missing transactions that were purged from all cluster members.
        NOTE: The target instance 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' has not been pre-provisioned (GTID set is empty). The Shell is unable to determine whether the instance has pre-existing data that would be overwritten with clone based recovery.
        
        Clone based recovery selected through the recoveryMethod option
        
        Validating instance configuration at gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306...
        
        This instance reports its own address as gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        
        Instance configuration is suitable.
        NOTE: Group Replication will communicate with other members using 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'. Use the localAddress option to override.
        
        * Checking connectivity and SSL configuration...
        A new instance will be added to the InnoDB Cluster. Depending on the amount of
        data on the cluster this might take from a few seconds to several hours.
        
        Adding instance to the cluster...
        
        Monitoring recovery process of the new cluster member. Press ^C to stop monitoring and let it continue in background.
        Clone based state recovery is now in progress.
        
        NOTE: A server restart is expected to happen as part of the clone process. If the
        server does not support the RESTART command or does not come back after a
        while, you may need to manually start it back.
        
        * Waiting for clone to finish...
        NOTE: gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306 is being cloned from gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        ** Stage DROP DATA: Completed
        ** Clone Transfer      FILE COPY      0%  In Progress    PAGE COPY      0%  Not Started    REDO COPY      0%  Not Started** Clone Transfer      FILE COPY    100%  Completed    PAGE COPY    100%  Completed    REDO COPY    100%  Completed
        NOTE: gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306 is shutting down...
        
        * Waiting for server restart... \ 
* Waiting for server restart... | 
* Waiting for server restart... / 
* Waiting for server restart... - 
	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:15:29 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Killing	Container mysql failed startup probe, will be restarted	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:15:29 +0000 UTC	Warning	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		FailedPreStopHook	PreStopHook failed	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:15:30 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 118ms (118ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:16:02 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-2		WaitForFirstConsumer	waiting for first consumer to be created before binding	persistentvolume-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:16:02 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-2		Provisioning	External provisioner is provisioning volume for claim "kuttl-test-ready-deer/datadir-gr-demand-backup-mysql-2"	pd.csi.storage.gke.io_gke-90e882eae5d44a1387ff-3819-ecb5-vm_9c1755ac-9c8b-4676-a238-b44d38aab72b	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:16:02 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-2		ExternalProvisioning	Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.	persistentvolume-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:16:06 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-2		ProvisioningSucceeded	Successfully provisioned volume pvc-a1b4b6dd-725d-4113-8780-41dd765af42c	pd.csi.storage.gke.io_gke-90e882eae5d44a1387ff-3819-ecb5-vm_9c1755ac-9c8b-4676-a238-b44d38aab72b	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:16:06 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-mysql-2 to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-gs77	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:16:13 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-a1b4b6dd-725d-4113-8780-41dd765af42c" 	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:16:14 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:16:14 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 146ms (146ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:16:15 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init}		Created	Created container: mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:16:15 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init}		Started	Started container mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:16:16 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-psmysql"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:16:16 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 109ms (109ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:16:16 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Created	Created container: mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:16:16 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Started	Started container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:16:16 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:16:16 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 111ms (111ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:16:16 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:16:16 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:16:47 +0000 UTC	Warning	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Unhealthy	Startup probe failed: 2025/03/26 02:16:34 Waiting for MySQL ready state
        2025/03/26 02:16:34 MySQL is ready
        2025/03/26 02:16:34 Bootstrap starting...
        2025/03/26 02:16:34 Running dba.configureLocalInstance('operator:*****@gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer', {'clearReadOnly': true})
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        WARNING: The clearReadOnly option is deprecated and will be removed in a future release.
        
        WARNING: This function is deprecated and will be removed in a future release of MySQL Shell, use dba.configureInstance() instead.
        Configuring local MySQL instance listening at port 3306 for use in an InnoDB cluster...
        
        This instance reports its own address as gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        
        applierWorkerThreads will be set to the default value of 4.
        
        NOTE: Some configuration options need to be fixed:
        
        
        +----------------------------------------+---------------+----------------+----------------------------+
        
        +----------------------------------------+---------------+----------------+----------------------------+
        | Variable                               | Current Value | Required Value | Note                       |+----------------------------------------+---------------+----------------+----------------------------+
        | binlog_transaction_dependency_tracking | COMMIT_ORDER  | WRITESET       | Update the server variable |Disabled super_read_only on the instance 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'
        Enabling super_read_only on the instance 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'
        Configuring instance...
        
        WARNING:*****@binlog_transaction_dependency_tracking' is deprecated and will be removed in a future release. (Code 1287).
        The instance 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' was configured to be used in an InnoDB cluster.
        2025/03/26 02:16:35 Instance (gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer) configured to join to the InnoDB cluster
        2025/03/26 02:16:35 peers: [gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer]
        2025/03/26 02:16:35 Running dba.getCluster('grdemandbackup')
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        2025/03/26 02:16:35 Connected to peer gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer
        2025/03/26 02:16:36 Cluster status:
        
        ClusterName: grdemandbackup
        Status: OK_NO_TOLERANCE
        StatusText: Cluster is NOT tolerant to any failures.
        SSL: REQUIRED
        Primary: gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        Topology:
        	
        	Member 0
        	Address: gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        	State: ONLINE
        	Errors: []
        
        		
        	Member 1
        	Address: gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        	State: ONLINE
        	Errors: []
        
        		
        2025/03/26 02:16:36 Adding instance (gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer) to InnoDB cluster
        2025/03/26 02:16:36 Running dba.getCluster('grdemandbackup').addInstance('operator:*****@gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer', {'recoveryMethod': 'clone', 'waitRecovery': 3})
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        WARNING: The waitRecovery option is deprecated. Please use the recoveryProgress option instead.
        
        NOTE: A GTID set check of the MySQL instance at 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' determined that it is missing transactions that were purged from all cluster members.
        NOTE: The target instance 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' has not been pre-provisioned (GTID set is empty). The Shell is unable to determine whether the instance has pre-existing data that would be overwritten with clone based recovery.
        
        Clone based recovery selected through the recoveryMethod option
        
        Validating instance configuration at gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306...
        
        This instance reports its own address as gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        
        Instance configuration is suitable.
        NOTE: Group Replication will communicate with other members using 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'. Use the localAddress option to override.
        
        * Checking connectivity and SSL configuration...
        A new instance will be added to the InnoDB Cluster. Depending on the amount of
        data on the cluster this might take from a few seconds to several hours.
        
        Adding instance to the cluster...
        
        Monitoring recovery process of the new cluster member. Press ^C to stop monitoring and let it continue in background.
        Clone based state recovery is now in progress.
        
        NOTE: A server restart is expected to happen as part of the clone process. If the
        server does not support the RESTART command or does not come back after a
        while, you may need to manually start it back.
        
        * Waiting for clone to finish...
        NOTE: gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306 is being cloned from gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        ** Stage DROP DATA: Completed
        ** Clone Transfer      FILE COPY    100%  Completed    PAGE COPY    100%  Completed    REDO COPY    100%  Completed
        NOTE: gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306 is shutting down...
        
        * Waiting for server restart... \ 
* Waiting for server restart... | 
* Waiting for server restart... / 
* Waiting for server restart... - 
	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:16:47 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Killing	Container mysql failed startup probe, will be restarted	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:16:47 +0000 UTC	Warning	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		FailedPreStopHook	PreStopHook failed	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:16:47 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 104ms (104ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:25 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-2k4zd	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-router-557848cfc-2k4zd to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3wfl	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:25 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-2k4zd.spec.initContainers{router-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:25 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-2k4zd.spec.initContainers{router-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 136ms (136ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:25 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-2k4zd.spec.initContainers{router-init}		Created	Created container: router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:25 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-xg58r	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-router-557848cfc-xg58r to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3z29	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:25 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-xg58r.spec.initContainers{router-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:25 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-xg58r.spec.initContainers{router-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 153ms (153ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:25 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-xg58r.spec.initContainers{router-init}		Created	Created container: router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:25 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-xg58r.spec.initContainers{router-init}		Started	Started container router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:25 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z76t7	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-router-557848cfc-z76t7 to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-gs77	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:25 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z76t7.spec.initContainers{router-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:25 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z76t7.spec.initContainers{router-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 151ms (151ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:25 +0000 UTC	Normal	ReplicaSet.apps gr-demand-backup-router-557848cfc		SuccessfulCreate	Created pod: gr-demand-backup-router-557848cfc-xg58r	replicaset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:25 +0000 UTC	Normal	ReplicaSet.apps gr-demand-backup-router-557848cfc		SuccessfulCreate	Created pod: gr-demand-backup-router-557848cfc-2k4zd	replicaset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:25 +0000 UTC	Normal	ReplicaSet.apps gr-demand-backup-router-557848cfc		SuccessfulCreate	Created pod: gr-demand-backup-router-557848cfc-z76t7	replicaset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:26 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-2k4zd.spec.initContainers{router-init}		Started	Started container router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:26 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z76t7.spec.initContainers{router-init}		Created	Created container: router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:26 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z76t7.spec.initContainers{router-init}		Started	Started container router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:27 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-xg58r.spec.containers{router}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-router"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:27 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-xg58r.spec.containers{router}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 120ms (120ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:27 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-xg58r.spec.containers{router}		Created	Created container: router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:27 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-xg58r.spec.containers{router}		Started	Started container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:27 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z76t7.spec.containers{router}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-router"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:27 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z76t7.spec.containers{router}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 97ms (97ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:27 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z76t7.spec.containers{router}		Created	Created container: router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:27 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z76t7.spec.containers{router}		Started	Started container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:28 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-2k4zd.spec.containers{router}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-router"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:28 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-2k4zd.spec.containers{router}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 110ms (110ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:28 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-2k4zd.spec.containers{router}		Created	Created container: router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:28 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-2k4zd.spec.containers{router}		Started	Started container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:50 +0000 UTC	Normal	Pod xb-gr-demand-backup-gcp-gcp-cs-jt7ns	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/xb-gr-demand-backup-gcp-gcp-cs-jt7ns to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3z29	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:50 +0000 UTC	Normal	Job.batch xb-gr-demand-backup-gcp-gcp-cs		SuccessfulCreate	Created pod: xb-gr-demand-backup-gcp-gcp-cs-jt7ns	job-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:51 +0000 UTC	Normal	Pod xb-gr-demand-backup-gcp-gcp-cs-jt7ns.spec.initContainers{xtrabackup-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:51 +0000 UTC	Normal	Pod xb-gr-demand-backup-gcp-gcp-cs-jt7ns.spec.initContainers{xtrabackup-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 180ms (180ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:51 +0000 UTC	Normal	Pod xb-gr-demand-backup-gcp-gcp-cs-jt7ns.spec.initContainers{xtrabackup-init}		Created	Created container: xtrabackup-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:51 +0000 UTC	Normal	Pod xb-gr-demand-backup-gcp-gcp-cs-jt7ns.spec.initContainers{xtrabackup-init}		Started	Started container xtrabackup-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:53 +0000 UTC	Normal	Pod xb-gr-demand-backup-gcp-gcp-cs-jt7ns.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:53 +0000 UTC	Normal	Pod xb-gr-demand-backup-gcp-gcp-cs-jt7ns.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 106ms (106ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:53 +0000 UTC	Normal	Pod xb-gr-demand-backup-gcp-gcp-cs-jt7ns.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:17:53 +0000 UTC	Normal	Pod xb-gr-demand-backup-gcp-gcp-cs-jt7ns.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:00 +0000 UTC	Normal	Job.batch xb-gr-demand-backup-gcp-gcp-cs		Completed	Job completed	job-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:15 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup}		Killing	Stopping container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:15 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Killing	Stopping container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:16 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-2k4zd.spec.containers{router}		Killing	Stopping container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:16 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-xg58r.spec.containers{router}		Killing	Stopping container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:16 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z76t7.spec.containers{router}		Killing	Stopping container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:16 +0000 UTC	Normal	ReplicaSet.apps gr-demand-backup-router-557848cfc		SuccessfulDelete	Deleted pod: gr-demand-backup-router-557848cfc-xg58r	replicaset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:16 +0000 UTC	Normal	ReplicaSet.apps gr-demand-backup-router-557848cfc		SuccessfulDelete	Deleted pod: gr-demand-backup-router-557848cfc-z76t7	replicaset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:16 +0000 UTC	Normal	ReplicaSet.apps gr-demand-backup-router-557848cfc		SuccessfulDelete	Deleted pod: gr-demand-backup-router-557848cfc-2k4zd	replicaset-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:19 +0000 UTC	Warning	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Unhealthy	Readiness probe failed: 2025/03/26 02:18:19 MySQL state is not ready...
        	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:23 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup}		Killing	Stopping container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:24 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Killing	Stopping container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:31 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Killing	Stopping container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:31 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup}		Killing	Stopping container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:34 +0000 UTC	Warning	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Unhealthy	Readiness probe failed: 2025/03/26 02:18:34 MySQL state is not ready...
        	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:38 +0000 UTC	Warning	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Unhealthy	Readiness probe errored: rpc error: code = Unknown desc = failed to exec in container: container is in CONTAINER_EXITED state	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:40 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-gcp-7sdfk	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/xb-restore-gr-demand-backup-restore-gcp-7sdfk to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3wfl	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:40 +0000 UTC	Warning	Pod xb-restore-gr-demand-backup-restore-gcp-7sdfk		FailedAttachVolume	Multi-Attach error for volume "pvc-fc9b3101-02dd-4334-af00-d8e38f4fe87c" Volume is already exclusively attached to one node and can't be attached to another	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:40 +0000 UTC	Normal	Job.batch xb-restore-gr-demand-backup-restore-gcp		SuccessfulCreate	Created pod: xb-restore-gr-demand-backup-restore-gcp-7sdfk	job-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:55 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-gcp-7sdfk		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-fc9b3101-02dd-4334-af00-d8e38f4fe87c" 	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:57 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-gcp-7sdfk.spec.initContainers{xtrabackup-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:57 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-gcp-7sdfk.spec.initContainers{xtrabackup-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 184ms (184ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:57 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-gcp-7sdfk.spec.initContainers{xtrabackup-init}		Created	Created container: xtrabackup-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:57 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-gcp-7sdfk.spec.initContainers{xtrabackup-init}		Started	Started container xtrabackup-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:59 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-gcp-7sdfk.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:59 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-gcp-7sdfk.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 101ms (101ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:59 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-gcp-7sdfk.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:18:59 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-gcp-7sdfk.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:19:13 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-mysql-0 to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3z29	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:19:13 +0000 UTC	Warning	Pod gr-demand-backup-mysql-0		FailedAttachVolume	Multi-Attach error for volume "pvc-fc9b3101-02dd-4334-af00-d8e38f4fe87c" Volume is already exclusively attached to one node and can't be attached to another	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:19:13 +0000 UTC	Normal	Job.batch xb-restore-gr-demand-backup-restore-gcp		Completed	Job completed	job-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:19:36 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-fc9b3101-02dd-4334-af00-d8e38f4fe87c" 	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:19:37 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:19:37 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 150ms (150ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:19:37 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init}		Created	Created container: mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:19:37 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init}		Started	Started container mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:19:39 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-psmysql"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:19:39 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 98ms (98ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:19:39 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Created	Created container: mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:19:39 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Started	Started container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:19:39 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:19:39 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 107ms (107ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:19:39 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:19:39 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:20:12 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-1		WaitForFirstConsumer	waiting for first consumer to be created before binding	persistentvolume-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:20:12 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-1		ExternalProvisioning	Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.	persistentvolume-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:20:12 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-1		Provisioning	External provisioner is provisioning volume for claim "kuttl-test-ready-deer/datadir-gr-demand-backup-mysql-1"	pd.csi.storage.gke.io_gke-90e882eae5d44a1387ff-3819-ecb5-vm_9c1755ac-9c8b-4676-a238-b44d38aab72b	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:20:16 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-1		ProvisioningSucceeded	Successfully provisioned volume pvc-db96aef2-5825-45ba-b8b4-695500c1930e	pd.csi.storage.gke.io_gke-90e882eae5d44a1387ff-3819-ecb5-vm_9c1755ac-9c8b-4676-a238-b44d38aab72b	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:20:16 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-mysql-1 to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3wfl	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:20:24 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-db96aef2-5825-45ba-b8b4-695500c1930e" 	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:20:25 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:20:25 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 139ms (139ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:20:25 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init}		Created	Created container: mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:20:25 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init}		Started	Started container mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:20:27 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-psmysql"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:20:27 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 103ms (103ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:20:27 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Created	Created container: mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:20:27 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Started	Started container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:20:27 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:20:27 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 97ms (97ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:20:27 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:20:28 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:20:58 +0000 UTC	Warning	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Unhealthy	Startup probe failed: 2025/03/26 02:20:45 Waiting for MySQL ready state
        2025/03/26 02:20:45 MySQL is ready
        2025/03/26 02:20:45 Bootstrap starting...
        2025/03/26 02:20:45 Running dba.configureLocalInstance('operator:*****@gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer', {'clearReadOnly': true})
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        WARNING: The clearReadOnly option is deprecated and will be removed in a future release.
        
        WARNING: This function is deprecated and will be removed in a future release of MySQL Shell, use dba.configureInstance() instead.
        Configuring local MySQL instance listening at port 3306 for use in an InnoDB cluster...
        
        This instance reports its own address as gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        
        applierWorkerThreads will be set to the default value of 4.
        
        NOTE: Some configuration options need to be fixed:
        
        
        +----------------------------------------+---------------+----------------+----------------------------+
        
        +----------------------------------------+---------------+----------------+----------------------------+
        | Variable                               | Current Value | Required Value | Note                       |+----------------------------------------+---------------+----------------+----------------------------+
        | binlog_transaction_dependency_tracking | COMMIT_ORDER  | WRITESET       | Update the server variable |Disabled super_read_only on the instance 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'
        Enabling super_read_only on the instance 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'
        Configuring instance...
        
        WARNING:*****@binlog_transaction_dependency_tracking' is deprecated and will be removed in a future release. (Code 1287).
        The instance 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' was configured to be used in an InnoDB cluster.
        2025/03/26 02:20:45 Instance (gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer) configured to join to the InnoDB cluster
        2025/03/26 02:20:45 peers: [gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer]
        2025/03/26 02:20:45 Running dba.getCluster('grdemandbackup')
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        2025/03/26 02:20:46 Connected to peer gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer
        2025/03/26 02:20:46 Cluster status:
        
        ClusterName: grdemandbackup
        Status: OK_NO_TOLERANCE
        StatusText: Cluster is NOT tolerant to any failures.
        SSL: REQUIRED
        Primary: gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        Topology:
        	
        	Member 0
        	Address: gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        	State: ONLINE
        	Errors: []
        
        		
        2025/03/26 02:20:46 Adding instance (gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer) to InnoDB cluster
        2025/03/26 02:20:46 Running dba.getCluster('grdemandbackup').addInstance('operator:*****@gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer', {'recoveryMethod': 'clone', 'waitRecovery': 3})
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        WARNING: The waitRecovery option is deprecated. Please use the recoveryProgress option instead.
        
        NOTE: A GTID set check of the MySQL instance at 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' determined that it is missing transactions that were purged from all cluster members.
        NOTE: The target instance 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' has not been pre-provisioned (GTID set is empty). The Shell is unable to determine whether the instance has pre-existing data that would be overwritten with clone based recovery.
        
        Clone based recovery selected through the recoveryMethod option
        
        Validating instance configuration at gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306...
        
        This instance reports its own address as gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        
        Instance configuration is suitable.
        NOTE: Group Replication will communicate with other members using 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'. Use the localAddress option to override.
        
        * Checking connectivity and SSL configuration...
        A new instance will be added to the InnoDB Cluster. Depending on the amount of
        data on the cluster this might take from a few seconds to several hours.
        
        Adding instance to the cluster...
        
        Monitoring recovery process of the new cluster member. Press ^C to stop monitoring and let it continue in background.
        Clone based state recovery is now in progress.
        
        NOTE: A server restart is expected to happen as part of the clone process. If the
        server does not support the RESTART command or does not come back after a
        while, you may need to manually start it back.
        
        * Waiting for clone to finish...
        NOTE: gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306 is being cloned from gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        ** Stage DROP DATA: Completed
        ** Clone Transfer      FILE COPY      0%  In Progress    PAGE COPY      0%  Not Started    REDO COPY      0%  Not Started** Clone Transfer      FILE COPY    100%  Completed    PAGE COPY    100%  Completed    REDO COPY    100%  Completed
        NOTE: gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306 is shutting down...
        
        * Waiting for server restart... \ 
* Waiting for server restart... | 
* Waiting for server restart... / 
	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:20:58 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Killing	Container mysql failed startup probe, will be restarted	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:20:58 +0000 UTC	Warning	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		FailedPreStopHook	PreStopHook failed	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:20:58 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 97ms (97ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:21:30 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-2		WaitForFirstConsumer	waiting for first consumer to be created before binding	persistentvolume-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:21:30 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-2		ExternalProvisioning	Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.	persistentvolume-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:21:30 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-2		Provisioning	External provisioner is provisioning volume for claim "kuttl-test-ready-deer/datadir-gr-demand-backup-mysql-2"	pd.csi.storage.gke.io_gke-90e882eae5d44a1387ff-3819-ecb5-vm_9c1755ac-9c8b-4676-a238-b44d38aab72b	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:21:34 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-2		ProvisioningSucceeded	Successfully provisioned volume pvc-b4f6363c-0aa5-4911-8791-7c591ac0ccfb	pd.csi.storage.gke.io_gke-90e882eae5d44a1387ff-3819-ecb5-vm_9c1755ac-9c8b-4676-a238-b44d38aab72b	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:21:35 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-mysql-2 to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-gs77	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:21:43 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-b4f6363c-0aa5-4911-8791-7c591ac0ccfb" 	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:21:45 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:21:45 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 135ms (135ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:21:45 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init}		Created	Created container: mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:21:45 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init}		Started	Started container mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:21:47 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-psmysql"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:21:47 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 110ms (110ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:21:47 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Created	Created container: mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:21:47 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Started	Started container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:21:47 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:21:47 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 127ms (127ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:21:47 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:21:47 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:22:17 +0000 UTC	Warning	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Unhealthy	Startup probe failed: 2025/03/26 02:22:05 Waiting for MySQL ready state
        2025/03/26 02:22:05 MySQL is ready
        2025/03/26 02:22:05 Bootstrap starting...
        2025/03/26 02:22:05 Running dba.configureLocalInstance('operator:*****@gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer', {'clearReadOnly': true})
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        WARNING: The clearReadOnly option is deprecated and will be removed in a future release.
        
        WARNING: This function is deprecated and will be removed in a future release of MySQL Shell, use dba.configureInstance() instead.
        Configuring local MySQL instance listening at port 3306 for use in an InnoDB cluster...
        
        This instance reports its own address as gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        
        applierWorkerThreads will be set to the default value of 4.
        
        NOTE: Some configuration options need to be fixed:
        
        
        +----------------------------------------+---------------+----------------+----------------------------+
        
        +----------------------------------------+---------------+----------------+----------------------------+
        | Variable                               | Current Value | Required Value | Note                       |+----------------------------------------+---------------+----------------+----------------------------+
        | binlog_transaction_dependency_tracking | COMMIT_ORDER  | WRITESET       | Update the server variable |Disabled super_read_only on the instance 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'
        Enabling super_read_only on the instance 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'
        Configuring instance...
        
        WARNING:*****@binlog_transaction_dependency_tracking' is deprecated and will be removed in a future release. (Code 1287).
        The instance 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' was configured to be used in an InnoDB cluster.
        2025/03/26 02:22:05 Instance (gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer) configured to join to the InnoDB cluster
        2025/03/26 02:22:05 peers: [gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer]
        2025/03/26 02:22:05 Running dba.getCluster('grdemandbackup')
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        2025/03/26 02:22:06 Connected to peer gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer
        2025/03/26 02:22:06 Cluster status:
        
        ClusterName: grdemandbackup
        Status: OK_NO_TOLERANCE
        StatusText: Cluster is NOT tolerant to any failures.
        SSL: REQUIRED
        Primary: gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        Topology:
        	
        	Member 0
        	Address: gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        	State: ONLINE
        	Errors: []
        
        		
        	Member 1
        	Address: gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        	State: ONLINE
        	Errors: []
        
        		
        2025/03/26 02:22:06 Adding instance (gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer) to InnoDB cluster
        2025/03/26 02:22:06 Running dba.getCluster('grdemandbackup').addInstance('operator:*****@gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer', {'recoveryMethod': 'clone', 'waitRecovery': 3})
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        WARNING: The waitRecovery option is deprecated. Please use the recoveryProgress option instead.
        
        NOTE: A GTID set check of the MySQL instance at 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' determined that it is missing transactions that were purged from all cluster members.
        NOTE: The target instance 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' has not been pre-provisioned (GTID set is empty). The Shell is unable to determine whether the instance has pre-existing data that would be overwritten with clone based recovery.
        
        Clone based recovery selected through the recoveryMethod option
        
        Validating instance configuration at gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306...
        
        This instance reports its own address as gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        
        Instance configuration is suitable.
        NOTE: Group Replication will communicate with other members using 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'. Use the localAddress option to override.
        
        * Checking connectivity and SSL configuration...
        A new instance will be added to the InnoDB Cluster. Depending on the amount of
        data on the cluster this might take from a few seconds to several hours.
        
        Adding instance to the cluster...
        
        Monitoring recovery process of the new cluster member. Press ^C to stop monitoring and let it continue in background.
        Clone based state recovery is now in progress.
        
        NOTE: A server restart is expected to happen as part of the clone process. If the
        server does not support the RESTART command or does not come back after a
        while, you may need to manually start it back.
        
        * Waiting for clone to finish...
        NOTE: gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306 is being cloned from gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        ** Stage DROP DATA: Completed
        ** Clone Transfer      FILE COPY    100%  Completed    PAGE COPY    100%  Completed    REDO COPY    100%  Completed
        NOTE: gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306 is shutting down...
        
        * Waiting for server restart... \ 
* Waiting for server restart... | 
* Waiting for server restart... / 
* Waiting for server restart... - 
	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:22:17 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Killing	Container mysql failed startup probe, will be restarted	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:22:17 +0000 UTC	Warning	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		FailedPreStopHook	PreStopHook failed	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:22:18 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 174ms (174ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:22:57 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-48k24	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-router-557848cfc-48k24 to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3wfl	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:22:57 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-v5f6h	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-router-557848cfc-v5f6h to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3z29	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:22:57 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z7t4t	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-router-557848cfc-z7t4t to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-gs77	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:22:57 +0000 UTC	Normal	ReplicaSet.apps gr-demand-backup-router-557848cfc		SuccessfulCreate	(combined from similar events): Created pod: gr-demand-backup-router-557848cfc-lpxpl		
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:22:58 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-48k24.spec.initContainers{router-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:22:58 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-48k24.spec.initContainers{router-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 160ms (160ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:22:58 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-48k24.spec.initContainers{router-init}		Created	Created container: router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:22:58 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-48k24.spec.initContainers{router-init}		Started	Started container router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:22:58 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-v5f6h.spec.initContainers{router-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:22:58 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-v5f6h.spec.initContainers{router-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 168ms (168ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:22:58 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-v5f6h.spec.initContainers{router-init}		Created	Created container: router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:22:58 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-v5f6h.spec.initContainers{router-init}		Started	Started container router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:22:58 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z7t4t.spec.initContainers{router-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:22:58 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z7t4t.spec.initContainers{router-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 142ms (142ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:22:58 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z7t4t.spec.initContainers{router-init}		Created	Created container: router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:22:58 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z7t4t.spec.initContainers{router-init}		Started	Started container router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:00 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-48k24.spec.containers{router}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-router"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:00 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-48k24.spec.containers{router}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 108ms (108ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:00 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-48k24.spec.containers{router}		Created	Created container: router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:00 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-48k24.spec.containers{router}		Started	Started container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:00 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-v5f6h.spec.containers{router}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-router"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:00 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-v5f6h.spec.containers{router}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 101ms (101ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:00 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-v5f6h.spec.containers{router}		Created	Created container: router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:00 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-v5f6h.spec.containers{router}		Started	Started container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:00 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z7t4t.spec.containers{router}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-router"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:00 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z7t4t.spec.containers{router}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 117ms (117ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:00 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z7t4t.spec.containers{router}		Created	Created container: router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:00 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z7t4t.spec.containers{router}		Started	Started container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:20 +0000 UTC	Normal	Pod xb-gr-demand-backup-azure-azure-blob-vrn8q	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/xb-gr-demand-backup-azure-azure-blob-vrn8q to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3z29	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:20 +0000 UTC	Normal	Job.batch xb-gr-demand-backup-azure-azure-blob		SuccessfulCreate	Created pod: xb-gr-demand-backup-azure-azure-blob-vrn8q	job-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:21 +0000 UTC	Normal	Pod xb-gr-demand-backup-azure-azure-blob-vrn8q.spec.initContainers{xtrabackup-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:21 +0000 UTC	Normal	Pod xb-gr-demand-backup-azure-azure-blob-vrn8q.spec.initContainers{xtrabackup-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 157ms (157ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:21 +0000 UTC	Normal	Pod xb-gr-demand-backup-azure-azure-blob-vrn8q.spec.initContainers{xtrabackup-init}		Created	Created container: xtrabackup-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:21 +0000 UTC	Normal	Pod xb-gr-demand-backup-azure-azure-blob-vrn8q.spec.initContainers{xtrabackup-init}		Started	Started container xtrabackup-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:23 +0000 UTC	Normal	Pod xb-gr-demand-backup-azure-azure-blob-vrn8q.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:23 +0000 UTC	Normal	Pod xb-gr-demand-backup-azure-azure-blob-vrn8q.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 142ms (142ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:23 +0000 UTC	Normal	Pod xb-gr-demand-backup-azure-azure-blob-vrn8q.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:23 +0000 UTC	Normal	Pod xb-gr-demand-backup-azure-azure-blob-vrn8q.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:29 +0000 UTC	Normal	Job.batch xb-gr-demand-backup-azure-azure-blob		Completed	Job completed	job-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:46 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup}		Killing	Stopping container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:46 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Killing	Stopping container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:46 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-48k24.spec.containers{router}		Killing	Stopping container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:46 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-v5f6h.spec.containers{router}		Killing	Stopping container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:46 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-z7t4t.spec.containers{router}		Killing	Stopping container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:46 +0000 UTC	Normal	ReplicaSet.apps gr-demand-backup-router-557848cfc		SuccessfulDelete	(combined from similar events): Deleted pod: gr-demand-backup-router-557848cfc-z7t4t		
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:50 +0000 UTC	Warning	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Unhealthy	Readiness probe failed: 2025/03/26 02:23:50 readiness check failed: Member state: OFFLINE
        	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:53 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup}		Killing	Stopping container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:23:53 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Killing	Stopping container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:24:00 +0000 UTC	Warning	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Unhealthy	Readiness probe failed: 2025/03/26 02:24:00 MySQL state is not ready...
        	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:24:01 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Killing	Stopping container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:24:01 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup}		Killing	Stopping container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:24:02 +0000 UTC	Warning	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Unhealthy	Readiness probe failed: 2025/03/26 02:24:02 MySQL state is not ready...
        	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:24:07 +0000 UTC	Warning	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Unhealthy	Readiness probe failed: 2025/03/26 02:24:07 MySQL state is not ready...
        	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:24:10 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-azure-trkqf	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/xb-restore-gr-demand-backup-restore-azure-trkqf to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3wfl	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:24:10 +0000 UTC	Normal	Job.batch xb-restore-gr-demand-backup-restore-azure		SuccessfulCreate	Created pod: xb-restore-gr-demand-backup-restore-azure-trkqf	job-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:24:22 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-azure-trkqf		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-fc9b3101-02dd-4334-af00-d8e38f4fe87c" 	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:24:24 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-azure-trkqf.spec.initContainers{xtrabackup-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:24:24 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-azure-trkqf.spec.initContainers{xtrabackup-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 189ms (189ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:24:24 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-azure-trkqf.spec.initContainers{xtrabackup-init}		Created	Created container: xtrabackup-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:24:24 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-azure-trkqf.spec.initContainers{xtrabackup-init}		Started	Started container xtrabackup-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:24:26 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-azure-trkqf.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:24:26 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-azure-trkqf.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 102ms (102ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:24:26 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-azure-trkqf.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:24:26 +0000 UTC	Normal	Pod xb-restore-gr-demand-backup-restore-azure-trkqf.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:24:41 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-mysql-0 to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3z29	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:24:41 +0000 UTC	Warning	Pod gr-demand-backup-mysql-0		FailedAttachVolume	Multi-Attach error for volume "pvc-fc9b3101-02dd-4334-af00-d8e38f4fe87c" Volume is already exclusively attached to one node and can't be attached to another	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:24:41 +0000 UTC	Normal	Job.batch xb-restore-gr-demand-backup-restore-azure		Completed	Job completed	job-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:03 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-fc9b3101-02dd-4334-af00-d8e38f4fe87c" 	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:04 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:04 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 155ms (155ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:04 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init}		Created	Created container: mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:05 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init}		Started	Started container mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:06 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-psmysql"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:06 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 113ms (113ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:06 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Created	Created container: mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:06 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Started	Started container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:06 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:06 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 115ms (115ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:06 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:06 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:39 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-1		WaitForFirstConsumer	waiting for first consumer to be created before binding	persistentvolume-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:39 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-1		ExternalProvisioning	Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.	persistentvolume-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:39 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-1		Provisioning	External provisioner is provisioning volume for claim "kuttl-test-ready-deer/datadir-gr-demand-backup-mysql-1"	pd.csi.storage.gke.io_gke-90e882eae5d44a1387ff-3819-ecb5-vm_9c1755ac-9c8b-4676-a238-b44d38aab72b	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:43 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-1		ProvisioningSucceeded	Successfully provisioned volume pvc-60e26dae-b28c-4ae9-904a-f29db1a1846b	pd.csi.storage.gke.io_gke-90e882eae5d44a1387ff-3819-ecb5-vm_9c1755ac-9c8b-4676-a238-b44d38aab72b	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:44 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-mysql-1 to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3wfl	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:52 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-60e26dae-b28c-4ae9-904a-f29db1a1846b" 	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:53 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:53 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 171ms (171ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:53 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init}		Created	Created container: mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:53 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init}		Started	Started container mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:55 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-psmysql"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:55 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 106ms (106ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:55 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Created	Created container: mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:55 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Started	Started container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:55 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:55 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 106ms (106ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:56 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:25:56 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:26:26 +0000 UTC	Warning	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Unhealthy	Startup probe failed: 2025/03/26 02:26:13 Waiting for MySQL ready state
        2025/03/26 02:26:13 MySQL is ready
        2025/03/26 02:26:13 Bootstrap starting...
        2025/03/26 02:26:13 Running dba.configureLocalInstance('operator:*****@gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer', {'clearReadOnly': true})
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        WARNING: The clearReadOnly option is deprecated and will be removed in a future release.
        
        WARNING: This function is deprecated and will be removed in a future release of MySQL Shell, use dba.configureInstance() instead.
        Configuring local MySQL instance listening at port 3306 for use in an InnoDB cluster...
        
        This instance reports its own address as gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        
        applierWorkerThreads will be set to the default value of 4.
        
        NOTE: Some configuration options need to be fixed:
        
        
        +----------------------------------------+---------------+----------------+----------------------------+
        
        +----------------------------------------+---------------+----------------+----------------------------+
        | Variable                               | Current Value | Required Value | Note                       |+----------------------------------------+---------------+----------------+----------------------------+
        | binlog_transaction_dependency_tracking | COMMIT_ORDER  | WRITESET       | Update the server variable |Disabled super_read_only on the instance 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'
        Enabling super_read_only on the instance 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'
        Configuring instance...
        
        WARNING:*****@binlog_transaction_dependency_tracking' is deprecated and will be removed in a future release. (Code 1287).
        The instance 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' was configured to be used in an InnoDB cluster.
        2025/03/26 02:26:14 Instance (gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer) configured to join to the InnoDB cluster
        2025/03/26 02:26:14 peers: [gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer]
        2025/03/26 02:26:14 Running dba.getCluster('grdemandbackup')
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        2025/03/26 02:26:14 Connected to peer gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer
        2025/03/26 02:26:14 Cluster status:
        
        ClusterName: grdemandbackup
        Status: OK_NO_TOLERANCE
        StatusText: Cluster is NOT tolerant to any failures.
        SSL: REQUIRED
        Primary: gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        Topology:
        	
        	Member 0
        	Address: gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        	State: ONLINE
        	Errors: []
        
        		
        2025/03/26 02:26:14 Adding instance (gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer) to InnoDB cluster
        2025/03/26 02:26:14 Running dba.getCluster('grdemandbackup').addInstance('operator:*****@gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer', {'recoveryMethod': 'clone', 'waitRecovery': 3})
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        WARNING: The waitRecovery option is deprecated. Please use the recoveryProgress option instead.
        
        NOTE: A GTID set check of the MySQL instance at 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' determined that it is missing transactions that were purged from all cluster members.
        NOTE: The target instance 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' has not been pre-provisioned (GTID set is empty). The Shell is unable to determine whether the instance has pre-existing data that would be overwritten with clone based recovery.
        
        Clone based recovery selected through the recoveryMethod option
        
        Validating instance configuration at gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306...
        
        This instance reports its own address as gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        
        Instance configuration is suitable.
        NOTE: Group Replication will communicate with other members using 'gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'. Use the localAddress option to override.
        
        * Checking connectivity and SSL configuration...
        A new instance will be added to the InnoDB Cluster. Depending on the amount of
        data on the cluster this might take from a few seconds to several hours.
        
        Adding instance to the cluster...
        
        Monitoring recovery process of the new cluster member. Press ^C to stop monitoring and let it continue in background.
        Clone based state recovery is now in progress.
        
        NOTE: A server restart is expected to happen as part of the clone process. If the
        server does not support the RESTART command or does not come back after a
        while, you may need to manually start it back.
        
        * Waiting for clone to finish...
        NOTE: gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306 is being cloned from gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        ** Stage DROP DATA: Completed
        ** Clone Transfer      FILE COPY      0%  In Progress    PAGE COPY      0%  Not Started    REDO COPY      0%  Not Started** Clone Transfer      FILE COPY    100%  Completed    PAGE COPY    100%  Completed    REDO COPY    100%  Completed
        NOTE: gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306 is shutting down...
        
        * Waiting for server restart... \ 
* Waiting for server restart... | 
* Waiting for server restart... / 
* Waiting for server restart... - 
	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:26:26 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Killing	Container mysql failed startup probe, will be restarted	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:26:26 +0000 UTC	Warning	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		FailedPreStopHook	PreStopHook failed	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:26:58 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-2		WaitForFirstConsumer	waiting for first consumer to be created before binding	persistentvolume-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:26:58 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-2		ExternalProvisioning	Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.	persistentvolume-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:26:58 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-2		Provisioning	External provisioner is provisioning volume for claim "kuttl-test-ready-deer/datadir-gr-demand-backup-mysql-2"	pd.csi.storage.gke.io_gke-90e882eae5d44a1387ff-3819-ecb5-vm_9c1755ac-9c8b-4676-a238-b44d38aab72b	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:27:02 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-demand-backup-mysql-2		ProvisioningSucceeded	Successfully provisioned volume pvc-8076e41d-a85f-425e-ab25-f8558fc261ea	pd.csi.storage.gke.io_gke-90e882eae5d44a1387ff-3819-ecb5-vm_9c1755ac-9c8b-4676-a238-b44d38aab72b	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:27:02 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-mysql-2 to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-gs77	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:27:10 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-8076e41d-a85f-425e-ab25-f8558fc261ea" 	attachdetach-controller	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:27:12 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:27:13 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 149ms (149ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:27:13 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init}		Created	Created container: mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:27:13 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init}		Started	Started container mysql-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:27:15 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-psmysql"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:27:15 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 96ms (96ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:27:15 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Created	Created container: mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:27:15 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Started	Started container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:27:15 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:27:15 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 112ms (112ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:27:15 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:27:15 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:27:45 +0000 UTC	Warning	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Unhealthy	Startup probe failed: 2025/03/26 02:27:32 Waiting for MySQL ready state
        2025/03/26 02:27:32 MySQL is ready
        2025/03/26 02:27:32 Bootstrap starting...
        2025/03/26 02:27:32 Running dba.configureLocalInstance('operator:*****@gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer', {'clearReadOnly': true})
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        WARNING: The clearReadOnly option is deprecated and will be removed in a future release.
        
        WARNING: This function is deprecated and will be removed in a future release of MySQL Shell, use dba.configureInstance() instead.
        Configuring local MySQL instance listening at port 3306 for use in an InnoDB cluster...
        
        This instance reports its own address as gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        
        applierWorkerThreads will be set to the default value of 4.
        
        NOTE: Some configuration options need to be fixed:
        
        
        +----------------------------------------+---------------+----------------+----------------------------+
        
        +----------------------------------------+---------------+----------------+----------------------------+
        | Variable                               | Current Value | Required Value | Note                       |+----------------------------------------+---------------+----------------+----------------------------+
        | binlog_transaction_dependency_tracking | COMMIT_ORDER  | WRITESET       | Update the server variable |Disabled super_read_only on the instance 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'
        Enabling super_read_only on the instance 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'
        Configuring instance...
        
        WARNING:*****@binlog_transaction_dependency_tracking' is deprecated and will be removed in a future release. (Code 1287).
        The instance 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' was configured to be used in an InnoDB cluster.
        2025/03/26 02:27:33 Instance (gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer) configured to join to the InnoDB cluster
        2025/03/26 02:27:33 peers: [gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer]
        2025/03/26 02:27:33 Running dba.getCluster('grdemandbackup')
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        2025/03/26 02:27:33 Connected to peer gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer
        2025/03/26 02:27:34 Cluster status:
        
        ClusterName: grdemandbackup
        Status: OK_NO_TOLERANCE
        StatusText: Cluster is NOT tolerant to any failures.
        SSL: REQUIRED
        Primary: gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        Topology:
        	
        	Member 0
        	Address: gr-demand-backup-mysql-0.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        	State: ONLINE
        	Errors: []
        
        		
        	Member 1
        	Address: gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        	State: ONLINE
        	Errors: []
        
        		
        2025/03/26 02:27:34 Adding instance (gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer) to InnoDB cluster
        2025/03/26 02:27:34 Running dba.getCluster('grdemandbackup').addInstance('operator:*****@gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer', {'recoveryMethod': 'clone', 'waitRecovery': 3})
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        WARNING: The waitRecovery option is deprecated. Please use the recoveryProgress option instead.
        
        NOTE: A GTID set check of the MySQL instance at 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' determined that it is missing transactions that were purged from all cluster members.
        NOTE: The target instance 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306' has not been pre-provisioned (GTID set is empty). The Shell is unable to determine whether the instance has pre-existing data that would be overwritten with clone based recovery.
        
        Clone based recovery selected through the recoveryMethod option
        
        Validating instance configuration at gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306...
        
        This instance reports its own address as gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        
        Instance configuration is suitable.
        NOTE: Group Replication will communicate with other members using 'gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306'. Use the localAddress option to override.
        
        * Checking connectivity and SSL configuration...
        A new instance will be added to the InnoDB Cluster. Depending on the amount of
        data on the cluster this might take from a few seconds to several hours.
        
        Adding instance to the cluster...
        
        Monitoring recovery process of the new cluster member. Press ^C to stop monitoring and let it continue in background.
        Clone based state recovery is now in progress.
        
        NOTE: A server restart is expected to happen as part of the clone process. If the
        server does not support the RESTART command or does not come back after a
        while, you may need to manually start it back.
        
        * Waiting for clone to finish...
        NOTE: gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306 is being cloned from gr-demand-backup-mysql-1.gr-demand-backup-mysql.kuttl-test-ready-deer:3306
        ** Stage DROP DATA: Completed
        ** Clone Transfer      FILE COPY    100%  Completed    PAGE COPY    100%  Completed    REDO COPY    100%  Completed
        NOTE: gr-demand-backup-mysql-2.gr-demand-backup-mysql.kuttl-test-ready-deer:3306 is shutting down...
        
        * Waiting for server restart... \ 
* Waiting for server restart... | 
* Waiting for server restart... / 
* Waiting for server restart... - 
	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:27:45 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Killing	Container mysql failed startup probe, will be restarted	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:27:45 +0000 UTC	Warning	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		FailedPreStopHook	PreStopHook failed	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:27:45 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 109ms (109ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:25 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-4rtkl	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-router-557848cfc-4rtkl to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3wfl	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:25 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-gl9hn	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-router-557848cfc-gl9hn to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-3z29	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:25 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-lpxpl	Binding	Scheduled	Successfully assigned kuttl-test-ready-deer/gr-demand-backup-router-557848cfc-lpxpl to gke-jen-ps-875-8c227d86--default-pool-2b2d2c60-gs77	default-scheduler	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:26 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-4rtkl.spec.initContainers{router-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:26 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-4rtkl.spec.initContainers{router-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 152ms (153ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:26 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-4rtkl.spec.initContainers{router-init}		Created	Created container: router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:26 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-4rtkl.spec.initContainers{router-init}		Started	Started container router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:26 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-gl9hn.spec.initContainers{router-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:26 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-gl9hn.spec.initContainers{router-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 194ms (194ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:26 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-gl9hn.spec.initContainers{router-init}		Created	Created container: router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:26 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-gl9hn.spec.initContainers{router-init}		Started	Started container router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:26 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-lpxpl.spec.initContainers{router-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-875-8c227d86"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:26 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-lpxpl.spec.initContainers{router-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-875-8c227d86" in 148ms (148ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:26 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-lpxpl.spec.initContainers{router-init}		Created	Created container: router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:26 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-lpxpl.spec.initContainers{router-init}		Started	Started container router-init	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:28 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-4rtkl.spec.containers{router}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-router"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:28 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-4rtkl.spec.containers{router}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 105ms (105ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:28 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-4rtkl.spec.containers{router}		Created	Created container: router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:28 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-4rtkl.spec.containers{router}		Started	Started container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:28 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-gl9hn.spec.containers{router}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-router"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:28 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-gl9hn.spec.containers{router}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 122ms (122ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:28 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-gl9hn.spec.containers{router}		Created	Created container: router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:28 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-gl9hn.spec.containers{router}		Started	Started container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:28 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-lpxpl.spec.containers{router}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-router"	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:28 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-lpxpl.spec.containers{router}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 97ms (97ms including waiting)	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:28 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-lpxpl.spec.containers{router}		Created	Created container: router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:28 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-lpxpl.spec.containers{router}		Started	Started container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:50 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Killing	Stopping container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:50 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup}		Killing	Stopping container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:50 +0000 UTC	Normal	Pod gr-demand-backup-mysql-1.spec.containers{mysql}		Killing	Stopping container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:50 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup}		Killing	Stopping container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:51 +0000 UTC	Normal	Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup}		Killing	Stopping container xtrabackup	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:51 +0000 UTC	Normal	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Killing	Stopping container mysql	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:51 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-4rtkl.spec.containers{router}		Killing	Stopping container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:51 +0000 UTC	Warning	Pod gr-demand-backup-router-557848cfc-4rtkl.spec.containers{router}		Unhealthy	Readiness probe errored: rpc error: code = NotFound desc = failed to exec in container: failed to load task: no running task found: task 4dab33704a2e1c0b4b44597030c55467231aa0ce5cdf49a9f4239cc7df378e84 not found: not found	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:51 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-gl9hn.spec.containers{router}		Killing	Stopping container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:51 +0000 UTC	Normal	Pod gr-demand-backup-router-557848cfc-lpxpl.spec.containers{router}		Killing	Stopping container router	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:51 +0000 UTC	Warning	Pod gr-demand-backup-router-557848cfc-lpxpl.spec.containers{router}		Unhealthy	Readiness probe errored: rpc error: code = Unknown desc = failed to exec in container: container is in CONTAINER_EXITED state	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:52 +0000 UTC	Warning	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Unhealthy	Readiness probe failed: 2025/03/26 02:28:52 MySQL state is not ready...
        	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:54 +0000 UTC	Warning	Pod gr-demand-backup-mysql-0.spec.containers{mysql}		Unhealthy	Readiness probe failed: 2025/03/26 02:28:54 MySQL state is not ready...
        	kubelet	
    logger.go:42: 02:28:58 | gr-demand-backup | 2025-03-26 02:28:57 +0000 UTC	Warning	Pod gr-demand-backup-mysql-2.spec.containers{mysql}		Unhealthy	Readiness probe failed: 2025/03/26 02:28:57 MySQL state is not ready...
        	kubelet	
    logger.go:42: 02:28:59 | gr-demand-backup | Deleting namespace: kuttl-test-ready-deer
=== NAME  kuttl
    harness.go:407: run tests finished
    harness.go:515: cleaning up
    harness.go:572: removing temp folder: ""
--- PASS: kuttl (1789.35s)
    --- PASS: kuttl/harness (0.00s)
        --- PASS: kuttl/harness/gr-demand-backup (1788.91s)
PASS