=== RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. I0206 08:46:01.022828 13049 request.go:682] Waited for 1.033042413s due to client-side throttling, not priority and fairness, request: GET:https://34.135.38.6/apis/flowcontrol.apiserver.k8s.io/v1beta2?timeout=32s harness.go:275: Successful connection to cluster at: https://34.135.38.6 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 180 seconds for each step harness.go:372: testsuite: e2e-tests/tests has 30 tests === RUN kuttl/harness === RUN kuttl/harness/gr-demand-backup-haproxy === PAUSE kuttl/harness/gr-demand-backup-haproxy === CONT kuttl/harness/gr-demand-backup-haproxy logger.go:42: 08:46:07 | gr-demand-backup-haproxy | Creating namespace: kuttl-test-real-hookworm logger.go:42: 08:46:07 | gr-demand-backup-haproxy/0-minio-secret | starting test step 0-minio-secret logger.go:42: 08:46:10 | gr-demand-backup-haproxy/0-minio-secret | Secret:kuttl-test-real-hookworm/minio-secret created logger.go:42: 08:46:10 | gr-demand-backup-haproxy/0-minio-secret | test step completed 0-minio-secret logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | starting test step 1-deploy-operator logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | running command: [sh -c set -o errexit set -o xtrace source ../../functions init_temp_dir # do this only in the first TestStep deploy_operator deploy_non_tls_cluster_secrets deploy_tls_cluster_secrets deploy_client deploy_minio] logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | + source ../../functions logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ realpath ../../.. logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | ++++ pwd logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-demand-backup-haproxy logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | ++ test_name=gr-demand-backup-haproxy logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup-haproxy logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup-haproxy logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ export GIT_BRANCH=PR-523 logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ GIT_BRANCH=PR-523 logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ export VERSION=PR-523-f00253e logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ VERSION=PR-523-f00253e logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | ++++ which gdate logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | ++++ which date logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ date=/usr/bin/date logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ command -v oc logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ kubectl get nodes logger.go:42: 08:46:10 | gr-demand-backup-haproxy/1-deploy-operator | +++ grep '^minikube' logger.go:42: 08:46:11 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:11.918117 13200 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:12 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:12.139355 13200 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:12 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:12.248586 13200 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:12 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:12.355256 13200 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:12 | gr-demand-backup-haproxy/1-deploy-operator | + init_temp_dir logger.go:42: 08:46:12 | gr-demand-backup-haproxy/1-deploy-operator | + rm -rf /tmp/kuttl/ps/gr-demand-backup-haproxy logger.go:42: 08:46:12 | gr-demand-backup-haproxy/1-deploy-operator | + mkdir -p /tmp/kuttl/ps/gr-demand-backup-haproxy logger.go:42: 08:46:12 | gr-demand-backup-haproxy/1-deploy-operator | + deploy_operator logger.go:42: 08:46:12 | gr-demand-backup-haproxy/1-deploy-operator | + kubectl -n kuttl-test-real-hookworm apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy/crd.yaml logger.go:42: 08:46:13 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:13.618201 13240 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:13 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:13.813503 13240 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:13 | gr-demand-backup-haproxy/1-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlbackups.ps.percona.com serverside-applied logger.go:42: 08:46:14 | gr-demand-backup-haproxy/1-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlrestores.ps.percona.com serverside-applied logger.go:42: 08:46:15 | gr-demand-backup-haproxy/1-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqls.ps.percona.com serverside-applied logger.go:42: 08:46:15 | gr-demand-backup-haproxy/1-deploy-operator | + kubectl -n kuttl-test-real-hookworm apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy/rbac.yaml logger.go:42: 08:46:16 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:16.672514 13280 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:16 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:16.783159 13280 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:17 | gr-demand-backup-haproxy/1-deploy-operator | serviceaccount/percona-server-mysql-operator created logger.go:42: 08:46:17 | gr-demand-backup-haproxy/1-deploy-operator | serviceaccount/percona-server-mysql-operator-orchestrator created logger.go:42: 08:46:18 | gr-demand-backup-haproxy/1-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 08:46:18 | gr-demand-backup-haproxy/1-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 08:46:18 | gr-demand-backup-haproxy/1-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 08:46:19 | gr-demand-backup-haproxy/1-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 08:46:19 | gr-demand-backup-haproxy/1-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 08:46:19 | gr-demand-backup-haproxy/1-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 08:46:19 | gr-demand-backup-haproxy/1-deploy-operator | + yq eval '(select(documentIndex==1).spec.template.spec.containers[] | select(.name=="manager").env[] | select(.name=="DISABLE_TELEMETRY").value) = "true"' logger.go:42: 08:46:19 | gr-demand-backup-haproxy/1-deploy-operator | + yq eval '(select(documentIndex==1).spec.template.spec.containers[] | select(.name=="manager").env[] | select(.name=="LOG_LEVEL").value) = "DEBUG"' logger.go:42: 08:46:19 | gr-demand-backup-haproxy/1-deploy-operator | + kubectl -n kuttl-test-real-hookworm apply -f - logger.go:42: 08:46:19 | gr-demand-backup-haproxy/1-deploy-operator | ++ printf 'select(documentIndex==1).spec.template.spec.containers[0].image="%s"' perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 08:46:19 | gr-demand-backup-haproxy/1-deploy-operator | + yq eval 'select(documentIndex==1).spec.template.spec.containers[0].image="perconalab/percona-server-mysql-operator:PR-523-f00253e"' /mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy/operator.yaml logger.go:42: 08:46:20 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:20.829747 13343 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:21 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:21.054034 13343 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:22 | gr-demand-backup-haproxy/1-deploy-operator | configmap/percona-server-mysql-operator-config created logger.go:42: 08:46:22 | gr-demand-backup-haproxy/1-deploy-operator | deployment.apps/percona-server-mysql-operator created logger.go:42: 08:46:22 | gr-demand-backup-haproxy/1-deploy-operator | + deploy_non_tls_cluster_secrets logger.go:42: 08:46:22 | gr-demand-backup-haproxy/1-deploy-operator | + kubectl -n kuttl-test-real-hookworm apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf/secrets.yaml logger.go:42: 08:46:23 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:23.199115 13391 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:23 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:23.423388 13391 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:23 | gr-demand-backup-haproxy/1-deploy-operator | secret/test-secrets created logger.go:42: 08:46:23 | gr-demand-backup-haproxy/1-deploy-operator | + deploy_tls_cluster_secrets logger.go:42: 08:46:23 | gr-demand-backup-haproxy/1-deploy-operator | + kubectl -n kuttl-test-real-hookworm apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf/ssl-secret.yaml logger.go:42: 08:46:24 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:24.590821 13422 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:24 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:24.913515 13422 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:25 | gr-demand-backup-haproxy/1-deploy-operator | secret/test-ssl created logger.go:42: 08:46:25 | gr-demand-backup-haproxy/1-deploy-operator | + deploy_client logger.go:42: 08:46:25 | gr-demand-backup-haproxy/1-deploy-operator | + kubectl -n kuttl-test-real-hookworm apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf/client.yaml logger.go:42: 08:46:26 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:26.190856 13443 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:26 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:26.296954 13443 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:26 | gr-demand-backup-haproxy/1-deploy-operator | pod/mysql-client created logger.go:42: 08:46:26 | gr-demand-backup-haproxy/1-deploy-operator | + deploy_minio logger.go:42: 08:46:26 | gr-demand-backup-haproxy/1-deploy-operator | + local access_key logger.go:42: 08:46:26 | gr-demand-backup-haproxy/1-deploy-operator | + local secret_key logger.go:42: 08:46:26 | gr-demand-backup-haproxy/1-deploy-operator | ++ kubectl -n kuttl-test-real-hookworm get secret minio-secret -o 'jsonpath={.data.AWS_ACCESS_KEY_ID}' logger.go:42: 08:46:26 | gr-demand-backup-haproxy/1-deploy-operator | ++ base64 -d logger.go:42: 08:46:27 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:27.311534 13466 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:27 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:27.422871 13466 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:27 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:27.529445 13466 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:27 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:27.636117 13466 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:27 | gr-demand-backup-haproxy/1-deploy-operator | + access_key='some-access$\n"-key' logger.go:42: 08:46:27 | gr-demand-backup-haproxy/1-deploy-operator | ++ kubectl -n kuttl-test-real-hookworm get secret minio-secret -o 'jsonpath={.data.AWS_SECRET_ACCESS_KEY}' logger.go:42: 08:46:27 | gr-demand-backup-haproxy/1-deploy-operator | ++ base64 -d logger.go:42: 08:46:28 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:28.329259 13494 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:28 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:28.550703 13494 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:28 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:28.658655 13494 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:28 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:28.765897 13494 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:28 | gr-demand-backup-haproxy/1-deploy-operator | + secret_key='some-$\n"secret-key' logger.go:42: 08:46:28 | gr-demand-backup-haproxy/1-deploy-operator | + helm uninstall -n kuttl-test-real-hookworm minio-service logger.go:42: 08:46:28 | gr-demand-backup-haproxy/1-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 08:46:28 | gr-demand-backup-haproxy/1-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 08:46:29 | gr-demand-backup-haproxy/1-deploy-operator | Error: uninstall: Release not loaded: minio-service: release: not found logger.go:42: 08:46:29 | gr-demand-backup-haproxy/1-deploy-operator | + : logger.go:42: 08:46:29 | gr-demand-backup-haproxy/1-deploy-operator | + helm repo remove minio logger.go:42: 08:46:29 | gr-demand-backup-haproxy/1-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 08:46:29 | gr-demand-backup-haproxy/1-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 08:46:29 | gr-demand-backup-haproxy/1-deploy-operator | Error: no repositories configured logger.go:42: 08:46:29 | gr-demand-backup-haproxy/1-deploy-operator | + : logger.go:42: 08:46:29 | gr-demand-backup-haproxy/1-deploy-operator | + helm repo add minio https://charts.min.io/ logger.go:42: 08:46:29 | gr-demand-backup-haproxy/1-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 08:46:29 | gr-demand-backup-haproxy/1-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 08:46:29 | gr-demand-backup-haproxy/1-deploy-operator | "minio" has been added to your repositories logger.go:42: 08:46:29 | gr-demand-backup-haproxy/1-deploy-operator | +++ printf %q 'some-access$\n"-key' logger.go:42: 08:46:29 | gr-demand-backup-haproxy/1-deploy-operator | ++ printf %q 'some-access\$\\n\"-key' logger.go:42: 08:46:29 | gr-demand-backup-haproxy/1-deploy-operator | +++ printf %q 'some-$\n"secret-key' logger.go:42: 08:46:29 | gr-demand-backup-haproxy/1-deploy-operator | ++ printf %q 'some-\$\\n\"secret-key' logger.go:42: 08:46:29 | gr-demand-backup-haproxy/1-deploy-operator | + retry 10 60 helm install minio-service -n kuttl-test-real-hookworm --version 5.0.14 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access\\\$\\\\n\\\"-key' --set 'users[0].secretKey=some-\\\$\\\\n\\\"secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/.minio/ --set persistence.size=2G --set securityContext.enabled=false minio/minio logger.go:42: 08:46:29 | gr-demand-backup-haproxy/1-deploy-operator | + local max=10 logger.go:42: 08:46:29 | gr-demand-backup-haproxy/1-deploy-operator | + local delay=60 logger.go:42: 08:46:29 | gr-demand-backup-haproxy/1-deploy-operator | + shift 2 logger.go:42: 08:46:29 | gr-demand-backup-haproxy/1-deploy-operator | + local n=1 logger.go:42: 08:46:29 | gr-demand-backup-haproxy/1-deploy-operator | + helm install minio-service -n kuttl-test-real-hookworm --version 5.0.14 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access\\\$\\\\n\\\"-key' --set 'users[0].secretKey=some-\\\$\\\\n\\\"secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/.minio/ --set persistence.size=2G --set securityContext.enabled=false minio/minio logger.go:42: 08:46:29 | gr-demand-backup-haproxy/1-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 08:46:29 | gr-demand-backup-haproxy/1-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 08:46:30 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:30.399808 13548 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:30 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:30.611057 13548 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:32 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:32.890355 13548 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:32 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:32.998839 13548 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:33 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:33.507309 13548 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:33 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:33.627704 13548 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:33 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:33.974699 13548 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:34 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:34.095084 13548 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:34 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:34.451197 13548 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:34 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:34.571556 13548 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:34 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:34.924131 13548 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:35 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:35.034322 13548 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:35 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:35.385616 13548 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:35 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:35.507319 13548 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:35 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:35.852715 13548 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:35 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:35.966029 13548 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:37 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:37.884293 13548 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:38 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:38.003441 13548 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:38 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:38.697832 13548 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:46:38 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:46:38.811183 13548 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:47:16.139471 13548 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:47:16.356166 13548 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | NAME: minio-service logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | LAST DEPLOYED: Tue Feb 6 08:46:30 2024 logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | NAMESPACE: kuttl-test-real-hookworm logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | STATUS: deployed logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | REVISION: 1 logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | TEST SUITE: None logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | NOTES: logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | MinIO can be accessed via port 9000 on the following DNS name from within your cluster: logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | minio-service.kuttl-test-real-hookworm.svc.cluster.local logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | To access MinIO from localhost, run the below commands: logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | 1. export POD_NAME=$(kubectl get pods --namespace kuttl-test-real-hookworm -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}") logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | 2. kubectl port-forward $POD_NAME 9000 --namespace kuttl-test-real-hookworm logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/ logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | You can now access MinIO server on http://localhost:9000. Follow the below steps to connect to MinIO server with mc client: logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | 1. Download the MinIO mc client - https://min.io/docs/minio/linux/reference/minio-mc.html#quickstart logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | 2. export MC_HOST_minio-service-local=http://$(kubectl get secret --namespace kuttl-test-real-hookworm minio-service -o jsonpath="{.data.rootUser}" | base64 --decode):$(kubectl get secret --namespace kuttl-test-real-hookworm minio-service -o jsonpath="{.data.rootPassword}" | base64 --decode)@localhost:9000 logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | 3. mc ls minio-service-local logger.go:42: 08:47:16 | gr-demand-backup-haproxy/1-deploy-operator | ++ kubectl -n kuttl-test-real-hookworm get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' logger.go:42: 08:47:17 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:47:17.499792 14023 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:47:17 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:47:17.631564 14023 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:47:17 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:47:17.738240 14023 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:47:17 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:47:17.845194 14023 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:47:17 | gr-demand-backup-haproxy/1-deploy-operator | + MINIO_POD=minio-service-85cdcd4d44-7fwb9 logger.go:42: 08:47:17 | gr-demand-backup-haproxy/1-deploy-operator | + wait_pod minio-service-85cdcd4d44-7fwb9 logger.go:42: 08:47:17 | gr-demand-backup-haproxy/1-deploy-operator | + local pod=minio-service-85cdcd4d44-7fwb9 logger.go:42: 08:47:17 | gr-demand-backup-haproxy/1-deploy-operator | + set +o xtrace logger.go:42: 08:47:19 | gr-demand-backup-haproxy/1-deploy-operator | minio-service-85cdcd4d44-7fwb9true logger.go:42: 08:47:19 | gr-demand-backup-haproxy/1-deploy-operator | + kubectl -n kuttl-test-real-hookworm run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- bash -c 'AWS_ACCESS_KEY_ID='\''some-access$\n"-key'\'' AWS_SECRET_ACCESS_KEY='\''some-$\n"secret-key'\'' AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url http://minio-service:9000 s3 mb s3://operator-testing' logger.go:42: 08:47:19 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:47:19.595932 14066 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:47:19 | gr-demand-backup-haproxy/1-deploy-operator | E0206 08:47:19.811450 14066 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 08:47:24 | gr-demand-backup-haproxy/1-deploy-operator | If you don't see a command prompt, try pressing enter. logger.go:42: 08:47:25 | gr-demand-backup-haproxy/1-deploy-operator | warning: couldn't attach to pod/aws-cli, falling back to streaming logs: Internal error occurred: error attaching to container: container is in CONTAINER_EXITED state logger.go:42: 08:47:25 | gr-demand-backup-haproxy/1-deploy-operator | make_bucket: operator-testing logger.go:42: 08:47:27 | gr-demand-backup-haproxy/1-deploy-operator | pod "aws-cli" deleted logger.go:42: 08:47:29 | gr-demand-backup-haproxy/1-deploy-operator | test step completed 1-deploy-operator logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | starting test step 2-create-cluster logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | running command: [sh -c set -o errexit set -o xtrace source ../../functions get_cr \ | yq eval '.spec.backup.storages.minio.type="s3"' - \ | yq eval '.spec.backup.storages.minio.s3.bucket="operator-testing"' - \ | yq eval '.spec.backup.storages.minio.s3.credentialsSecret="minio-secret"' - \ | yq eval '.spec.backup.storages.minio.s3.endpointUrl="http://minio-service:9000"' - \ | yq eval '.spec.backup.storages.minio.s3.region="us-east-1"' - \ | yq eval '.spec.mysql.clusterType="group-replication"' - \ | yq eval '.spec.proxy.router.enabled=false' - \ | yq eval '.spec.proxy.haproxy.enabled=true' - \ | kubectl -n "${NAMESPACE}" apply -f -] logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + source ../../functions logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ realpath ../../.. logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | ++++ pwd logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-demand-backup-haproxy logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | ++ test_name=gr-demand-backup-haproxy logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup-haproxy logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup-haproxy logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ export GIT_BRANCH=PR-523 logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ GIT_BRANCH=PR-523 logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ export VERSION=PR-523-f00253e logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ VERSION=PR-523-f00253e logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | ++++ which gdate logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | ++++ which date logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ date=/usr/bin/date logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ command -v oc logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ kubectl get nodes logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | +++ grep '^minikube' logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + get_cr logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + local name_suffix= logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + yq eval '.spec.backup.storages.minio.type="s3"' - logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + yq eval '.spec.backup.storages.minio.s3.credentialsSecret="minio-secret"' - logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + yq eval .spec.proxy.haproxy.enabled=true - logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + kubectl -n kuttl-test-real-hookworm apply -f - logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + yq eval .spec.proxy.router.enabled=false - logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + yq eval '.spec.backup.storages.minio.s3.endpointUrl="http://minio-service:9000"' - logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + yq eval '.spec.mysql.clusterType="group-replication"' - logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + '[' -n '' ']' logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + yq eval - logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + yq eval '.spec.backup.storages.minio.s3.region="us-east-1"' - logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + yq eval '.spec.secretsName="test-secrets"' - logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + yq eval '.spec.sslSecretName="test-ssl"' - logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | ++ printf '.spec.pmm.image="%s"' perconalab/pmm-client:dev-latest logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + yq eval '.spec.backup.storages.minio.s3.bucket="operator-testing"' - logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + yq eval '.spec.upgradeOptions.apply="disabled"' - logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + yq eval '.spec.pmm.image="perconalab/pmm-client:dev-latest"' - logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | ++ printf '.spec.backup.image="%s"' perconalab/percona-server-mysql-operator:main-backup logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | ++ printf '.spec.orchestrator.image="%s"' perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + yq eval '.spec.orchestrator.image="perconalab/percona-server-mysql-operator:main-orchestrator"' - logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | ++ printf '.spec.proxy.haproxy.image="%s"' perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + yq eval '.spec.backup.image="perconalab/percona-server-mysql-operator:main-backup"' - logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | ++ printf '.spec.toolkit.image="%s"' perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | ++ printf '.metadata.name="%s"' gr-demand-backup-haproxy logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + yq eval '.spec.toolkit.image="perconalab/percona-server-mysql-operator:main-toolkit"' - logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + yq eval '.metadata.name="gr-demand-backup-haproxy"' /mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy/cr.yaml logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | ++ printf '.spec.mysql.image="%s"' perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + yq eval '.spec.mysql.image="perconalab/percona-server-mysql-operator:main-psmysql"' - logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + yq eval '.spec.proxy.haproxy.image="perconalab/percona-server-mysql-operator:main-haproxy"' - logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | ++ printf '.spec.initImage="%s"' perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + yq eval '.spec.initImage="perconalab/percona-server-mysql-operator:PR-523-f00253e"' - logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | ++ printf '.spec.proxy.router.image="%s"' perconalab/percona-server-mysql-operator:main-router logger.go:42: 08:47:29 | gr-demand-backup-haproxy/2-create-cluster | + yq eval '.spec.proxy.router.image="perconalab/percona-server-mysql-operator:main-router"' - logger.go:42: 08:47:31 | gr-demand-backup-haproxy/2-create-cluster | perconaservermysql.ps.percona.com/gr-demand-backup-haproxy created logger.go:42: 08:52:43 | gr-demand-backup-haproxy/2-create-cluster | test step completed 2-create-cluster logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | starting test step 3-write-data logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | running command: [sh -c set -o errexit set -o pipefail set -o xtrace source ../../functions run_mysql \ "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password" run_mysql \ "INSERT myDB.myTable (id) VALUES (100500)" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password"] logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | + source ../../functions logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ realpath ../../.. logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | ++++ pwd logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-demand-backup-haproxy logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | ++ test_name=gr-demand-backup-haproxy logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup-haproxy logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup-haproxy logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ export GIT_BRANCH=PR-523 logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ GIT_BRANCH=PR-523 logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ export VERSION=PR-523-f00253e logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ VERSION=PR-523-f00253e logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | ++++ which gdate logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | ++++ which date logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ date=/usr/bin/date logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ command -v oc logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ kubectl get nodes logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ grep '^minikube' logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ get_cluster_name logger.go:42: 08:52:43 | gr-demand-backup-haproxy/3-write-data | +++ kubectl -n kuttl-test-real-hookworm get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 08:52:44 | gr-demand-backup-haproxy/3-write-data | ++ get_haproxy_svc gr-demand-backup-haproxy logger.go:42: 08:52:44 | gr-demand-backup-haproxy/3-write-data | ++ local cluster=gr-demand-backup-haproxy logger.go:42: 08:52:44 | gr-demand-backup-haproxy/3-write-data | ++ echo gr-demand-backup-haproxy-haproxy logger.go:42: 08:52:44 | gr-demand-backup-haproxy/3-write-data | + run_mysql 'CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' '-h gr-demand-backup-haproxy-haproxy -uroot -proot_password' logger.go:42: 08:52:44 | gr-demand-backup-haproxy/3-write-data | + local 'command=CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' logger.go:42: 08:52:44 | gr-demand-backup-haproxy/3-write-data | + local 'uri=-h gr-demand-backup-haproxy-haproxy -uroot -proot_password' logger.go:42: 08:52:44 | gr-demand-backup-haproxy/3-write-data | + local pod= logger.go:42: 08:52:44 | gr-demand-backup-haproxy/3-write-data | ++ get_client_pod logger.go:42: 08:52:44 | gr-demand-backup-haproxy/3-write-data | ++ kubectl -n kuttl-test-real-hookworm get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 08:52:44 | gr-demand-backup-haproxy/3-write-data | + client_pod=mysql-client logger.go:42: 08:52:44 | gr-demand-backup-haproxy/3-write-data | + wait_pod mysql-client logger.go:42: 08:52:44 | gr-demand-backup-haproxy/3-write-data | + local pod=mysql-client logger.go:42: 08:52:44 | gr-demand-backup-haproxy/3-write-data | + set +o xtrace logger.go:42: 08:52:44 | gr-demand-backup-haproxy/3-write-data | mysql-clienttrue logger.go:42: 08:52:44 | gr-demand-backup-haproxy/3-write-data | + kubectl -n kuttl-test-real-hookworm exec mysql-client -- bash -c 'printf '\''%s\n'\'' "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" | mysql -sN -h gr-demand-backup-haproxy-haproxy -uroot -proot_password' logger.go:42: 08:52:44 | gr-demand-backup-haproxy/3-write-data | + sed -e 's/mysql: //' logger.go:42: 08:52:44 | gr-demand-backup-haproxy/3-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 08:52:45 | gr-demand-backup-haproxy/3-write-data | + : logger.go:42: 08:52:45 | gr-demand-backup-haproxy/3-write-data | +++ get_cluster_name logger.go:42: 08:52:45 | gr-demand-backup-haproxy/3-write-data | +++ kubectl -n kuttl-test-real-hookworm get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 08:52:46 | gr-demand-backup-haproxy/3-write-data | ++ get_haproxy_svc gr-demand-backup-haproxy logger.go:42: 08:52:46 | gr-demand-backup-haproxy/3-write-data | ++ local cluster=gr-demand-backup-haproxy logger.go:42: 08:52:46 | gr-demand-backup-haproxy/3-write-data | ++ echo gr-demand-backup-haproxy-haproxy logger.go:42: 08:52:46 | gr-demand-backup-haproxy/3-write-data | + run_mysql 'INSERT myDB.myTable (id) VALUES (100500)' '-h gr-demand-backup-haproxy-haproxy -uroot -proot_password' logger.go:42: 08:52:46 | gr-demand-backup-haproxy/3-write-data | + local 'command=INSERT myDB.myTable (id) VALUES (100500)' logger.go:42: 08:52:46 | gr-demand-backup-haproxy/3-write-data | + local 'uri=-h gr-demand-backup-haproxy-haproxy -uroot -proot_password' logger.go:42: 08:52:46 | gr-demand-backup-haproxy/3-write-data | + local pod= logger.go:42: 08:52:46 | gr-demand-backup-haproxy/3-write-data | ++ get_client_pod logger.go:42: 08:52:46 | gr-demand-backup-haproxy/3-write-data | ++ kubectl -n kuttl-test-real-hookworm get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 08:52:46 | gr-demand-backup-haproxy/3-write-data | + client_pod=mysql-client logger.go:42: 08:52:46 | gr-demand-backup-haproxy/3-write-data | + wait_pod mysql-client logger.go:42: 08:52:46 | gr-demand-backup-haproxy/3-write-data | + local pod=mysql-client logger.go:42: 08:52:46 | gr-demand-backup-haproxy/3-write-data | + set +o xtrace logger.go:42: 08:52:47 | gr-demand-backup-haproxy/3-write-data | mysql-clienttrue logger.go:42: 08:52:47 | gr-demand-backup-haproxy/3-write-data | + sed -e 's/mysql: //' logger.go:42: 08:52:47 | gr-demand-backup-haproxy/3-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 08:52:47 | gr-demand-backup-haproxy/3-write-data | + kubectl -n kuttl-test-real-hookworm exec mysql-client -- bash -c 'printf '\''%s\n'\'' "INSERT myDB.myTable (id) VALUES (100500)" | mysql -sN -h gr-demand-backup-haproxy-haproxy -uroot -proot_password' logger.go:42: 08:52:48 | gr-demand-backup-haproxy/3-write-data | + : logger.go:42: 08:52:49 | gr-demand-backup-haproxy/3-write-data | test step completed 3-write-data logger.go:42: 08:52:49 | gr-demand-backup-haproxy/4-create-backup-minio | starting test step 4-create-backup-minio logger.go:42: 08:52:50 | gr-demand-backup-haproxy/4-create-backup-minio | PerconaServerMySQLBackup:kuttl-test-real-hookworm/gr-demand-backup-haproxy-minio created logger.go:42: 08:53:06 | gr-demand-backup-haproxy/4-create-backup-minio | test step completed 4-create-backup-minio logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | starting test step 5-delete-data logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "TRUNCATE TABLE myDB.myTable" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password" cluster_name=$(get_cluster_name) for i in 0 1 2; do data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 04-delete-data-minio-${i} --from-literal=data="${data}" done] logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | + source ../../functions logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ realpath ../../.. logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | ++++ pwd logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-demand-backup-haproxy logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | ++ test_name=gr-demand-backup-haproxy logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup-haproxy logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup-haproxy logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ export GIT_BRANCH=PR-523 logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ GIT_BRANCH=PR-523 logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ export VERSION=PR-523-f00253e logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ VERSION=PR-523-f00253e logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | ++++ which gdate logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | ++++ which date logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ date=/usr/bin/date logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ command -v oc logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ kubectl get nodes logger.go:42: 08:53:06 | gr-demand-backup-haproxy/5-delete-data | +++ grep '^minikube' logger.go:42: 08:53:07 | gr-demand-backup-haproxy/5-delete-data | +++ get_cluster_name logger.go:42: 08:53:07 | gr-demand-backup-haproxy/5-delete-data | +++ kubectl -n kuttl-test-real-hookworm get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 08:53:07 | gr-demand-backup-haproxy/5-delete-data | ++ get_haproxy_svc gr-demand-backup-haproxy logger.go:42: 08:53:07 | gr-demand-backup-haproxy/5-delete-data | ++ local cluster=gr-demand-backup-haproxy logger.go:42: 08:53:07 | gr-demand-backup-haproxy/5-delete-data | ++ echo gr-demand-backup-haproxy-haproxy logger.go:42: 08:53:07 | gr-demand-backup-haproxy/5-delete-data | + run_mysql 'TRUNCATE TABLE myDB.myTable' '-h gr-demand-backup-haproxy-haproxy -uroot -proot_password' logger.go:42: 08:53:07 | gr-demand-backup-haproxy/5-delete-data | + local 'command=TRUNCATE TABLE myDB.myTable' logger.go:42: 08:53:07 | gr-demand-backup-haproxy/5-delete-data | + local 'uri=-h gr-demand-backup-haproxy-haproxy -uroot -proot_password' logger.go:42: 08:53:07 | gr-demand-backup-haproxy/5-delete-data | + local pod= logger.go:42: 08:53:07 | gr-demand-backup-haproxy/5-delete-data | ++ get_client_pod logger.go:42: 08:53:07 | gr-demand-backup-haproxy/5-delete-data | ++ kubectl -n kuttl-test-real-hookworm get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 08:53:08 | gr-demand-backup-haproxy/5-delete-data | + client_pod=mysql-client logger.go:42: 08:53:08 | gr-demand-backup-haproxy/5-delete-data | + wait_pod mysql-client logger.go:42: 08:53:08 | gr-demand-backup-haproxy/5-delete-data | + local pod=mysql-client logger.go:42: 08:53:08 | gr-demand-backup-haproxy/5-delete-data | + set +o xtrace logger.go:42: 08:53:08 | gr-demand-backup-haproxy/5-delete-data | mysql-clienttrue logger.go:42: 08:53:08 | gr-demand-backup-haproxy/5-delete-data | + kubectl -n kuttl-test-real-hookworm exec mysql-client -- bash -c 'printf '\''%s\n'\'' "TRUNCATE TABLE myDB.myTable" | mysql -sN -h gr-demand-backup-haproxy-haproxy -uroot -proot_password' logger.go:42: 08:53:08 | gr-demand-backup-haproxy/5-delete-data | + sed -e 's/mysql: //' logger.go:42: 08:53:08 | gr-demand-backup-haproxy/5-delete-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 08:53:09 | gr-demand-backup-haproxy/5-delete-data | + : logger.go:42: 08:53:09 | gr-demand-backup-haproxy/5-delete-data | ++ get_cluster_name logger.go:42: 08:53:09 | gr-demand-backup-haproxy/5-delete-data | ++ kubectl -n kuttl-test-real-hookworm get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 08:53:09 | gr-demand-backup-haproxy/5-delete-data | + cluster_name=gr-demand-backup-haproxy logger.go:42: 08:53:09 | gr-demand-backup-haproxy/5-delete-data | + for i in 0 1 2 logger.go:42: 08:53:09 | gr-demand-backup-haproxy/5-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-haproxy-mysql-0.gr-demand-backup-haproxy-mysql -uroot -proot_password' logger.go:42: 08:53:09 | gr-demand-backup-haproxy/5-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 08:53:09 | gr-demand-backup-haproxy/5-delete-data | ++ local 'uri=-h gr-demand-backup-haproxy-mysql-0.gr-demand-backup-haproxy-mysql -uroot -proot_password' logger.go:42: 08:53:09 | gr-demand-backup-haproxy/5-delete-data | ++ local pod= logger.go:42: 08:53:09 | gr-demand-backup-haproxy/5-delete-data | +++ get_client_pod logger.go:42: 08:53:09 | gr-demand-backup-haproxy/5-delete-data | +++ kubectl -n kuttl-test-real-hookworm get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 08:53:10 | gr-demand-backup-haproxy/5-delete-data | ++ client_pod=mysql-client logger.go:42: 08:53:10 | gr-demand-backup-haproxy/5-delete-data | ++ wait_pod mysql-client logger.go:42: 08:53:10 | gr-demand-backup-haproxy/5-delete-data | ++ local pod=mysql-client logger.go:42: 08:53:10 | gr-demand-backup-haproxy/5-delete-data | ++ set +o xtrace logger.go:42: 08:53:10 | gr-demand-backup-haproxy/5-delete-data | mysql-clienttrue logger.go:42: 08:53:10 | gr-demand-backup-haproxy/5-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 08:53:10 | gr-demand-backup-haproxy/5-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 08:53:10 | gr-demand-backup-haproxy/5-delete-data | ++ kubectl -n kuttl-test-real-hookworm exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-haproxy-mysql-0.gr-demand-backup-haproxy-mysql -uroot -proot_password' logger.go:42: 08:53:11 | gr-demand-backup-haproxy/5-delete-data | ++ : logger.go:42: 08:53:11 | gr-demand-backup-haproxy/5-delete-data | + data= logger.go:42: 08:53:11 | gr-demand-backup-haproxy/5-delete-data | + kubectl create configmap -n kuttl-test-real-hookworm 04-delete-data-minio-0 --from-literal=data= logger.go:42: 08:53:12 | gr-demand-backup-haproxy/5-delete-data | configmap/04-delete-data-minio-0 created logger.go:42: 08:53:12 | gr-demand-backup-haproxy/5-delete-data | + for i in 0 1 2 logger.go:42: 08:53:12 | gr-demand-backup-haproxy/5-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-haproxy-mysql-1.gr-demand-backup-haproxy-mysql -uroot -proot_password' logger.go:42: 08:53:12 | gr-demand-backup-haproxy/5-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 08:53:12 | gr-demand-backup-haproxy/5-delete-data | ++ local 'uri=-h gr-demand-backup-haproxy-mysql-1.gr-demand-backup-haproxy-mysql -uroot -proot_password' logger.go:42: 08:53:12 | gr-demand-backup-haproxy/5-delete-data | ++ local pod= logger.go:42: 08:53:12 | gr-demand-backup-haproxy/5-delete-data | +++ get_client_pod logger.go:42: 08:53:12 | gr-demand-backup-haproxy/5-delete-data | +++ kubectl -n kuttl-test-real-hookworm get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 08:53:12 | gr-demand-backup-haproxy/5-delete-data | ++ client_pod=mysql-client logger.go:42: 08:53:12 | gr-demand-backup-haproxy/5-delete-data | ++ wait_pod mysql-client logger.go:42: 08:53:12 | gr-demand-backup-haproxy/5-delete-data | ++ local pod=mysql-client logger.go:42: 08:53:12 | gr-demand-backup-haproxy/5-delete-data | ++ set +o xtrace logger.go:42: 08:53:12 | gr-demand-backup-haproxy/5-delete-data | mysql-clienttrue logger.go:42: 08:53:12 | gr-demand-backup-haproxy/5-delete-data | ++ kubectl -n kuttl-test-real-hookworm exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-haproxy-mysql-1.gr-demand-backup-haproxy-mysql -uroot -proot_password' logger.go:42: 08:53:12 | gr-demand-backup-haproxy/5-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 08:53:12 | gr-demand-backup-haproxy/5-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 08:53:14 | gr-demand-backup-haproxy/5-delete-data | ++ : logger.go:42: 08:53:14 | gr-demand-backup-haproxy/5-delete-data | + data= logger.go:42: 08:53:14 | gr-demand-backup-haproxy/5-delete-data | + kubectl create configmap -n kuttl-test-real-hookworm 04-delete-data-minio-1 --from-literal=data= logger.go:42: 08:53:14 | gr-demand-backup-haproxy/5-delete-data | configmap/04-delete-data-minio-1 created logger.go:42: 08:53:14 | gr-demand-backup-haproxy/5-delete-data | + for i in 0 1 2 logger.go:42: 08:53:14 | gr-demand-backup-haproxy/5-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-haproxy-mysql-2.gr-demand-backup-haproxy-mysql -uroot -proot_password' logger.go:42: 08:53:14 | gr-demand-backup-haproxy/5-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 08:53:14 | gr-demand-backup-haproxy/5-delete-data | ++ local 'uri=-h gr-demand-backup-haproxy-mysql-2.gr-demand-backup-haproxy-mysql -uroot -proot_password' logger.go:42: 08:53:14 | gr-demand-backup-haproxy/5-delete-data | ++ local pod= logger.go:42: 08:53:14 | gr-demand-backup-haproxy/5-delete-data | +++ get_client_pod logger.go:42: 08:53:14 | gr-demand-backup-haproxy/5-delete-data | +++ kubectl -n kuttl-test-real-hookworm get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 08:53:14 | gr-demand-backup-haproxy/5-delete-data | ++ client_pod=mysql-client logger.go:42: 08:53:14 | gr-demand-backup-haproxy/5-delete-data | ++ wait_pod mysql-client logger.go:42: 08:53:14 | gr-demand-backup-haproxy/5-delete-data | ++ local pod=mysql-client logger.go:42: 08:53:14 | gr-demand-backup-haproxy/5-delete-data | ++ set +o xtrace logger.go:42: 08:53:15 | gr-demand-backup-haproxy/5-delete-data | mysql-clienttrue logger.go:42: 08:53:15 | gr-demand-backup-haproxy/5-delete-data | ++ kubectl -n kuttl-test-real-hookworm exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-haproxy-mysql-2.gr-demand-backup-haproxy-mysql -uroot -proot_password' logger.go:42: 08:53:15 | gr-demand-backup-haproxy/5-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 08:53:15 | gr-demand-backup-haproxy/5-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 08:53:16 | gr-demand-backup-haproxy/5-delete-data | ++ : logger.go:42: 08:53:16 | gr-demand-backup-haproxy/5-delete-data | + data= logger.go:42: 08:53:16 | gr-demand-backup-haproxy/5-delete-data | + kubectl create configmap -n kuttl-test-real-hookworm 04-delete-data-minio-2 --from-literal=data= logger.go:42: 08:53:16 | gr-demand-backup-haproxy/5-delete-data | configmap/04-delete-data-minio-2 created logger.go:42: 08:53:18 | gr-demand-backup-haproxy/5-delete-data | test step completed 5-delete-data logger.go:42: 08:53:18 | gr-demand-backup-haproxy/6-restore-from-minio | starting test step 6-restore-from-minio logger.go:42: 08:53:19 | gr-demand-backup-haproxy/6-restore-from-minio | PerconaServerMySQLRestore:kuttl-test-real-hookworm/gr-demand-backup-haproxy-restore-minio created logger.go:42: 08:58:01 | gr-demand-backup-haproxy/6-restore-from-minio | test step completed 6-restore-from-minio logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | starting test step 7-read-data logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions cluster_name=$(get_cluster_name) for i in 0 1 2; do data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 06-read-data-minio-${i} --from-literal=data="${data}" done] logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | + source ../../functions logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ realpath ../../.. logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | ++++ pwd logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-demand-backup-haproxy logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | ++ test_name=gr-demand-backup-haproxy logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup-haproxy logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup-haproxy logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ export GIT_BRANCH=PR-523 logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ GIT_BRANCH=PR-523 logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ export VERSION=PR-523-f00253e logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ VERSION=PR-523-f00253e logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | ++++ which gdate logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | ++++ which date logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ date=/usr/bin/date logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ command -v oc logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ kubectl get nodes logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | +++ grep '^minikube' logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | ++ get_cluster_name logger.go:42: 08:58:01 | gr-demand-backup-haproxy/7-read-data | ++ kubectl -n kuttl-test-real-hookworm get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 08:58:02 | gr-demand-backup-haproxy/7-read-data | + cluster_name=gr-demand-backup-haproxy logger.go:42: 08:58:02 | gr-demand-backup-haproxy/7-read-data | + for i in 0 1 2 logger.go:42: 08:58:02 | gr-demand-backup-haproxy/7-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-haproxy-mysql-0.gr-demand-backup-haproxy-mysql -uroot -proot_password' logger.go:42: 08:58:02 | gr-demand-backup-haproxy/7-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 08:58:02 | gr-demand-backup-haproxy/7-read-data | ++ local 'uri=-h gr-demand-backup-haproxy-mysql-0.gr-demand-backup-haproxy-mysql -uroot -proot_password' logger.go:42: 08:58:02 | gr-demand-backup-haproxy/7-read-data | ++ local pod= logger.go:42: 08:58:02 | gr-demand-backup-haproxy/7-read-data | +++ get_client_pod logger.go:42: 08:58:02 | gr-demand-backup-haproxy/7-read-data | +++ kubectl -n kuttl-test-real-hookworm get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 08:58:02 | gr-demand-backup-haproxy/7-read-data | ++ client_pod=mysql-client logger.go:42: 08:58:02 | gr-demand-backup-haproxy/7-read-data | ++ wait_pod mysql-client logger.go:42: 08:58:02 | gr-demand-backup-haproxy/7-read-data | ++ local pod=mysql-client logger.go:42: 08:58:02 | gr-demand-backup-haproxy/7-read-data | ++ set +o xtrace logger.go:42: 08:58:03 | gr-demand-backup-haproxy/7-read-data | mysql-clienttrue logger.go:42: 08:58:03 | gr-demand-backup-haproxy/7-read-data | ++ kubectl -n kuttl-test-real-hookworm exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-haproxy-mysql-0.gr-demand-backup-haproxy-mysql -uroot -proot_password' logger.go:42: 08:58:03 | gr-demand-backup-haproxy/7-read-data | ++ sed -e 's/mysql: //' logger.go:42: 08:58:03 | gr-demand-backup-haproxy/7-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 08:58:04 | gr-demand-backup-haproxy/7-read-data | + data=100500 logger.go:42: 08:58:04 | gr-demand-backup-haproxy/7-read-data | + kubectl create configmap -n kuttl-test-real-hookworm 06-read-data-minio-0 --from-literal=data=100500 logger.go:42: 08:58:04 | gr-demand-backup-haproxy/7-read-data | configmap/06-read-data-minio-0 created logger.go:42: 08:58:04 | gr-demand-backup-haproxy/7-read-data | + for i in 0 1 2 logger.go:42: 08:58:04 | gr-demand-backup-haproxy/7-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-haproxy-mysql-1.gr-demand-backup-haproxy-mysql -uroot -proot_password' logger.go:42: 08:58:04 | gr-demand-backup-haproxy/7-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 08:58:04 | gr-demand-backup-haproxy/7-read-data | ++ local 'uri=-h gr-demand-backup-haproxy-mysql-1.gr-demand-backup-haproxy-mysql -uroot -proot_password' logger.go:42: 08:58:04 | gr-demand-backup-haproxy/7-read-data | ++ local pod= logger.go:42: 08:58:04 | gr-demand-backup-haproxy/7-read-data | +++ get_client_pod logger.go:42: 08:58:04 | gr-demand-backup-haproxy/7-read-data | +++ kubectl -n kuttl-test-real-hookworm get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 08:58:05 | gr-demand-backup-haproxy/7-read-data | ++ client_pod=mysql-client logger.go:42: 08:58:05 | gr-demand-backup-haproxy/7-read-data | ++ wait_pod mysql-client logger.go:42: 08:58:05 | gr-demand-backup-haproxy/7-read-data | ++ local pod=mysql-client logger.go:42: 08:58:05 | gr-demand-backup-haproxy/7-read-data | ++ set +o xtrace logger.go:42: 08:58:05 | gr-demand-backup-haproxy/7-read-data | mysql-clienttrue logger.go:42: 08:58:05 | gr-demand-backup-haproxy/7-read-data | ++ kubectl -n kuttl-test-real-hookworm exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-haproxy-mysql-1.gr-demand-backup-haproxy-mysql -uroot -proot_password' logger.go:42: 08:58:05 | gr-demand-backup-haproxy/7-read-data | ++ sed -e 's/mysql: //' logger.go:42: 08:58:05 | gr-demand-backup-haproxy/7-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 08:58:06 | gr-demand-backup-haproxy/7-read-data | + data=100500 logger.go:42: 08:58:06 | gr-demand-backup-haproxy/7-read-data | + kubectl create configmap -n kuttl-test-real-hookworm 06-read-data-minio-1 --from-literal=data=100500 logger.go:42: 08:58:06 | gr-demand-backup-haproxy/7-read-data | configmap/06-read-data-minio-1 created logger.go:42: 08:58:06 | gr-demand-backup-haproxy/7-read-data | + for i in 0 1 2 logger.go:42: 08:58:06 | gr-demand-backup-haproxy/7-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-haproxy-mysql-2.gr-demand-backup-haproxy-mysql -uroot -proot_password' logger.go:42: 08:58:06 | gr-demand-backup-haproxy/7-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 08:58:06 | gr-demand-backup-haproxy/7-read-data | ++ local 'uri=-h gr-demand-backup-haproxy-mysql-2.gr-demand-backup-haproxy-mysql -uroot -proot_password' logger.go:42: 08:58:06 | gr-demand-backup-haproxy/7-read-data | ++ local pod= logger.go:42: 08:58:06 | gr-demand-backup-haproxy/7-read-data | +++ get_client_pod logger.go:42: 08:58:06 | gr-demand-backup-haproxy/7-read-data | +++ kubectl -n kuttl-test-real-hookworm get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 08:58:07 | gr-demand-backup-haproxy/7-read-data | ++ client_pod=mysql-client logger.go:42: 08:58:07 | gr-demand-backup-haproxy/7-read-data | ++ wait_pod mysql-client logger.go:42: 08:58:07 | gr-demand-backup-haproxy/7-read-data | ++ local pod=mysql-client logger.go:42: 08:58:07 | gr-demand-backup-haproxy/7-read-data | ++ set +o xtrace logger.go:42: 08:58:07 | gr-demand-backup-haproxy/7-read-data | mysql-clienttrue logger.go:42: 08:58:07 | gr-demand-backup-haproxy/7-read-data | ++ sed -e 's/mysql: //' logger.go:42: 08:58:07 | gr-demand-backup-haproxy/7-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 08:58:07 | gr-demand-backup-haproxy/7-read-data | ++ kubectl -n kuttl-test-real-hookworm exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-haproxy-mysql-2.gr-demand-backup-haproxy-mysql -uroot -proot_password' logger.go:42: 08:58:08 | gr-demand-backup-haproxy/7-read-data | + data=100500 logger.go:42: 08:58:08 | gr-demand-backup-haproxy/7-read-data | + kubectl create configmap -n kuttl-test-real-hookworm 06-read-data-minio-2 --from-literal=data=100500 logger.go:42: 08:58:09 | gr-demand-backup-haproxy/7-read-data | configmap/06-read-data-minio-2 created logger.go:42: 08:58:10 | gr-demand-backup-haproxy/7-read-data | test step completed 7-read-data logger.go:42: 08:58:10 | gr-demand-backup-haproxy/99-drop-finalizer | starting test step 99-drop-finalizer logger.go:42: 08:58:11 | gr-demand-backup-haproxy/99-drop-finalizer | PerconaServerMySQL:kuttl-test-real-hookworm/gr-demand-backup-haproxy updated logger.go:42: 08:58:11 | gr-demand-backup-haproxy/99-drop-finalizer | test step completed 99-drop-finalizer logger.go:42: 08:58:11 | gr-demand-backup-haproxy | gr-demand-backup-haproxy events from ns kuttl-test-real-hookworm: logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:22 +0000 UTC Normal Pod percona-server-mysql-operator-b4c599bbb-tj6g6 Scheduled Successfully assigned kuttl-test-real-hookworm/percona-server-mysql-operator-b4c599bbb-tj6g6 to gke-jen-ps-523-f00253e-5-default-pool-6e6f9275-77nf default-scheduler logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:22 +0000 UTC Normal ReplicaSet.apps percona-server-mysql-operator-b4c599bbb SuccessfulCreate Created pod: percona-server-mysql-operator-b4c599bbb-tj6g6 replicaset-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:22 +0000 UTC Normal Deployment.apps percona-server-mysql-operator ScalingReplicaSet Scaled up replica set percona-server-mysql-operator-b4c599bbb to 1 deployment-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:24 +0000 UTC Normal Pod percona-server-mysql-operator-b4c599bbb-tj6g6.spec.containers{manager} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:26 +0000 UTC Normal Pod mysql-client Scheduled Successfully assigned kuttl-test-real-hookworm/mysql-client to gke-jen-ps-523-f00253e-5-default-pool-6e6f9275-0brr default-scheduler logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:27 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Pulling Pulling image "percona/percona-server:8.0.33" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:37 +0000 UTC Normal ReplicaSet.apps minio-service-85cdcd4d44 SuccessfulCreate Created pod: minio-service-85cdcd4d44-7fwb9 replicaset-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:37 +0000 UTC Normal PersistentVolumeClaim minio-service WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:37 +0000 UTC Normal Deployment.apps minio-service ScalingReplicaSet Scaled up replica set minio-service-85cdcd4d44 to 1 deployment-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:37 +0000 UTC Normal PersistentVolumeClaim minio-service ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:37 +0000 UTC Normal PersistentVolumeClaim minio-service Provisioning External provisioner is provisioning volume for claim "kuttl-test-real-hookworm/minio-service" pd.csi.storage.gke.io_gke-215547eb7cc0414a9857-296b-163f-vm_12151542-af00-4a77-bf44-27e93719a638 logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:38 +0000 UTC Normal Lease.coordination.k8s.io 08db2feb.percona.com LeaderElection percona-server-mysql-operator-b4c599bbb-tj6g6_536ee573-95ca-4dfb-bb79-fa4a1a0f4736 became leader percona-server-mysql-operator-b4c599bbb-tj6g6_536ee573-95ca-4dfb-bb79-fa4a1a0f4736 logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:38 +0000 UTC Normal Pod percona-server-mysql-operator-b4c599bbb-tj6g6.spec.containers{manager} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 13.600703862s (13.600834431s including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:38 +0000 UTC Normal Pod percona-server-mysql-operator-b4c599bbb-tj6g6.spec.containers{manager} Created Created container manager kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:38 +0000 UTC Normal Pod percona-server-mysql-operator-b4c599bbb-tj6g6.spec.containers{manager} Started Started container manager kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:39 +0000 UTC Normal Pod minio-service-post-job-d6brl Scheduled Successfully assigned kuttl-test-real-hookworm/minio-service-post-job-d6brl to gke-jen-ps-523-f00253e-5-default-pool-6e6f9275-0brr default-scheduler logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:39 +0000 UTC Normal Pod minio-service-post-job-d6brl.spec.containers{minio-make-user} Pulling Pulling image "quay.io/minio/mc:RELEASE.2023-09-29T16-41-22Z" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:39 +0000 UTC Normal Job.batch minio-service-post-job SuccessfulCreate Created pod: minio-service-post-job-d6brl job-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:41 +0000 UTC Normal Pod minio-service-85cdcd4d44-7fwb9 Scheduled Successfully assigned kuttl-test-real-hookworm/minio-service-85cdcd4d44-7fwb9 to gke-jen-ps-523-f00253e-5-default-pool-6e6f9275-0brr default-scheduler logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:41 +0000 UTC Normal PersistentVolumeClaim minio-service ProvisioningSucceeded Successfully provisioned volume pvc-95176ead-c0dc-4209-9487-c02f0bb3b4d1 pd.csi.storage.gke.io_gke-215547eb7cc0414a9857-296b-163f-vm_12151542-af00-4a77-bf44-27e93719a638 logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:49 +0000 UTC Normal Pod minio-service-85cdcd4d44-7fwb9 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-95176ead-c0dc-4209-9487-c02f0bb3b4d1" attachdetach-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:49 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Pulled Successfully pulled image "percona/percona-server:8.0.33" in 22.030161422s (22.030372267s including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:49 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Created Created container mysql-client kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:49 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Started Started container mysql-client kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:50 +0000 UTC Normal Pod minio-service-85cdcd4d44-7fwb9.spec.containers{minio} Pulling Pulling image "quay.io/minio/minio:RELEASE.2023-09-30T07-02-29Z" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:46:59 +0000 UTC Normal Pod minio-service-post-job-d6brl.spec.containers{minio-make-user} Pulled Successfully pulled image "quay.io/minio/mc:RELEASE.2023-09-29T16-41-22Z" in 9.225954386s (19.485630493s including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:47:00 +0000 UTC Normal Pod minio-service-post-job-d6brl.spec.containers{minio-make-user} Created Created container minio-make-user kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:47:00 +0000 UTC Normal Pod minio-service-post-job-d6brl.spec.containers{minio-make-user} Started Started container minio-make-user kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:47:03 +0000 UTC Normal Pod minio-service-85cdcd4d44-7fwb9.spec.containers{minio} Pulled Successfully pulled image "quay.io/minio/minio:RELEASE.2023-09-30T07-02-29Z" in 4.409220406s (12.964309771s including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:47:03 +0000 UTC Normal Pod minio-service-85cdcd4d44-7fwb9.spec.containers{minio} Created Created container minio kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:47:03 +0000 UTC Normal Pod minio-service-85cdcd4d44-7fwb9.spec.containers{minio} Started Started container minio kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:47:15 +0000 UTC Normal Job.batch minio-service-post-job Completed Job completed job-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:47:19 +0000 UTC Normal Pod aws-cli Scheduled Successfully assigned kuttl-test-real-hookworm/aws-cli to gke-jen-ps-523-f00253e-5-default-pool-6e6f9275-0brr default-scheduler logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:47:20 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Pulling Pulling image "perconalab/awscli" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:47:23 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Pulled Successfully pulled image "perconalab/awscli" in 3.163929246s (3.163946685s including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:47:23 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Created Created container aws-cli kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:47:23 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Started Started container aws-cli kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:47:34 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-haproxy-mysql-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:47:34 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-haproxy-mysql-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-real-hookworm/datadir-gr-demand-backup-haproxy-mysql-0" pd.csi.storage.gke.io_gke-215547eb7cc0414a9857-296b-163f-vm_12151542-af00-4a77-bf44-27e93719a638 logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:47:34 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-haproxy-mysql-0 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:47:34 +0000 UTC Normal StatefulSet.apps gr-demand-backup-haproxy-mysql SuccessfulCreate create Claim datadir-gr-demand-backup-haproxy-mysql-0 Pod gr-demand-backup-haproxy-mysql-0 in StatefulSet gr-demand-backup-haproxy-mysql success statefulset-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:47:34 +0000 UTC Normal StatefulSet.apps gr-demand-backup-haproxy-mysql SuccessfulCreate create Pod gr-demand-backup-haproxy-mysql-0 in StatefulSet gr-demand-backup-haproxy-mysql successful statefulset-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:47:38 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-haproxy-mysql-0 ProvisioningSucceeded Successfully provisioned volume pvc-05875a35-edc2-4040-be55-ef6636277ccd pd.csi.storage.gke.io_gke-215547eb7cc0414a9857-296b-163f-vm_12151542-af00-4a77-bf44-27e93719a638 logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:47:38 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0 Scheduled Successfully assigned kuttl-test-real-hookworm/gr-demand-backup-haproxy-mysql-0 to gke-jen-ps-523-f00253e-5-default-pool-6e6f9275-77nf default-scheduler logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:47:46 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-05875a35-edc2-4040-be55-ef6636277ccd" attachdetach-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:47:48 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:47:48 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 90.594746ms (90.60287ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:47:48 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:47:48 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:47:49 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:48:09 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 19.761387166s (19.761402231s including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:48:09 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:48:09 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:48:09 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:48:17 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 7.712817639s (7.712835304s including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:48:17 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:48:17 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:48:42 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-haproxy-mysql-1 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:48:42 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-haproxy-mysql-1 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:48:42 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-haproxy-mysql-1 Provisioning External provisioner is provisioning volume for claim "kuttl-test-real-hookworm/datadir-gr-demand-backup-haproxy-mysql-1" pd.csi.storage.gke.io_gke-215547eb7cc0414a9857-296b-163f-vm_12151542-af00-4a77-bf44-27e93719a638 logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:48:42 +0000 UTC Normal StatefulSet.apps gr-demand-backup-haproxy-mysql SuccessfulCreate create Claim datadir-gr-demand-backup-haproxy-mysql-1 Pod gr-demand-backup-haproxy-mysql-1 in StatefulSet gr-demand-backup-haproxy-mysql success statefulset-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:48:42 +0000 UTC Normal StatefulSet.apps gr-demand-backup-haproxy-mysql SuccessfulCreate create Pod gr-demand-backup-haproxy-mysql-1 in StatefulSet gr-demand-backup-haproxy-mysql successful statefulset-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:48:46 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-haproxy-mysql-1 ProvisioningSucceeded Successfully provisioned volume pvc-faeaa0f8-e486-41f5-b22d-9cd42fa934f0 pd.csi.storage.gke.io_gke-215547eb7cc0414a9857-296b-163f-vm_12151542-af00-4a77-bf44-27e93719a638 logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:48:46 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0 Scheduled Successfully assigned kuttl-test-real-hookworm/gr-demand-backup-haproxy-haproxy-0 to gke-jen-ps-523-f00253e-5-default-pool-6e6f9275-0brr default-scheduler logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:48:46 +0000 UTC Normal StatefulSet.apps gr-demand-backup-haproxy-haproxy SuccessfulCreate create Pod gr-demand-backup-haproxy-haproxy-0 in StatefulSet gr-demand-backup-haproxy-haproxy successful statefulset-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:48:46 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1 Scheduled Successfully assigned kuttl-test-real-hookworm/gr-demand-backup-haproxy-mysql-1 to gke-jen-ps-523-f00253e-5-default-pool-6e6f9275-0brr default-scheduler logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:48:47 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:48:52 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-faeaa0f8-e486-41f5-b22d-9cd42fa934f0" attachdetach-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:48:53 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:02 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 14.789201943s (14.789291911s including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:02 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.initContainers{haproxy-init} Created Created container haproxy-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:02 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.initContainers{haproxy-init} Started Started container haproxy-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:02 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 126.734211ms (8.948963474s including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:02 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:02 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:10 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:10 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:13 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 3.154891454s (3.155034489s including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:13 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.containers{haproxy} Created Created container haproxy kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:13 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.containers{haproxy} Started Started container haproxy kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:13 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:35 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 102.227146ms (22.168467906s including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:35 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:35 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:35 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 22.170444762s (25.323407167s including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:35 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:35 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:35 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:36 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1 Scheduled Successfully assigned kuttl-test-real-hookworm/gr-demand-backup-haproxy-haproxy-1 to gke-jen-ps-523-f00253e-5-default-pool-6e6f9275-qcjw default-scheduler logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:36 +0000 UTC Normal StatefulSet.apps gr-demand-backup-haproxy-haproxy SuccessfulCreate create Pod gr-demand-backup-haproxy-haproxy-1 in StatefulSet gr-demand-backup-haproxy-haproxy successful statefulset-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:37 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:46 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 11.084106923s (11.084118659s including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:46 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:46 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:51 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 14.0695149s (14.069538797s including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:51 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.initContainers{haproxy-init} Created Created container haproxy-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:51 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.initContainers{haproxy-init} Started Started container haproxy-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:49:58 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:01 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 3.116198728s (3.116207415s including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:01 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.containers{haproxy} Created Created container haproxy kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:01 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.containers{haproxy} Started Started container haproxy kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:01 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:01 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 136.666032ms (136.683016ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:01 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:01 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:02 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2 Scheduled Successfully assigned kuttl-test-real-hookworm/gr-demand-backup-haproxy-haproxy-2 to gke-jen-ps-523-f00253e-5-default-pool-6e6f9275-77nf default-scheduler logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:02 +0000 UTC Normal StatefulSet.apps gr-demand-backup-haproxy-haproxy SuccessfulCreate create Pod gr-demand-backup-haproxy-haproxy-2 in StatefulSet gr-demand-backup-haproxy-haproxy successful statefulset-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:03 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:03 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 95.892703ms (95.908214ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:03 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.initContainers{haproxy-init} Created Created container haproxy-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:03 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.initContainers{haproxy-init} Started Started container haproxy-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:05 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:08 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 2.962416157s (2.962429589s including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:08 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.containers{haproxy} Created Created container haproxy kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:08 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.containers{haproxy} Started Started container haproxy kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:08 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:08 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 105.939297ms (105.95227ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:08 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:08 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:12 +0000 UTC Warning Pod gr-demand-backup-haproxy-mysql-1.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:12 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:12 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 108.535454ms (108.550563ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:43 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-haproxy-mysql-2 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:43 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-haproxy-mysql-2 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:43 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-haproxy-mysql-2 Provisioning External provisioner is provisioning volume for claim "kuttl-test-real-hookworm/datadir-gr-demand-backup-haproxy-mysql-2" pd.csi.storage.gke.io_gke-215547eb7cc0414a9857-296b-163f-vm_12151542-af00-4a77-bf44-27e93719a638 logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:43 +0000 UTC Normal StatefulSet.apps gr-demand-backup-haproxy-mysql SuccessfulCreate create Claim datadir-gr-demand-backup-haproxy-mysql-2 Pod gr-demand-backup-haproxy-mysql-2 in StatefulSet gr-demand-backup-haproxy-mysql success statefulset-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:43 +0000 UTC Normal StatefulSet.apps gr-demand-backup-haproxy-mysql SuccessfulCreate create Pod gr-demand-backup-haproxy-mysql-2 in StatefulSet gr-demand-backup-haproxy-mysql successful statefulset-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:47 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-haproxy-mysql-2 ProvisioningSucceeded Successfully provisioned volume pvc-bfe7215e-1895-4045-aee0-f66c44ea2f9b pd.csi.storage.gke.io_gke-215547eb7cc0414a9857-296b-163f-vm_12151542-af00-4a77-bf44-27e93719a638 logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:47 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2 Scheduled Successfully assigned kuttl-test-real-hookworm/gr-demand-backup-haproxy-mysql-2 to gke-jen-ps-523-f00253e-5-default-pool-6e6f9275-qcjw default-scheduler logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:55 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-bfe7215e-1895-4045-aee0-f66c44ea2f9b" attachdetach-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:57 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:57 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 110.873048ms (110.888097ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:57 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:57 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:50:59 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:51:20 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 21.407329755s (21.40734645s including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:51:20 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:51:20 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:51:20 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:51:31 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 10.567269996s (10.567286167s including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:51:31 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:51:31 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:51:55 +0000 UTC Warning Pod gr-demand-backup-haproxy-mysql-2.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:51:55 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:51:58 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 81.919397ms (81.954622ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:52:50 +0000 UTC Normal Pod xb-gr-demand-backup-haproxy-minio-minio-pnbbc Scheduled Successfully assigned kuttl-test-real-hookworm/xb-gr-demand-backup-haproxy-minio-minio-pnbbc to gke-jen-ps-523-f00253e-5-default-pool-6e6f9275-0brr default-scheduler logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:52:50 +0000 UTC Normal Job.batch xb-gr-demand-backup-haproxy-minio-minio SuccessfulCreate Created pod: xb-gr-demand-backup-haproxy-minio-minio-pnbbc job-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:52:51 +0000 UTC Normal Pod xb-gr-demand-backup-haproxy-minio-minio-pnbbc.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:52:51 +0000 UTC Normal Pod xb-gr-demand-backup-haproxy-minio-minio-pnbbc.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 89.456564ms (89.473942ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:52:51 +0000 UTC Normal Pod xb-gr-demand-backup-haproxy-minio-minio-pnbbc.spec.initContainers{xtrabackup-init} Created Created container xtrabackup-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:52:51 +0000 UTC Normal Pod xb-gr-demand-backup-haproxy-minio-minio-pnbbc.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:52:53 +0000 UTC Normal Pod xb-gr-demand-backup-haproxy-minio-minio-pnbbc.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:52:53 +0000 UTC Normal Pod xb-gr-demand-backup-haproxy-minio-minio-pnbbc.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 78.756089ms (78.766311ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:52:53 +0000 UTC Normal Pod xb-gr-demand-backup-haproxy-minio-minio-pnbbc.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:52:53 +0000 UTC Normal Pod xb-gr-demand-backup-haproxy-minio-minio-pnbbc.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:00 +0000 UTC Normal Job.batch xb-gr-demand-backup-haproxy-minio-minio Completed Job completed job-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:21 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.containers{haproxy} Killing Stopping container haproxy kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:21 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.containers{mysql-monit} Killing Stopping container mysql-monit kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:21 +0000 UTC Normal StatefulSet.apps gr-demand-backup-haproxy-haproxy SuccessfulDelete delete Pod gr-demand-backup-haproxy-haproxy-2 in StatefulSet gr-demand-backup-haproxy-haproxy successful statefulset-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:21 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.containers{xtrabackup} Killing Stopping container xtrabackup kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:21 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:21 +0000 UTC Normal StatefulSet.apps gr-demand-backup-haproxy-mysql SuccessfulDelete delete Pod gr-demand-backup-haproxy-mysql-2 in StatefulSet gr-demand-backup-haproxy-mysql successful statefulset-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:22 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.containers{haproxy} Killing Stopping container haproxy kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:22 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.containers{mysql-monit} Killing Stopping container mysql-monit kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:22 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.containers{haproxy} Killing Stopping container haproxy kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:22 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.containers{mysql-monit} Killing Stopping container mysql-monit kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:22 +0000 UTC Normal StatefulSet.apps gr-demand-backup-haproxy-haproxy SuccessfulDelete delete Pod gr-demand-backup-haproxy-haproxy-1 in StatefulSet gr-demand-backup-haproxy-haproxy successful statefulset-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:22 +0000 UTC Normal StatefulSet.apps gr-demand-backup-haproxy-haproxy SuccessfulDelete delete Pod gr-demand-backup-haproxy-haproxy-0 in StatefulSet gr-demand-backup-haproxy-haproxy successful statefulset-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:24 +0000 UTC Warning Pod gr-demand-backup-haproxy-mysql-2.spec.containers{mysql} Unhealthy Readiness probe failed: [mysql] 2024/02/06 08:53:24 packets.go:37: read tcp 10.65.178.10:55116->10.65.178.10:33062: read: connection reset by peer 2024/02/06 08:53:24 readiness check failed: connect to db: ping DB: dial tcp 10.65.178.10:33062: connect: connection refused kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:27 +0000 UTC Warning Pod gr-demand-backup-haproxy-mysql-2.spec.containers{mysql} Unhealthy Readiness probe failed: 2024/02/06 08:53:27 readiness check failed: connect to db: ping DB: dial tcp 10.65.178.10:33062: connect: connection refused kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:28 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.containers{xtrabackup} Killing Stopping container xtrabackup kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:28 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:28 +0000 UTC Normal StatefulSet.apps gr-demand-backup-haproxy-mysql SuccessfulDelete delete Pod gr-demand-backup-haproxy-mysql-1 in StatefulSet gr-demand-backup-haproxy-mysql successful statefulset-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:33 +0000 UTC Warning Pod gr-demand-backup-haproxy-mysql-1.spec.containers{mysql} Unhealthy Readiness probe failed: 2024/02/06 08:53:33 readiness check failed: connect to db: ping DB: dial tcp 10.65.176.9:33062: connect: connection refused kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:36 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:36 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.containers{xtrabackup} Killing Stopping container xtrabackup kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:36 +0000 UTC Normal StatefulSet.apps gr-demand-backup-haproxy-mysql SuccessfulDelete delete Pod gr-demand-backup-haproxy-mysql-0 in StatefulSet gr-demand-backup-haproxy-mysql successful statefulset-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:39 +0000 UTC Warning Pod gr-demand-backup-haproxy-mysql-0.spec.containers{mysql} Unhealthy Readiness probe failed: [mysql] 2024/02/06 08:53:39 packets.go:37: read tcp 10.65.177.5:52986->10.65.177.5:33062: read: connection reset by peer 2024/02/06 08:53:39 readiness check failed: connect to db: ping DB: dial tcp 10.65.177.5:33062: connect: connection refused kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:42 +0000 UTC Warning Pod gr-demand-backup-haproxy-mysql-0.spec.containers{mysql} Unhealthy Readiness probe errored: rpc error: code = Unknown desc = failed to exec in container: container is in CONTAINER_EXITED state kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:44 +0000 UTC Normal Pod xb-restore-gr-demand-backup-haproxy-restore-minio-96nzf Scheduled Successfully assigned kuttl-test-real-hookworm/xb-restore-gr-demand-backup-haproxy-restore-minio-96nzf to gke-jen-ps-523-f00253e-5-default-pool-6e6f9275-0brr default-scheduler logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:44 +0000 UTC Warning Pod xb-restore-gr-demand-backup-haproxy-restore-minio-96nzf FailedAttachVolume Multi-Attach error for volume "pvc-05875a35-edc2-4040-be55-ef6636277ccd" Volume is already exclusively attached to one node and can't be attached to another attachdetach-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:53:44 +0000 UTC Normal Job.batch xb-restore-gr-demand-backup-haproxy-restore-minio SuccessfulCreate Created pod: xb-restore-gr-demand-backup-haproxy-restore-minio-96nzf job-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:54:01 +0000 UTC Normal Pod xb-restore-gr-demand-backup-haproxy-restore-minio-96nzf SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-05875a35-edc2-4040-be55-ef6636277ccd" attachdetach-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:54:02 +0000 UTC Normal Pod xb-restore-gr-demand-backup-haproxy-restore-minio-96nzf.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:54:03 +0000 UTC Normal Pod xb-restore-gr-demand-backup-haproxy-restore-minio-96nzf.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 116.186192ms (116.201469ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:54:03 +0000 UTC Normal Pod xb-restore-gr-demand-backup-haproxy-restore-minio-96nzf.spec.initContainers{xtrabackup-init} Created Created container xtrabackup-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:54:03 +0000 UTC Normal Pod xb-restore-gr-demand-backup-haproxy-restore-minio-96nzf.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:54:04 +0000 UTC Normal Pod xb-restore-gr-demand-backup-haproxy-restore-minio-96nzf.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:54:04 +0000 UTC Normal Pod xb-restore-gr-demand-backup-haproxy-restore-minio-96nzf.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 94.272289ms (94.292555ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:54:04 +0000 UTC Normal Pod xb-restore-gr-demand-backup-haproxy-restore-minio-96nzf.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:54:05 +0000 UTC Normal Pod xb-restore-gr-demand-backup-haproxy-restore-minio-96nzf.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:54:14 +0000 UTC Normal Job.batch xb-restore-gr-demand-backup-haproxy-restore-minio Completed Job completed job-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:54:19 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0 Scheduled Successfully assigned kuttl-test-real-hookworm/gr-demand-backup-haproxy-mysql-0 to gke-jen-ps-523-f00253e-5-default-pool-6e6f9275-77nf default-scheduler logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:54:19 +0000 UTC Warning Pod gr-demand-backup-haproxy-mysql-0 FailedAttachVolume Multi-Attach error for volume "pvc-05875a35-edc2-4040-be55-ef6636277ccd" Volume is already exclusively attached to one node and can't be attached to another attachdetach-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:54:38 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-05875a35-edc2-4040-be55-ef6636277ccd" attachdetach-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:54:40 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:54:40 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 108.141327ms (108.17667ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:54:40 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:54:40 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:54:41 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:54:41 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 114.256437ms (114.273763ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:54:42 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:54:42 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:54:42 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:54:42 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 92.711954ms (92.761102ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:54:42 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:54:42 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:14 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-haproxy-mysql-1 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:14 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-haproxy-mysql-1 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:14 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-haproxy-mysql-1 Provisioning External provisioner is provisioning volume for claim "kuttl-test-real-hookworm/datadir-gr-demand-backup-haproxy-mysql-1" pd.csi.storage.gke.io_gke-215547eb7cc0414a9857-296b-163f-vm_12151542-af00-4a77-bf44-27e93719a638 logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:15 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0 Scheduled Successfully assigned kuttl-test-real-hookworm/gr-demand-backup-haproxy-haproxy-0 to gke-jen-ps-523-f00253e-5-default-pool-6e6f9275-0brr default-scheduler logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:16 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:16 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 143.263023ms (143.280171ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:16 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.initContainers{haproxy-init} Created Created container haproxy-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:16 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.initContainers{haproxy-init} Started Started container haproxy-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:18 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-haproxy-mysql-1 ProvisioningSucceeded Successfully provisioned volume pvc-6390a60f-b775-4578-a8d2-adb476123ece pd.csi.storage.gke.io_gke-215547eb7cc0414a9857-296b-163f-vm_12151542-af00-4a77-bf44-27e93719a638 logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:18 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:18 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 106.783588ms (106.799344ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:18 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.containers{haproxy} Created Created container haproxy kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:18 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.containers{haproxy} Started Started container haproxy kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:18 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:18 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 109.102806ms (109.136817ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:18 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:18 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-0.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:18 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1 Scheduled Successfully assigned kuttl-test-real-hookworm/gr-demand-backup-haproxy-mysql-1 to gke-jen-ps-523-f00253e-5-default-pool-6e6f9275-qcjw default-scheduler logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:19 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1 Scheduled Successfully assigned kuttl-test-real-hookworm/gr-demand-backup-haproxy-haproxy-1 to gke-jen-ps-523-f00253e-5-default-pool-6e6f9275-qcjw default-scheduler logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:19 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:20 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 118.787342ms (118.803498ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:20 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.initContainers{haproxy-init} Created Created container haproxy-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:20 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.initContainers{haproxy-init} Started Started container haproxy-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:22 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:22 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 110.797577ms (110.812537ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:22 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.containers{haproxy} Created Created container haproxy kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:22 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.containers{haproxy} Started Started container haproxy kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:22 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:22 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 94.006345ms (94.025022ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:22 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:22 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-1.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:23 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2 Scheduled Successfully assigned kuttl-test-real-hookworm/gr-demand-backup-haproxy-haproxy-2 to gke-jen-ps-523-f00253e-5-default-pool-6e6f9275-77nf default-scheduler logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:23 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:24 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 101.829797ms (101.850243ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:24 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.initContainers{haproxy-init} Created Created container haproxy-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:24 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.initContainers{haproxy-init} Started Started container haproxy-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:25 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:26 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 97.752295ms (97.767445ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:26 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.containers{haproxy} Created Created container haproxy kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:26 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.containers{haproxy} Started Started container haproxy kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:26 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:26 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 113.01452ms (113.03799ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:26 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:26 +0000 UTC Normal Pod gr-demand-backup-haproxy-haproxy-2.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:26 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-6390a60f-b775-4578-a8d2-adb476123ece" attachdetach-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:29 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:29 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 88.394826ms (88.414492ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:29 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:29 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:31 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:31 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 99.744382ms (99.77325ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:31 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:31 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:31 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:31 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 84.872099ms (84.89104ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:31 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:55:31 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:56:02 +0000 UTC Warning Pod gr-demand-backup-haproxy-mysql-1.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:56:02 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:56:02 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 90.290927ms (90.308842ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:56:34 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-haproxy-mysql-2 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:56:34 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-haproxy-mysql-2 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:56:34 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-haproxy-mysql-2 Provisioning External provisioner is provisioning volume for claim "kuttl-test-real-hookworm/datadir-gr-demand-backup-haproxy-mysql-2" pd.csi.storage.gke.io_gke-215547eb7cc0414a9857-296b-163f-vm_12151542-af00-4a77-bf44-27e93719a638 logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:56:38 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-haproxy-mysql-2 ProvisioningSucceeded Successfully provisioned volume pvc-8036f0d4-88a6-46db-b5e6-7b1d19adbf97 pd.csi.storage.gke.io_gke-215547eb7cc0414a9857-296b-163f-vm_12151542-af00-4a77-bf44-27e93719a638 logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:56:38 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2 Scheduled Successfully assigned kuttl-test-real-hookworm/gr-demand-backup-haproxy-mysql-2 to gke-jen-ps-523-f00253e-5-default-pool-6e6f9275-0brr default-scheduler logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:56:46 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-8036f0d4-88a6-46db-b5e6-7b1d19adbf97" attachdetach-controller logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:56:47 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:56:47 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 130.465634ms (130.480881ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:56:47 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:56:47 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:56:49 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:56:49 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 97.015236ms (97.022583ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:56:49 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:56:49 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:56:49 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:56:49 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 98.080116ms (98.087161ms including waiting) kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:56:49 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:56:49 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:57:21 +0000 UTC Warning Pod gr-demand-backup-haproxy-mysql-2.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:57:21 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 08:58:11 | gr-demand-backup-haproxy | 2024-02-06 08:57:21 +0000 UTC Normal Pod gr-demand-backup-haproxy-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 116.766176ms (116.781517ms including waiting) kubelet logger.go:42: 08:58:12 | gr-demand-backup-haproxy | Deleting namespace: kuttl-test-real-hookworm === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (773.74s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/gr-demand-backup-haproxy (765.26s) PASS