=== RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. I0205 14:06:10.629594 17589 request.go:682] Waited for 1.018521867s due to client-side throttling, not priority and fairness, request: GET:https://34.173.52.98/apis/rbac.authorization.k8s.io/v1?timeout=32s harness.go:275: Successful connection to cluster at: https://34.173.52.98 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 180 seconds for each step harness.go:372: testsuite: e2e-tests/tests has 30 tests === RUN kuttl/harness === RUN kuttl/harness/gr-demand-backup === PAUSE kuttl/harness/gr-demand-backup === CONT kuttl/harness/gr-demand-backup logger.go:42: 14:06:17 | gr-demand-backup | Creating namespace: kuttl-test-crack-yak logger.go:42: 14:06:17 | gr-demand-backup/0-minio-secret | starting test step 0-minio-secret logger.go:42: 14:06:20 | gr-demand-backup/0-minio-secret | Secret:kuttl-test-crack-yak/minio-secret created logger.go:42: 14:06:20 | gr-demand-backup/0-minio-secret | test step completed 0-minio-secret logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | starting test step 1-deploy-operator logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | running command: [sh -c set -o errexit set -o xtrace source ../../functions init_temp_dir # do this only in the first TestStep kubectl -n "${NAMESPACE}" apply -f "${TESTS_CONFIG_DIR}/cloud-secret.yml" deploy_operator deploy_non_tls_cluster_secrets deploy_tls_cluster_secrets deploy_client deploy_minio] logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | + source ../../functions logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ realpath ../../.. logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | ++++ pwd logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-demand-backup logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | ++ test_name=gr-demand-backup logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ export GIT_BRANCH=PR-523 logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ GIT_BRANCH=PR-523 logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ export VERSION=PR-523-f00253e logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ VERSION=PR-523-f00253e logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | ++++ which gdate logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | ++++ which date logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ date=/usr/bin/date logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ command -v oc logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ kubectl get nodes logger.go:42: 14:06:20 | gr-demand-backup/1-deploy-operator | +++ grep '^minikube' logger.go:42: 14:06:21 | gr-demand-backup/1-deploy-operator | E0205 14:06:21.720867 17738 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:21 | gr-demand-backup/1-deploy-operator | E0205 14:06:21.943425 17738 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:22 | gr-demand-backup/1-deploy-operator | E0205 14:06:22.055204 17738 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:22 | gr-demand-backup/1-deploy-operator | E0205 14:06:22.165686 17738 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:22 | gr-demand-backup/1-deploy-operator | + init_temp_dir logger.go:42: 14:06:22 | gr-demand-backup/1-deploy-operator | + rm -rf /tmp/kuttl/ps/gr-demand-backup logger.go:42: 14:06:22 | gr-demand-backup/1-deploy-operator | + mkdir -p /tmp/kuttl/ps/gr-demand-backup logger.go:42: 14:06:22 | gr-demand-backup/1-deploy-operator | + kubectl -n kuttl-test-crack-yak apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf/cloud-secret.yml logger.go:42: 14:06:23 | gr-demand-backup/1-deploy-operator | E0205 14:06:23.814748 17785 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:23 | gr-demand-backup/1-deploy-operator | E0205 14:06:23.929298 17785 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:24 | gr-demand-backup/1-deploy-operator | secret/aws-s3-secret created logger.go:42: 14:06:24 | gr-demand-backup/1-deploy-operator | secret/gcp-cs-secret created logger.go:42: 14:06:25 | gr-demand-backup/1-deploy-operator | secret/azure-secret created logger.go:42: 14:06:25 | gr-demand-backup/1-deploy-operator | + deploy_operator logger.go:42: 14:06:25 | gr-demand-backup/1-deploy-operator | + kubectl -n kuttl-test-crack-yak apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy/crd.yaml logger.go:42: 14:06:26 | gr-demand-backup/1-deploy-operator | E0205 14:06:26.299373 17826 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:26 | gr-demand-backup/1-deploy-operator | E0205 14:06:26.489737 17826 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:26 | gr-demand-backup/1-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlbackups.ps.percona.com serverside-applied logger.go:42: 14:06:27 | gr-demand-backup/1-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlrestores.ps.percona.com serverside-applied logger.go:42: 14:06:28 | gr-demand-backup/1-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqls.ps.percona.com serverside-applied logger.go:42: 14:06:28 | gr-demand-backup/1-deploy-operator | + kubectl -n kuttl-test-crack-yak apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy/rbac.yaml logger.go:42: 14:06:29 | gr-demand-backup/1-deploy-operator | E0205 14:06:29.361060 17867 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:29 | gr-demand-backup/1-deploy-operator | E0205 14:06:29.625865 17867 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:30 | gr-demand-backup/1-deploy-operator | serviceaccount/percona-server-mysql-operator created logger.go:42: 14:06:30 | gr-demand-backup/1-deploy-operator | serviceaccount/percona-server-mysql-operator-orchestrator created logger.go:42: 14:06:31 | gr-demand-backup/1-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 14:06:31 | gr-demand-backup/1-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 14:06:31 | gr-demand-backup/1-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 14:06:32 | gr-demand-backup/1-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 14:06:32 | gr-demand-backup/1-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 14:06:32 | gr-demand-backup/1-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 14:06:32 | gr-demand-backup/1-deploy-operator | + yq eval '(select(documentIndex==1).spec.template.spec.containers[] | select(.name=="manager").env[] | select(.name=="DISABLE_TELEMETRY").value) = "true"' logger.go:42: 14:06:32 | gr-demand-backup/1-deploy-operator | ++ printf 'select(documentIndex==1).spec.template.spec.containers[0].image="%s"' perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:06:32 | gr-demand-backup/1-deploy-operator | + yq eval '(select(documentIndex==1).spec.template.spec.containers[] | select(.name=="manager").env[] | select(.name=="LOG_LEVEL").value) = "DEBUG"' logger.go:42: 14:06:32 | gr-demand-backup/1-deploy-operator | + kubectl -n kuttl-test-crack-yak apply -f - logger.go:42: 14:06:32 | gr-demand-backup/1-deploy-operator | + yq eval 'select(documentIndex==1).spec.template.spec.containers[0].image="perconalab/percona-server-mysql-operator:PR-523-f00253e"' /mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy/operator.yaml logger.go:42: 14:06:33 | gr-demand-backup/1-deploy-operator | E0205 14:06:33.648346 17932 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:33 | gr-demand-backup/1-deploy-operator | E0205 14:06:33.871463 17932 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:34 | gr-demand-backup/1-deploy-operator | configmap/percona-server-mysql-operator-config created logger.go:42: 14:06:35 | gr-demand-backup/1-deploy-operator | deployment.apps/percona-server-mysql-operator created logger.go:42: 14:06:35 | gr-demand-backup/1-deploy-operator | + deploy_non_tls_cluster_secrets logger.go:42: 14:06:35 | gr-demand-backup/1-deploy-operator | + kubectl -n kuttl-test-crack-yak apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf/secrets.yaml logger.go:42: 14:06:36 | gr-demand-backup/1-deploy-operator | E0205 14:06:36.217553 17979 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:36 | gr-demand-backup/1-deploy-operator | E0205 14:06:36.336212 17979 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:36 | gr-demand-backup/1-deploy-operator | secret/test-secrets created logger.go:42: 14:06:36 | gr-demand-backup/1-deploy-operator | + deploy_tls_cluster_secrets logger.go:42: 14:06:36 | gr-demand-backup/1-deploy-operator | + kubectl -n kuttl-test-crack-yak apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf/ssl-secret.yaml logger.go:42: 14:06:37 | gr-demand-backup/1-deploy-operator | E0205 14:06:37.535083 18010 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:37 | gr-demand-backup/1-deploy-operator | E0205 14:06:37.755919 18010 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:38 | gr-demand-backup/1-deploy-operator | secret/test-ssl created logger.go:42: 14:06:38 | gr-demand-backup/1-deploy-operator | + deploy_client logger.go:42: 14:06:38 | gr-demand-backup/1-deploy-operator | + kubectl -n kuttl-test-crack-yak apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf/client.yaml logger.go:42: 14:06:39 | gr-demand-backup/1-deploy-operator | E0205 14:06:39.069050 18030 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:39 | gr-demand-backup/1-deploy-operator | E0205 14:06:39.179483 18030 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:39 | gr-demand-backup/1-deploy-operator | pod/mysql-client created logger.go:42: 14:06:39 | gr-demand-backup/1-deploy-operator | + deploy_minio logger.go:42: 14:06:39 | gr-demand-backup/1-deploy-operator | + local access_key logger.go:42: 14:06:39 | gr-demand-backup/1-deploy-operator | + local secret_key logger.go:42: 14:06:39 | gr-demand-backup/1-deploy-operator | ++ kubectl -n kuttl-test-crack-yak get secret minio-secret -o 'jsonpath={.data.AWS_ACCESS_KEY_ID}' logger.go:42: 14:06:39 | gr-demand-backup/1-deploy-operator | ++ base64 -d logger.go:42: 14:06:40 | gr-demand-backup/1-deploy-operator | E0205 14:06:40.252564 18053 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:40 | gr-demand-backup/1-deploy-operator | E0205 14:06:40.368177 18053 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:40 | gr-demand-backup/1-deploy-operator | E0205 14:06:40.479431 18053 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:40 | gr-demand-backup/1-deploy-operator | E0205 14:06:40.591059 18053 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:40 | gr-demand-backup/1-deploy-operator | + access_key='some-access$\n"-key' logger.go:42: 14:06:40 | gr-demand-backup/1-deploy-operator | ++ kubectl -n kuttl-test-crack-yak get secret minio-secret -o 'jsonpath={.data.AWS_SECRET_ACCESS_KEY}' logger.go:42: 14:06:40 | gr-demand-backup/1-deploy-operator | ++ base64 -d logger.go:42: 14:06:41 | gr-demand-backup/1-deploy-operator | E0205 14:06:41.363256 18084 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:41 | gr-demand-backup/1-deploy-operator | E0205 14:06:41.592590 18084 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:41 | gr-demand-backup/1-deploy-operator | E0205 14:06:41.710272 18084 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:41 | gr-demand-backup/1-deploy-operator | E0205 14:06:41.832970 18084 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:41 | gr-demand-backup/1-deploy-operator | + secret_key='some-$\n"secret-key' logger.go:42: 14:06:41 | gr-demand-backup/1-deploy-operator | + helm uninstall -n kuttl-test-crack-yak minio-service logger.go:42: 14:06:41 | gr-demand-backup/1-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 14:06:41 | gr-demand-backup/1-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 14:06:42 | gr-demand-backup/1-deploy-operator | Error: uninstall: Release not loaded: minio-service: release: not found logger.go:42: 14:06:42 | gr-demand-backup/1-deploy-operator | + : logger.go:42: 14:06:42 | gr-demand-backup/1-deploy-operator | + helm repo remove minio logger.go:42: 14:06:42 | gr-demand-backup/1-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 14:06:42 | gr-demand-backup/1-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 14:06:42 | gr-demand-backup/1-deploy-operator | Error: no repositories configured logger.go:42: 14:06:42 | gr-demand-backup/1-deploy-operator | + : logger.go:42: 14:06:42 | gr-demand-backup/1-deploy-operator | + helm repo add minio https://charts.min.io/ logger.go:42: 14:06:42 | gr-demand-backup/1-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 14:06:42 | gr-demand-backup/1-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 14:06:42 | gr-demand-backup/1-deploy-operator | "minio" has been added to your repositories logger.go:42: 14:06:42 | gr-demand-backup/1-deploy-operator | +++ printf %q 'some-access$\n"-key' logger.go:42: 14:06:42 | gr-demand-backup/1-deploy-operator | ++ printf %q 'some-access\$\\n\"-key' logger.go:42: 14:06:42 | gr-demand-backup/1-deploy-operator | +++ printf %q 'some-$\n"secret-key' logger.go:42: 14:06:42 | gr-demand-backup/1-deploy-operator | ++ printf %q 'some-\$\\n\"secret-key' logger.go:42: 14:06:42 | gr-demand-backup/1-deploy-operator | + retry 10 60 helm install minio-service -n kuttl-test-crack-yak --version 5.0.14 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access\\\$\\\\n\\\"-key' --set 'users[0].secretKey=some-\\\$\\\\n\\\"secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/.minio/ --set persistence.size=2G --set securityContext.enabled=false minio/minio logger.go:42: 14:06:42 | gr-demand-backup/1-deploy-operator | + local max=10 logger.go:42: 14:06:42 | gr-demand-backup/1-deploy-operator | + local delay=60 logger.go:42: 14:06:42 | gr-demand-backup/1-deploy-operator | + shift 2 logger.go:42: 14:06:42 | gr-demand-backup/1-deploy-operator | + local n=1 logger.go:42: 14:06:42 | gr-demand-backup/1-deploy-operator | + helm install minio-service -n kuttl-test-crack-yak --version 5.0.14 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access\\\$\\\\n\\\"-key' --set 'users[0].secretKey=some-\\\$\\\\n\\\"secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/.minio/ --set persistence.size=2G --set securityContext.enabled=false minio/minio logger.go:42: 14:06:42 | gr-demand-backup/1-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 14:06:42 | gr-demand-backup/1-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 14:06:43 | gr-demand-backup/1-deploy-operator | E0205 14:06:43.695262 18134 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:43 | gr-demand-backup/1-deploy-operator | E0205 14:06:43.961520 18134 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:46 | gr-demand-backup/1-deploy-operator | E0205 14:06:46.317115 18134 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:46 | gr-demand-backup/1-deploy-operator | E0205 14:06:46.431609 18134 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:46 | gr-demand-backup/1-deploy-operator | E0205 14:06:46.830893 18134 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:46 | gr-demand-backup/1-deploy-operator | E0205 14:06:46.951670 18134 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:47 | gr-demand-backup/1-deploy-operator | E0205 14:06:47.331754 18134 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:47 | gr-demand-backup/1-deploy-operator | E0205 14:06:47.445125 18134 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:47 | gr-demand-backup/1-deploy-operator | E0205 14:06:47.828158 18134 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:47 | gr-demand-backup/1-deploy-operator | E0205 14:06:47.940728 18134 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:48 | gr-demand-backup/1-deploy-operator | E0205 14:06:48.309351 18134 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:48 | gr-demand-backup/1-deploy-operator | E0205 14:06:48.426081 18134 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:48 | gr-demand-backup/1-deploy-operator | E0205 14:06:48.806430 18134 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:48 | gr-demand-backup/1-deploy-operator | E0205 14:06:48.922469 18134 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:49 | gr-demand-backup/1-deploy-operator | E0205 14:06:49.290774 18134 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:49 | gr-demand-backup/1-deploy-operator | E0205 14:06:49.405638 18134 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:51 | gr-demand-backup/1-deploy-operator | E0205 14:06:51.497849 18134 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:51 | gr-demand-backup/1-deploy-operator | E0205 14:06:51.611603 18134 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:52 | gr-demand-backup/1-deploy-operator | E0205 14:06:52.361639 18134 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:06:52 | gr-demand-backup/1-deploy-operator | E0205 14:06:52.476301 18134 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | NAME: minio-service logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | LAST DEPLOYED: Mon Feb 5 14:06:43 2024 logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | NAMESPACE: kuttl-test-crack-yak logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | STATUS: deployed logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | REVISION: 1 logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | TEST SUITE: None logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | NOTES: logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | MinIO can be accessed via port 9000 on the following DNS name from within your cluster: logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | minio-service.kuttl-test-crack-yak.svc.cluster.local logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | To access MinIO from localhost, run the below commands: logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | 1. export POD_NAME=$(kubectl get pods --namespace kuttl-test-crack-yak -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}") logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | 2. kubectl port-forward $POD_NAME 9000 --namespace kuttl-test-crack-yak logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/ logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | You can now access MinIO server on http://localhost:9000. Follow the below steps to connect to MinIO server with mc client: logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | 1. Download the MinIO mc client - https://min.io/docs/minio/linux/reference/minio-mc.html#quickstart logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | 2. export MC_HOST_minio-service-local=http://$(kubectl get secret --namespace kuttl-test-crack-yak minio-service -o jsonpath="{.data.rootUser}" | base64 --decode):$(kubectl get secret --namespace kuttl-test-crack-yak minio-service -o jsonpath="{.data.rootPassword}" | base64 --decode)@localhost:9000 logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | 3. mc ls minio-service-local logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | ++ kubectl -n kuttl-test-crack-yak get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | + MINIO_POD=minio-service-85cdcd4d44-v7kmg logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | + wait_pod minio-service-85cdcd4d44-v7kmg logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | + local pod=minio-service-85cdcd4d44-v7kmg logger.go:42: 14:07:39 | gr-demand-backup/1-deploy-operator | + set +o xtrace logger.go:42: 14:07:40 | gr-demand-backup/1-deploy-operator | minio-service-85cdcd4d44-v7kmgtrue logger.go:42: 14:07:40 | gr-demand-backup/1-deploy-operator | + kubectl -n kuttl-test-crack-yak run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- bash -c 'AWS_ACCESS_KEY_ID='\''some-access$\n"-key'\'' AWS_SECRET_ACCESS_KEY='\''some-$\n"secret-key'\'' AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url http://minio-service:9000 s3 mb s3://operator-testing' logger.go:42: 14:07:44 | gr-demand-backup/1-deploy-operator | If you don't see a command prompt, try pressing enter. logger.go:42: 14:07:49 | gr-demand-backup/1-deploy-operator | pod "aws-cli" deleted logger.go:42: 14:07:50 | gr-demand-backup/1-deploy-operator | test step completed 1-deploy-operator logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | starting test step 2-create-cluster logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | running command: [sh -c set -o errexit set -o xtrace source ../../functions get_cr \ | yq eval '.spec.backup.storages.minio.type="s3"' - \ | yq eval '.spec.backup.storages.minio.s3.bucket="operator-testing"' - \ | yq eval '.spec.backup.storages.minio.s3.credentialsSecret="minio-secret"' - \ | yq eval '.spec.backup.storages.minio.s3.endpointUrl="http://minio-service:9000"' - \ | yq eval '.spec.backup.storages.minio.s3.region="us-east-1"' - \ | yq eval '.spec.backup.storages.aws-s3.type="s3"' - \ | yq eval '.spec.backup.storages.aws-s3.verifyTLS=true' - \ | yq eval '.spec.backup.storages.aws-s3.s3.bucket="operator-testing"' - \ | yq eval '.spec.backup.storages.aws-s3.s3.credentialsSecret="aws-s3-secret"' - \ | yq eval '.spec.backup.storages.aws-s3.s3.region="us-east-1"' - \ | yq eval '.spec.backup.storages.aws-s3.s3.prefix="ps"' - \ | yq eval '.spec.backup.storages.gcp-cs.type="gcs"' - \ | yq eval '.spec.backup.storages.gcp-cs.verifyTLS=true' - \ | yq eval '.spec.backup.storages.gcp-cs.gcs.bucket="operator-testing"' - \ | yq eval '.spec.backup.storages.gcp-cs.gcs.credentialsSecret="gcp-cs-secret"' - \ | yq eval '.spec.backup.storages.gcp-cs.gcs.endpointUrl="https://storage.googleapis.com"' - \ | yq eval '.spec.backup.storages.gcp-cs.gcs.prefix="ps"' - \ | yq eval '.spec.backup.storages.azure-blob.type="azure"' - \ | yq eval '.spec.backup.storages.azure-blob.verifyTLS=true' - \ | yq eval '.spec.backup.storages.azure-blob.azure.containerName="operator-testing"' - \ | yq eval '.spec.backup.storages.azure-blob.azure.credentialsSecret="azure-secret"' - \ | yq eval '.spec.backup.storages.azure-blob.azure.prefix="ps"' - \ | yq eval '.spec.mysql.clusterType="group-replication"' - \ | yq eval '.spec.proxy.router.enabled=true' - \ | yq eval '.spec.proxy.haproxy.enabled=false' - \ | kubectl -n "${NAMESPACE}" apply -f -] logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + source ../../functions logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ realpath ../../.. logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | ++++ pwd logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-demand-backup logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | ++ test_name=gr-demand-backup logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ export GIT_BRANCH=PR-523 logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ GIT_BRANCH=PR-523 logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ export VERSION=PR-523-f00253e logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ VERSION=PR-523-f00253e logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | ++++ which gdate logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | ++++ which date logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ date=/usr/bin/date logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ command -v oc logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ kubectl get nodes logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | +++ grep '^minikube' logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.minio.type="s3"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + get_cr logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.minio.s3.bucket="operator-testing"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + local name_suffix= logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.minio.s3.credentialsSecret="minio-secret"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.aws-s3.s3.credentialsSecret="aws-s3-secret"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.aws-s3.s3.region="us-east-1"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.aws-s3.s3.prefix="ps"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.gcp-cs.type="gcs"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.gcp-cs.gcs.bucket="operator-testing"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.aws-s3.type="s3"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.gcp-cs.gcs.credentialsSecret="gcp-cs-secret"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.secretsName="test-secrets"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval .spec.proxy.haproxy.enabled=false - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.upgradeOptions.apply="disabled"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + kubectl -n kuttl-test-crack-yak apply -f - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | ++ printf '.spec.backup.image="%s"' perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.gcp-cs.gcs.endpointUrl="https://storage.googleapis.com"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.image="perconalab/percona-server-mysql-operator:main-backup"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + '[' -n '' ']' logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval .spec.backup.storages.gcp-cs.verifyTLS=true - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.gcp-cs.gcs.prefix="ps"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | ++ printf '.spec.mysql.image="%s"' perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.mysql.image="perconalab/percona-server-mysql-operator:main-psmysql"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.azure-blob.type="azure"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.azure-blob.azure.credentialsSecret="azure-secret"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.aws-s3.s3.bucket="operator-testing"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | ++ printf '.spec.proxy.router.image="%s"' perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.azure-blob.azure.prefix="ps"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.proxy.router.image="perconalab/percona-server-mysql-operator:main-router"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval .spec.proxy.router.enabled=true - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval .spec.backup.storages.azure-blob.verifyTLS=true - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.sslSecretName="test-ssl"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | ++ printf '.spec.toolkit.image="%s"' perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.minio.s3.region="us-east-1"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.toolkit.image="perconalab/percona-server-mysql-operator:main-toolkit"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.azure-blob.azure.containerName="operator-testing"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.mysql.clusterType="group-replication"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval .spec.backup.storages.aws-s3.verifyTLS=true - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | ++ printf '.spec.orchestrator.image="%s"' perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.orchestrator.image="perconalab/percona-server-mysql-operator:main-orchestrator"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | ++ printf '.spec.pmm.image="%s"' perconalab/pmm-client:dev-latest logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | ++ printf '.spec.proxy.haproxy.image="%s"' perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.proxy.haproxy.image="perconalab/percona-server-mysql-operator:main-haproxy"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.pmm.image="perconalab/pmm-client:dev-latest"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.minio.s3.endpointUrl="http://minio-service:9000"' - logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | ++ printf '.spec.initImage="%s"' perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:07:50 | gr-demand-backup/2-create-cluster | + yq eval '.spec.initImage="perconalab/percona-server-mysql-operator:PR-523-f00253e"' - logger.go:42: 14:07:51 | gr-demand-backup/2-create-cluster | ++ printf '.metadata.name="%s"' gr-demand-backup logger.go:42: 14:07:51 | gr-demand-backup/2-create-cluster | + yq eval '.metadata.name="gr-demand-backup"' /mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy/cr.yaml logger.go:42: 14:07:52 | gr-demand-backup/2-create-cluster | perconaservermysql.ps.percona.com/gr-demand-backup created logger.go:42: 14:13:46 | gr-demand-backup/2-create-cluster | test step completed 2-create-cluster logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | starting test step 3-write-data logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | running command: [sh -c set -o errexit set -o pipefail set -o xtrace source ../../functions run_mysql \ "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" \ "-h $(get_mysql_router_service $(get_cluster_name)) -uroot -proot_password" run_mysql \ "INSERT myDB.myTable (id) VALUES (100500)" \ "-h $(get_mysql_router_service $(get_cluster_name)) -uroot -proot_password"] logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | + source ../../functions logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ realpath ../../.. logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | ++++ pwd logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-demand-backup logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | ++ test_name=gr-demand-backup logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ export GIT_BRANCH=PR-523 logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ GIT_BRANCH=PR-523 logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ export VERSION=PR-523-f00253e logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ VERSION=PR-523-f00253e logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | ++++ which gdate logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | ++++ which date logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ date=/usr/bin/date logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ command -v oc logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ kubectl get nodes logger.go:42: 14:13:46 | gr-demand-backup/3-write-data | +++ grep '^minikube' logger.go:42: 14:13:47 | gr-demand-backup/3-write-data | +++ get_cluster_name logger.go:42: 14:13:47 | gr-demand-backup/3-write-data | +++ kubectl -n kuttl-test-crack-yak get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:13:47 | gr-demand-backup/3-write-data | ++ get_mysql_router_service gr-demand-backup logger.go:42: 14:13:47 | gr-demand-backup/3-write-data | ++ local cluster=gr-demand-backup logger.go:42: 14:13:47 | gr-demand-backup/3-write-data | ++ echo gr-demand-backup-router logger.go:42: 14:13:47 | gr-demand-backup/3-write-data | + run_mysql 'CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' '-h gr-demand-backup-router -uroot -proot_password' logger.go:42: 14:13:47 | gr-demand-backup/3-write-data | + local 'command=CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' logger.go:42: 14:13:47 | gr-demand-backup/3-write-data | + local 'uri=-h gr-demand-backup-router -uroot -proot_password' logger.go:42: 14:13:47 | gr-demand-backup/3-write-data | + local pod= logger.go:42: 14:13:47 | gr-demand-backup/3-write-data | ++ get_client_pod logger.go:42: 14:13:47 | gr-demand-backup/3-write-data | ++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:13:48 | gr-demand-backup/3-write-data | + client_pod=mysql-client logger.go:42: 14:13:48 | gr-demand-backup/3-write-data | + wait_pod mysql-client logger.go:42: 14:13:48 | gr-demand-backup/3-write-data | + local pod=mysql-client logger.go:42: 14:13:48 | gr-demand-backup/3-write-data | + set +o xtrace logger.go:42: 14:13:48 | gr-demand-backup/3-write-data | mysql-clienttrue logger.go:42: 14:13:48 | gr-demand-backup/3-write-data | + sed -e 's/mysql: //' logger.go:42: 14:13:48 | gr-demand-backup/3-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:13:48 | gr-demand-backup/3-write-data | + kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" | mysql -sN -h gr-demand-backup-router -uroot -proot_password' logger.go:42: 14:13:49 | gr-demand-backup/3-write-data | + : logger.go:42: 14:13:49 | gr-demand-backup/3-write-data | +++ get_cluster_name logger.go:42: 14:13:49 | gr-demand-backup/3-write-data | +++ kubectl -n kuttl-test-crack-yak get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:13:50 | gr-demand-backup/3-write-data | ++ get_mysql_router_service gr-demand-backup logger.go:42: 14:13:50 | gr-demand-backup/3-write-data | ++ local cluster=gr-demand-backup logger.go:42: 14:13:50 | gr-demand-backup/3-write-data | ++ echo gr-demand-backup-router logger.go:42: 14:13:50 | gr-demand-backup/3-write-data | + run_mysql 'INSERT myDB.myTable (id) VALUES (100500)' '-h gr-demand-backup-router -uroot -proot_password' logger.go:42: 14:13:50 | gr-demand-backup/3-write-data | + local 'command=INSERT myDB.myTable (id) VALUES (100500)' logger.go:42: 14:13:50 | gr-demand-backup/3-write-data | + local 'uri=-h gr-demand-backup-router -uroot -proot_password' logger.go:42: 14:13:50 | gr-demand-backup/3-write-data | + local pod= logger.go:42: 14:13:50 | gr-demand-backup/3-write-data | ++ get_client_pod logger.go:42: 14:13:50 | gr-demand-backup/3-write-data | ++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:13:50 | gr-demand-backup/3-write-data | + client_pod=mysql-client logger.go:42: 14:13:50 | gr-demand-backup/3-write-data | + wait_pod mysql-client logger.go:42: 14:13:50 | gr-demand-backup/3-write-data | + local pod=mysql-client logger.go:42: 14:13:50 | gr-demand-backup/3-write-data | + set +o xtrace logger.go:42: 14:13:50 | gr-demand-backup/3-write-data | mysql-clienttrue logger.go:42: 14:13:50 | gr-demand-backup/3-write-data | + kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "INSERT myDB.myTable (id) VALUES (100500)" | mysql -sN -h gr-demand-backup-router -uroot -proot_password' logger.go:42: 14:13:50 | gr-demand-backup/3-write-data | + sed -e 's/mysql: //' logger.go:42: 14:13:50 | gr-demand-backup/3-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:13:52 | gr-demand-backup/3-write-data | + : logger.go:42: 14:13:52 | gr-demand-backup/3-write-data | test step completed 3-write-data logger.go:42: 14:13:52 | gr-demand-backup/4-create-backup-minio | starting test step 4-create-backup-minio logger.go:42: 14:13:54 | gr-demand-backup/4-create-backup-minio | PerconaServerMySQLBackup:kuttl-test-crack-yak/gr-demand-backup-minio created logger.go:42: 14:14:09 | gr-demand-backup/4-create-backup-minio | test step completed 4-create-backup-minio logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | starting test step 5-delete-data logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "TRUNCATE TABLE myDB.myTable" \ "-h $(get_mysql_router_service $(get_cluster_name)) -uroot -proot_password" cluster_name=$(get_cluster_name) for i in 0 1 2; do data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 04-delete-data-minio-${i} --from-literal=data="${data}" done] logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | + source ../../functions logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ realpath ../../.. logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | ++++ pwd logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-demand-backup logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | ++ test_name=gr-demand-backup logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ export GIT_BRANCH=PR-523 logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ GIT_BRANCH=PR-523 logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ export VERSION=PR-523-f00253e logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ VERSION=PR-523-f00253e logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | ++++ which gdate logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | ++++ which date logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ date=/usr/bin/date logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ command -v oc logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ kubectl get nodes logger.go:42: 14:14:09 | gr-demand-backup/5-delete-data | +++ grep '^minikube' logger.go:42: 14:14:10 | gr-demand-backup/5-delete-data | +++ get_cluster_name logger.go:42: 14:14:10 | gr-demand-backup/5-delete-data | +++ kubectl -n kuttl-test-crack-yak get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:14:10 | gr-demand-backup/5-delete-data | ++ get_mysql_router_service gr-demand-backup logger.go:42: 14:14:10 | gr-demand-backup/5-delete-data | ++ local cluster=gr-demand-backup logger.go:42: 14:14:10 | gr-demand-backup/5-delete-data | ++ echo gr-demand-backup-router logger.go:42: 14:14:10 | gr-demand-backup/5-delete-data | + run_mysql 'TRUNCATE TABLE myDB.myTable' '-h gr-demand-backup-router -uroot -proot_password' logger.go:42: 14:14:10 | gr-demand-backup/5-delete-data | + local 'command=TRUNCATE TABLE myDB.myTable' logger.go:42: 14:14:10 | gr-demand-backup/5-delete-data | + local 'uri=-h gr-demand-backup-router -uroot -proot_password' logger.go:42: 14:14:10 | gr-demand-backup/5-delete-data | + local pod= logger.go:42: 14:14:10 | gr-demand-backup/5-delete-data | ++ get_client_pod logger.go:42: 14:14:10 | gr-demand-backup/5-delete-data | ++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:14:11 | gr-demand-backup/5-delete-data | + client_pod=mysql-client logger.go:42: 14:14:11 | gr-demand-backup/5-delete-data | + wait_pod mysql-client logger.go:42: 14:14:11 | gr-demand-backup/5-delete-data | + local pod=mysql-client logger.go:42: 14:14:11 | gr-demand-backup/5-delete-data | + set +o xtrace logger.go:42: 14:14:11 | gr-demand-backup/5-delete-data | mysql-clienttrue logger.go:42: 14:14:11 | gr-demand-backup/5-delete-data | + kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "TRUNCATE TABLE myDB.myTable" | mysql -sN -h gr-demand-backup-router -uroot -proot_password' logger.go:42: 14:14:11 | gr-demand-backup/5-delete-data | + sed -e 's/mysql: //' logger.go:42: 14:14:11 | gr-demand-backup/5-delete-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:14:12 | gr-demand-backup/5-delete-data | + : logger.go:42: 14:14:12 | gr-demand-backup/5-delete-data | ++ get_cluster_name logger.go:42: 14:14:12 | gr-demand-backup/5-delete-data | ++ kubectl -n kuttl-test-crack-yak get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:14:13 | gr-demand-backup/5-delete-data | + cluster_name=gr-demand-backup logger.go:42: 14:14:13 | gr-demand-backup/5-delete-data | + for i in 0 1 2 logger.go:42: 14:14:13 | gr-demand-backup/5-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:14:13 | gr-demand-backup/5-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:14:13 | gr-demand-backup/5-delete-data | ++ local 'uri=-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:14:13 | gr-demand-backup/5-delete-data | ++ local pod= logger.go:42: 14:14:13 | gr-demand-backup/5-delete-data | +++ get_client_pod logger.go:42: 14:14:13 | gr-demand-backup/5-delete-data | +++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:14:13 | gr-demand-backup/5-delete-data | ++ client_pod=mysql-client logger.go:42: 14:14:13 | gr-demand-backup/5-delete-data | ++ wait_pod mysql-client logger.go:42: 14:14:13 | gr-demand-backup/5-delete-data | ++ local pod=mysql-client logger.go:42: 14:14:13 | gr-demand-backup/5-delete-data | ++ set +o xtrace logger.go:42: 14:14:14 | gr-demand-backup/5-delete-data | mysql-clienttrue logger.go:42: 14:14:14 | gr-demand-backup/5-delete-data | ++ kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:14:14 | gr-demand-backup/5-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 14:14:14 | gr-demand-backup/5-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:14:15 | gr-demand-backup/5-delete-data | ++ : logger.go:42: 14:14:15 | gr-demand-backup/5-delete-data | + data= logger.go:42: 14:14:15 | gr-demand-backup/5-delete-data | + kubectl create configmap -n kuttl-test-crack-yak 04-delete-data-minio-0 --from-literal=data= logger.go:42: 14:14:15 | gr-demand-backup/5-delete-data | configmap/04-delete-data-minio-0 created logger.go:42: 14:14:15 | gr-demand-backup/5-delete-data | + for i in 0 1 2 logger.go:42: 14:14:15 | gr-demand-backup/5-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:14:15 | gr-demand-backup/5-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:14:15 | gr-demand-backup/5-delete-data | ++ local 'uri=-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:14:15 | gr-demand-backup/5-delete-data | ++ local pod= logger.go:42: 14:14:15 | gr-demand-backup/5-delete-data | +++ get_client_pod logger.go:42: 14:14:15 | gr-demand-backup/5-delete-data | +++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:14:16 | gr-demand-backup/5-delete-data | ++ client_pod=mysql-client logger.go:42: 14:14:16 | gr-demand-backup/5-delete-data | ++ wait_pod mysql-client logger.go:42: 14:14:16 | gr-demand-backup/5-delete-data | ++ local pod=mysql-client logger.go:42: 14:14:16 | gr-demand-backup/5-delete-data | ++ set +o xtrace logger.go:42: 14:14:16 | gr-demand-backup/5-delete-data | mysql-clienttrue logger.go:42: 14:14:16 | gr-demand-backup/5-delete-data | ++ kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:14:16 | gr-demand-backup/5-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 14:14:16 | gr-demand-backup/5-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:14:17 | gr-demand-backup/5-delete-data | ++ : logger.go:42: 14:14:17 | gr-demand-backup/5-delete-data | + data= logger.go:42: 14:14:17 | gr-demand-backup/5-delete-data | + kubectl create configmap -n kuttl-test-crack-yak 04-delete-data-minio-1 --from-literal=data= logger.go:42: 14:14:18 | gr-demand-backup/5-delete-data | configmap/04-delete-data-minio-1 created logger.go:42: 14:14:18 | gr-demand-backup/5-delete-data | + for i in 0 1 2 logger.go:42: 14:14:18 | gr-demand-backup/5-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:14:18 | gr-demand-backup/5-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:14:18 | gr-demand-backup/5-delete-data | ++ local 'uri=-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:14:18 | gr-demand-backup/5-delete-data | ++ local pod= logger.go:42: 14:14:18 | gr-demand-backup/5-delete-data | +++ get_client_pod logger.go:42: 14:14:18 | gr-demand-backup/5-delete-data | +++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:14:18 | gr-demand-backup/5-delete-data | ++ client_pod=mysql-client logger.go:42: 14:14:18 | gr-demand-backup/5-delete-data | ++ wait_pod mysql-client logger.go:42: 14:14:18 | gr-demand-backup/5-delete-data | ++ local pod=mysql-client logger.go:42: 14:14:18 | gr-demand-backup/5-delete-data | ++ set +o xtrace logger.go:42: 14:14:18 | gr-demand-backup/5-delete-data | mysql-clienttrue logger.go:42: 14:14:18 | gr-demand-backup/5-delete-data | ++ kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:14:18 | gr-demand-backup/5-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 14:14:18 | gr-demand-backup/5-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:14:20 | gr-demand-backup/5-delete-data | ++ : logger.go:42: 14:14:20 | gr-demand-backup/5-delete-data | + data= logger.go:42: 14:14:20 | gr-demand-backup/5-delete-data | + kubectl create configmap -n kuttl-test-crack-yak 04-delete-data-minio-2 --from-literal=data= logger.go:42: 14:14:20 | gr-demand-backup/5-delete-data | configmap/04-delete-data-minio-2 created logger.go:42: 14:14:21 | gr-demand-backup/5-delete-data | test step completed 5-delete-data logger.go:42: 14:14:21 | gr-demand-backup/6-restore-from-minio | starting test step 6-restore-from-minio logger.go:42: 14:14:22 | gr-demand-backup/6-restore-from-minio | PerconaServerMySQLRestore:kuttl-test-crack-yak/gr-demand-backup-restore-minio created logger.go:42: 14:19:21 | gr-demand-backup/6-restore-from-minio | test step completed 6-restore-from-minio logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | starting test step 7-read-data logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions cluster_name=$(get_cluster_name) for i in 0 1 2; do data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 06-read-data-minio-${i} --from-literal=data="${data}" done] logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | + source ../../functions logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ realpath ../../.. logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | ++++ pwd logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-demand-backup logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | ++ test_name=gr-demand-backup logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ export GIT_BRANCH=PR-523 logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ GIT_BRANCH=PR-523 logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ export VERSION=PR-523-f00253e logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ VERSION=PR-523-f00253e logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | ++++ which gdate logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | ++++ which date logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ date=/usr/bin/date logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ command -v oc logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ kubectl get nodes logger.go:42: 14:19:21 | gr-demand-backup/7-read-data | +++ grep '^minikube' logger.go:42: 14:19:22 | gr-demand-backup/7-read-data | ++ get_cluster_name logger.go:42: 14:19:22 | gr-demand-backup/7-read-data | ++ kubectl -n kuttl-test-crack-yak get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:19:22 | gr-demand-backup/7-read-data | + cluster_name=gr-demand-backup logger.go:42: 14:19:22 | gr-demand-backup/7-read-data | + for i in 0 1 2 logger.go:42: 14:19:22 | gr-demand-backup/7-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:19:22 | gr-demand-backup/7-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:19:22 | gr-demand-backup/7-read-data | ++ local 'uri=-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:19:22 | gr-demand-backup/7-read-data | ++ local pod= logger.go:42: 14:19:22 | gr-demand-backup/7-read-data | +++ get_client_pod logger.go:42: 14:19:22 | gr-demand-backup/7-read-data | +++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:19:23 | gr-demand-backup/7-read-data | ++ client_pod=mysql-client logger.go:42: 14:19:23 | gr-demand-backup/7-read-data | ++ wait_pod mysql-client logger.go:42: 14:19:23 | gr-demand-backup/7-read-data | ++ local pod=mysql-client logger.go:42: 14:19:23 | gr-demand-backup/7-read-data | ++ set +o xtrace logger.go:42: 14:19:23 | gr-demand-backup/7-read-data | mysql-clienttrue logger.go:42: 14:19:23 | gr-demand-backup/7-read-data | ++ kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:19:23 | gr-demand-backup/7-read-data | ++ sed -e 's/mysql: //' logger.go:42: 14:19:23 | gr-demand-backup/7-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:19:24 | gr-demand-backup/7-read-data | + data=100500 logger.go:42: 14:19:24 | gr-demand-backup/7-read-data | + kubectl create configmap -n kuttl-test-crack-yak 06-read-data-minio-0 --from-literal=data=100500 logger.go:42: 14:19:25 | gr-demand-backup/7-read-data | configmap/06-read-data-minio-0 created logger.go:42: 14:19:25 | gr-demand-backup/7-read-data | + for i in 0 1 2 logger.go:42: 14:19:25 | gr-demand-backup/7-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:19:25 | gr-demand-backup/7-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:19:25 | gr-demand-backup/7-read-data | ++ local 'uri=-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:19:25 | gr-demand-backup/7-read-data | ++ local pod= logger.go:42: 14:19:25 | gr-demand-backup/7-read-data | +++ get_client_pod logger.go:42: 14:19:25 | gr-demand-backup/7-read-data | +++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:19:25 | gr-demand-backup/7-read-data | ++ client_pod=mysql-client logger.go:42: 14:19:25 | gr-demand-backup/7-read-data | ++ wait_pod mysql-client logger.go:42: 14:19:25 | gr-demand-backup/7-read-data | ++ local pod=mysql-client logger.go:42: 14:19:25 | gr-demand-backup/7-read-data | ++ set +o xtrace logger.go:42: 14:19:26 | gr-demand-backup/7-read-data | mysql-clienttrue logger.go:42: 14:19:26 | gr-demand-backup/7-read-data | ++ kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:19:26 | gr-demand-backup/7-read-data | ++ sed -e 's/mysql: //' logger.go:42: 14:19:26 | gr-demand-backup/7-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:19:27 | gr-demand-backup/7-read-data | + data=100500 logger.go:42: 14:19:27 | gr-demand-backup/7-read-data | + kubectl create configmap -n kuttl-test-crack-yak 06-read-data-minio-1 --from-literal=data=100500 logger.go:42: 14:19:27 | gr-demand-backup/7-read-data | configmap/06-read-data-minio-1 created logger.go:42: 14:19:27 | gr-demand-backup/7-read-data | + for i in 0 1 2 logger.go:42: 14:19:27 | gr-demand-backup/7-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:19:27 | gr-demand-backup/7-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:19:27 | gr-demand-backup/7-read-data | ++ local 'uri=-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:19:27 | gr-demand-backup/7-read-data | ++ local pod= logger.go:42: 14:19:27 | gr-demand-backup/7-read-data | +++ get_client_pod logger.go:42: 14:19:27 | gr-demand-backup/7-read-data | +++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:19:28 | gr-demand-backup/7-read-data | ++ client_pod=mysql-client logger.go:42: 14:19:28 | gr-demand-backup/7-read-data | ++ wait_pod mysql-client logger.go:42: 14:19:28 | gr-demand-backup/7-read-data | ++ local pod=mysql-client logger.go:42: 14:19:28 | gr-demand-backup/7-read-data | ++ set +o xtrace logger.go:42: 14:19:28 | gr-demand-backup/7-read-data | mysql-clienttrue logger.go:42: 14:19:28 | gr-demand-backup/7-read-data | ++ kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:19:28 | gr-demand-backup/7-read-data | ++ sed -e 's/mysql: //' logger.go:42: 14:19:28 | gr-demand-backup/7-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:19:29 | gr-demand-backup/7-read-data | + data=100500 logger.go:42: 14:19:29 | gr-demand-backup/7-read-data | + kubectl create configmap -n kuttl-test-crack-yak 06-read-data-minio-2 --from-literal=data=100500 logger.go:42: 14:19:30 | gr-demand-backup/7-read-data | configmap/06-read-data-minio-2 created logger.go:42: 14:19:31 | gr-demand-backup/7-read-data | test step completed 7-read-data logger.go:42: 14:19:31 | gr-demand-backup/8-create-backup-s3 | starting test step 8-create-backup-s3 logger.go:42: 14:19:32 | gr-demand-backup/8-create-backup-s3 | PerconaServerMySQLBackup:kuttl-test-crack-yak/gr-demand-backup-s3 created logger.go:42: 14:19:44 | gr-demand-backup/8-create-backup-s3 | test step completed 8-create-backup-s3 logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | starting test step 9-delete-data logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "TRUNCATE TABLE myDB.myTable" \ "-h $(get_mysql_router_service $(get_cluster_name)) -uroot -proot_password" cluster_name=$(get_cluster_name) for i in 0 1 2; do data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 08-delete-data-s3-${i} --from-literal=data="${data}" done] logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | + source ../../functions logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ realpath ../../.. logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | ++++ pwd logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-demand-backup logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | ++ test_name=gr-demand-backup logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ export GIT_BRANCH=PR-523 logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ GIT_BRANCH=PR-523 logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ export VERSION=PR-523-f00253e logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ VERSION=PR-523-f00253e logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | ++++ which gdate logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | ++++ which date logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ date=/usr/bin/date logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ command -v oc logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ kubectl get nodes logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ grep '^minikube' logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ get_cluster_name logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | +++ kubectl -n kuttl-test-crack-yak get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | ++ get_mysql_router_service gr-demand-backup logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | ++ local cluster=gr-demand-backup logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | ++ echo gr-demand-backup-router logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | + run_mysql 'TRUNCATE TABLE myDB.myTable' '-h gr-demand-backup-router -uroot -proot_password' logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | + local 'command=TRUNCATE TABLE myDB.myTable' logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | + local 'uri=-h gr-demand-backup-router -uroot -proot_password' logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | + local pod= logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | ++ get_client_pod logger.go:42: 14:19:44 | gr-demand-backup/9-delete-data | ++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:19:45 | gr-demand-backup/9-delete-data | + client_pod=mysql-client logger.go:42: 14:19:45 | gr-demand-backup/9-delete-data | + wait_pod mysql-client logger.go:42: 14:19:45 | gr-demand-backup/9-delete-data | + local pod=mysql-client logger.go:42: 14:19:45 | gr-demand-backup/9-delete-data | + set +o xtrace logger.go:42: 14:19:45 | gr-demand-backup/9-delete-data | mysql-clienttrue logger.go:42: 14:19:45 | gr-demand-backup/9-delete-data | + kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "TRUNCATE TABLE myDB.myTable" | mysql -sN -h gr-demand-backup-router -uroot -proot_password' logger.go:42: 14:19:45 | gr-demand-backup/9-delete-data | + sed -e 's/mysql: //' logger.go:42: 14:19:45 | gr-demand-backup/9-delete-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:19:46 | gr-demand-backup/9-delete-data | + : logger.go:42: 14:19:46 | gr-demand-backup/9-delete-data | ++ get_cluster_name logger.go:42: 14:19:46 | gr-demand-backup/9-delete-data | ++ kubectl -n kuttl-test-crack-yak get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:19:47 | gr-demand-backup/9-delete-data | + cluster_name=gr-demand-backup logger.go:42: 14:19:47 | gr-demand-backup/9-delete-data | + for i in 0 1 2 logger.go:42: 14:19:47 | gr-demand-backup/9-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:19:47 | gr-demand-backup/9-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:19:47 | gr-demand-backup/9-delete-data | ++ local 'uri=-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:19:47 | gr-demand-backup/9-delete-data | ++ local pod= logger.go:42: 14:19:47 | gr-demand-backup/9-delete-data | +++ get_client_pod logger.go:42: 14:19:47 | gr-demand-backup/9-delete-data | +++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:19:47 | gr-demand-backup/9-delete-data | ++ client_pod=mysql-client logger.go:42: 14:19:47 | gr-demand-backup/9-delete-data | ++ wait_pod mysql-client logger.go:42: 14:19:47 | gr-demand-backup/9-delete-data | ++ local pod=mysql-client logger.go:42: 14:19:47 | gr-demand-backup/9-delete-data | ++ set +o xtrace logger.go:42: 14:19:48 | gr-demand-backup/9-delete-data | mysql-clienttrue logger.go:42: 14:19:48 | gr-demand-backup/9-delete-data | ++ kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:19:48 | gr-demand-backup/9-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 14:19:48 | gr-demand-backup/9-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:19:49 | gr-demand-backup/9-delete-data | ++ : logger.go:42: 14:19:49 | gr-demand-backup/9-delete-data | + data= logger.go:42: 14:19:49 | gr-demand-backup/9-delete-data | + kubectl create configmap -n kuttl-test-crack-yak 08-delete-data-s3-0 --from-literal=data= logger.go:42: 14:19:49 | gr-demand-backup/9-delete-data | configmap/08-delete-data-s3-0 created logger.go:42: 14:19:49 | gr-demand-backup/9-delete-data | + for i in 0 1 2 logger.go:42: 14:19:49 | gr-demand-backup/9-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:19:49 | gr-demand-backup/9-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:19:49 | gr-demand-backup/9-delete-data | ++ local 'uri=-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:19:49 | gr-demand-backup/9-delete-data | ++ local pod= logger.go:42: 14:19:49 | gr-demand-backup/9-delete-data | +++ get_client_pod logger.go:42: 14:19:49 | gr-demand-backup/9-delete-data | +++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:19:50 | gr-demand-backup/9-delete-data | ++ client_pod=mysql-client logger.go:42: 14:19:50 | gr-demand-backup/9-delete-data | ++ wait_pod mysql-client logger.go:42: 14:19:50 | gr-demand-backup/9-delete-data | ++ local pod=mysql-client logger.go:42: 14:19:50 | gr-demand-backup/9-delete-data | ++ set +o xtrace logger.go:42: 14:19:50 | gr-demand-backup/9-delete-data | mysql-clienttrue logger.go:42: 14:19:50 | gr-demand-backup/9-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 14:19:50 | gr-demand-backup/9-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:19:50 | gr-demand-backup/9-delete-data | ++ kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:19:51 | gr-demand-backup/9-delete-data | ++ : logger.go:42: 14:19:51 | gr-demand-backup/9-delete-data | + data= logger.go:42: 14:19:51 | gr-demand-backup/9-delete-data | + kubectl create configmap -n kuttl-test-crack-yak 08-delete-data-s3-1 --from-literal=data= logger.go:42: 14:19:52 | gr-demand-backup/9-delete-data | configmap/08-delete-data-s3-1 created logger.go:42: 14:19:52 | gr-demand-backup/9-delete-data | + for i in 0 1 2 logger.go:42: 14:19:52 | gr-demand-backup/9-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:19:52 | gr-demand-backup/9-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:19:52 | gr-demand-backup/9-delete-data | ++ local 'uri=-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:19:52 | gr-demand-backup/9-delete-data | ++ local pod= logger.go:42: 14:19:52 | gr-demand-backup/9-delete-data | +++ get_client_pod logger.go:42: 14:19:52 | gr-demand-backup/9-delete-data | +++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:19:52 | gr-demand-backup/9-delete-data | ++ client_pod=mysql-client logger.go:42: 14:19:52 | gr-demand-backup/9-delete-data | ++ wait_pod mysql-client logger.go:42: 14:19:52 | gr-demand-backup/9-delete-data | ++ local pod=mysql-client logger.go:42: 14:19:52 | gr-demand-backup/9-delete-data | ++ set +o xtrace logger.go:42: 14:19:52 | gr-demand-backup/9-delete-data | mysql-clienttrue logger.go:42: 14:19:52 | gr-demand-backup/9-delete-data | ++ kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:19:52 | gr-demand-backup/9-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 14:19:52 | gr-demand-backup/9-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:19:54 | gr-demand-backup/9-delete-data | ++ : logger.go:42: 14:19:54 | gr-demand-backup/9-delete-data | + data= logger.go:42: 14:19:54 | gr-demand-backup/9-delete-data | + kubectl create configmap -n kuttl-test-crack-yak 08-delete-data-s3-2 --from-literal=data= logger.go:42: 14:19:54 | gr-demand-backup/9-delete-data | configmap/08-delete-data-s3-2 created logger.go:42: 14:19:55 | gr-demand-backup/9-delete-data | test step completed 9-delete-data logger.go:42: 14:19:55 | gr-demand-backup/10-restore-from-s3 | starting test step 10-restore-from-s3 logger.go:42: 14:19:57 | gr-demand-backup/10-restore-from-s3 | PerconaServerMySQLRestore:kuttl-test-crack-yak/gr-demand-backup-restore-s3 created logger.go:42: 14:25:01 | gr-demand-backup/10-restore-from-s3 | test step completed 10-restore-from-s3 logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | starting test step 11-read-data logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions cluster_name=$(get_cluster_name) for i in 0 1 2; do data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 06-read-data-s3-${i} --from-literal=data="${data}" done] logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | + source ../../functions logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ realpath ../../.. logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | ++++ pwd logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-demand-backup logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | ++ test_name=gr-demand-backup logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ export GIT_BRANCH=PR-523 logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ GIT_BRANCH=PR-523 logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ export VERSION=PR-523-f00253e logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ VERSION=PR-523-f00253e logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | ++++ which gdate logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | ++++ which date logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ date=/usr/bin/date logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ command -v oc logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ kubectl get nodes logger.go:42: 14:25:01 | gr-demand-backup/11-read-data | +++ grep '^minikube' logger.go:42: 14:25:02 | gr-demand-backup/11-read-data | ++ get_cluster_name logger.go:42: 14:25:02 | gr-demand-backup/11-read-data | ++ kubectl -n kuttl-test-crack-yak get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:25:02 | gr-demand-backup/11-read-data | + cluster_name=gr-demand-backup logger.go:42: 14:25:02 | gr-demand-backup/11-read-data | + for i in 0 1 2 logger.go:42: 14:25:02 | gr-demand-backup/11-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:25:02 | gr-demand-backup/11-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:25:02 | gr-demand-backup/11-read-data | ++ local 'uri=-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:25:02 | gr-demand-backup/11-read-data | ++ local pod= logger.go:42: 14:25:02 | gr-demand-backup/11-read-data | +++ get_client_pod logger.go:42: 14:25:02 | gr-demand-backup/11-read-data | +++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:25:03 | gr-demand-backup/11-read-data | ++ client_pod=mysql-client logger.go:42: 14:25:03 | gr-demand-backup/11-read-data | ++ wait_pod mysql-client logger.go:42: 14:25:03 | gr-demand-backup/11-read-data | ++ local pod=mysql-client logger.go:42: 14:25:03 | gr-demand-backup/11-read-data | ++ set +o xtrace logger.go:42: 14:25:03 | gr-demand-backup/11-read-data | mysql-clienttrue logger.go:42: 14:25:03 | gr-demand-backup/11-read-data | ++ kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:25:03 | gr-demand-backup/11-read-data | ++ sed -e 's/mysql: //' logger.go:42: 14:25:03 | gr-demand-backup/11-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:25:04 | gr-demand-backup/11-read-data | + data=100500 logger.go:42: 14:25:04 | gr-demand-backup/11-read-data | + kubectl create configmap -n kuttl-test-crack-yak 06-read-data-s3-0 --from-literal=data=100500 logger.go:42: 14:25:05 | gr-demand-backup/11-read-data | configmap/06-read-data-s3-0 created logger.go:42: 14:25:05 | gr-demand-backup/11-read-data | + for i in 0 1 2 logger.go:42: 14:25:05 | gr-demand-backup/11-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:25:05 | gr-demand-backup/11-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:25:05 | gr-demand-backup/11-read-data | ++ local 'uri=-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:25:05 | gr-demand-backup/11-read-data | ++ local pod= logger.go:42: 14:25:05 | gr-demand-backup/11-read-data | +++ get_client_pod logger.go:42: 14:25:05 | gr-demand-backup/11-read-data | +++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:25:05 | gr-demand-backup/11-read-data | ++ client_pod=mysql-client logger.go:42: 14:25:05 | gr-demand-backup/11-read-data | ++ wait_pod mysql-client logger.go:42: 14:25:05 | gr-demand-backup/11-read-data | ++ local pod=mysql-client logger.go:42: 14:25:05 | gr-demand-backup/11-read-data | ++ set +o xtrace logger.go:42: 14:25:05 | gr-demand-backup/11-read-data | mysql-clienttrue logger.go:42: 14:25:05 | gr-demand-backup/11-read-data | ++ sed -e 's/mysql: //' logger.go:42: 14:25:05 | gr-demand-backup/11-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:25:05 | gr-demand-backup/11-read-data | ++ kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:25:07 | gr-demand-backup/11-read-data | + data=100500 logger.go:42: 14:25:07 | gr-demand-backup/11-read-data | + kubectl create configmap -n kuttl-test-crack-yak 06-read-data-s3-1 --from-literal=data=100500 logger.go:42: 14:25:07 | gr-demand-backup/11-read-data | configmap/06-read-data-s3-1 created logger.go:42: 14:25:07 | gr-demand-backup/11-read-data | + for i in 0 1 2 logger.go:42: 14:25:07 | gr-demand-backup/11-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:25:07 | gr-demand-backup/11-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:25:07 | gr-demand-backup/11-read-data | ++ local 'uri=-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:25:07 | gr-demand-backup/11-read-data | ++ local pod= logger.go:42: 14:25:07 | gr-demand-backup/11-read-data | +++ get_client_pod logger.go:42: 14:25:07 | gr-demand-backup/11-read-data | +++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:25:07 | gr-demand-backup/11-read-data | ++ client_pod=mysql-client logger.go:42: 14:25:07 | gr-demand-backup/11-read-data | ++ wait_pod mysql-client logger.go:42: 14:25:07 | gr-demand-backup/11-read-data | ++ local pod=mysql-client logger.go:42: 14:25:07 | gr-demand-backup/11-read-data | ++ set +o xtrace logger.go:42: 14:25:08 | gr-demand-backup/11-read-data | mysql-clienttrue logger.go:42: 14:25:08 | gr-demand-backup/11-read-data | ++ kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:25:08 | gr-demand-backup/11-read-data | ++ sed -e 's/mysql: //' logger.go:42: 14:25:08 | gr-demand-backup/11-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:25:09 | gr-demand-backup/11-read-data | + data=100500 logger.go:42: 14:25:09 | gr-demand-backup/11-read-data | + kubectl create configmap -n kuttl-test-crack-yak 06-read-data-s3-2 --from-literal=data=100500 logger.go:42: 14:25:09 | gr-demand-backup/11-read-data | configmap/06-read-data-s3-2 created logger.go:42: 14:25:11 | gr-demand-backup/11-read-data | test step completed 11-read-data logger.go:42: 14:25:11 | gr-demand-backup/12-create-backup-gcp | starting test step 12-create-backup-gcp logger.go:42: 14:25:12 | gr-demand-backup/12-create-backup-gcp | PerconaServerMySQLBackup:kuttl-test-crack-yak/gr-demand-backup-gcp created logger.go:42: 14:25:29 | gr-demand-backup/12-create-backup-gcp | test step completed 12-create-backup-gcp logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | starting test step 13-delete-data logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "TRUNCATE TABLE myDB.myTable" \ "-h $(get_mysql_router_service $(get_cluster_name)) -uroot -proot_password" cluster_name=$(get_cluster_name) for i in 0 1 2; do data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 12-delete-data-gcp-${i} --from-literal=data="${data}" done] logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | + source ../../functions logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ realpath ../../.. logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | ++++ pwd logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-demand-backup logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | ++ test_name=gr-demand-backup logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ export GIT_BRANCH=PR-523 logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ GIT_BRANCH=PR-523 logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ export VERSION=PR-523-f00253e logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ VERSION=PR-523-f00253e logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | ++++ which gdate logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | ++++ which date logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ date=/usr/bin/date logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ command -v oc logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ kubectl get nodes logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ grep '^minikube' logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ get_cluster_name logger.go:42: 14:25:29 | gr-demand-backup/13-delete-data | +++ kubectl -n kuttl-test-crack-yak get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:25:30 | gr-demand-backup/13-delete-data | ++ get_mysql_router_service gr-demand-backup logger.go:42: 14:25:30 | gr-demand-backup/13-delete-data | ++ local cluster=gr-demand-backup logger.go:42: 14:25:30 | gr-demand-backup/13-delete-data | ++ echo gr-demand-backup-router logger.go:42: 14:25:30 | gr-demand-backup/13-delete-data | + run_mysql 'TRUNCATE TABLE myDB.myTable' '-h gr-demand-backup-router -uroot -proot_password' logger.go:42: 14:25:30 | gr-demand-backup/13-delete-data | + local 'command=TRUNCATE TABLE myDB.myTable' logger.go:42: 14:25:30 | gr-demand-backup/13-delete-data | + local 'uri=-h gr-demand-backup-router -uroot -proot_password' logger.go:42: 14:25:30 | gr-demand-backup/13-delete-data | + local pod= logger.go:42: 14:25:30 | gr-demand-backup/13-delete-data | ++ get_client_pod logger.go:42: 14:25:30 | gr-demand-backup/13-delete-data | ++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:25:30 | gr-demand-backup/13-delete-data | + client_pod=mysql-client logger.go:42: 14:25:30 | gr-demand-backup/13-delete-data | + wait_pod mysql-client logger.go:42: 14:25:30 | gr-demand-backup/13-delete-data | + local pod=mysql-client logger.go:42: 14:25:30 | gr-demand-backup/13-delete-data | + set +o xtrace logger.go:42: 14:25:31 | gr-demand-backup/13-delete-data | mysql-clienttrue logger.go:42: 14:25:31 | gr-demand-backup/13-delete-data | + kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "TRUNCATE TABLE myDB.myTable" | mysql -sN -h gr-demand-backup-router -uroot -proot_password' logger.go:42: 14:25:31 | gr-demand-backup/13-delete-data | + sed -e 's/mysql: //' logger.go:42: 14:25:31 | gr-demand-backup/13-delete-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:25:32 | gr-demand-backup/13-delete-data | + : logger.go:42: 14:25:32 | gr-demand-backup/13-delete-data | ++ get_cluster_name logger.go:42: 14:25:32 | gr-demand-backup/13-delete-data | ++ kubectl -n kuttl-test-crack-yak get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:25:32 | gr-demand-backup/13-delete-data | + cluster_name=gr-demand-backup logger.go:42: 14:25:32 | gr-demand-backup/13-delete-data | + for i in 0 1 2 logger.go:42: 14:25:32 | gr-demand-backup/13-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:25:32 | gr-demand-backup/13-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:25:32 | gr-demand-backup/13-delete-data | ++ local 'uri=-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:25:32 | gr-demand-backup/13-delete-data | ++ local pod= logger.go:42: 14:25:32 | gr-demand-backup/13-delete-data | +++ get_client_pod logger.go:42: 14:25:32 | gr-demand-backup/13-delete-data | +++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:25:33 | gr-demand-backup/13-delete-data | ++ client_pod=mysql-client logger.go:42: 14:25:33 | gr-demand-backup/13-delete-data | ++ wait_pod mysql-client logger.go:42: 14:25:33 | gr-demand-backup/13-delete-data | ++ local pod=mysql-client logger.go:42: 14:25:33 | gr-demand-backup/13-delete-data | ++ set +o xtrace logger.go:42: 14:25:33 | gr-demand-backup/13-delete-data | mysql-clienttrue logger.go:42: 14:25:33 | gr-demand-backup/13-delete-data | ++ kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:25:33 | gr-demand-backup/13-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 14:25:33 | gr-demand-backup/13-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:25:34 | gr-demand-backup/13-delete-data | ++ : logger.go:42: 14:25:34 | gr-demand-backup/13-delete-data | + data= logger.go:42: 14:25:34 | gr-demand-backup/13-delete-data | + kubectl create configmap -n kuttl-test-crack-yak 12-delete-data-gcp-0 --from-literal=data= logger.go:42: 14:25:35 | gr-demand-backup/13-delete-data | configmap/12-delete-data-gcp-0 created logger.go:42: 14:25:35 | gr-demand-backup/13-delete-data | + for i in 0 1 2 logger.go:42: 14:25:35 | gr-demand-backup/13-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:25:35 | gr-demand-backup/13-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:25:35 | gr-demand-backup/13-delete-data | ++ local 'uri=-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:25:35 | gr-demand-backup/13-delete-data | ++ local pod= logger.go:42: 14:25:35 | gr-demand-backup/13-delete-data | +++ get_client_pod logger.go:42: 14:25:35 | gr-demand-backup/13-delete-data | +++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:25:35 | gr-demand-backup/13-delete-data | ++ client_pod=mysql-client logger.go:42: 14:25:35 | gr-demand-backup/13-delete-data | ++ wait_pod mysql-client logger.go:42: 14:25:35 | gr-demand-backup/13-delete-data | ++ local pod=mysql-client logger.go:42: 14:25:35 | gr-demand-backup/13-delete-data | ++ set +o xtrace logger.go:42: 14:25:35 | gr-demand-backup/13-delete-data | mysql-clienttrue logger.go:42: 14:25:35 | gr-demand-backup/13-delete-data | ++ kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:25:35 | gr-demand-backup/13-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 14:25:35 | gr-demand-backup/13-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:25:37 | gr-demand-backup/13-delete-data | ++ : logger.go:42: 14:25:37 | gr-demand-backup/13-delete-data | + data= logger.go:42: 14:25:37 | gr-demand-backup/13-delete-data | + kubectl create configmap -n kuttl-test-crack-yak 12-delete-data-gcp-1 --from-literal=data= logger.go:42: 14:25:37 | gr-demand-backup/13-delete-data | configmap/12-delete-data-gcp-1 created logger.go:42: 14:25:37 | gr-demand-backup/13-delete-data | + for i in 0 1 2 logger.go:42: 14:25:37 | gr-demand-backup/13-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:25:37 | gr-demand-backup/13-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:25:37 | gr-demand-backup/13-delete-data | ++ local 'uri=-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:25:37 | gr-demand-backup/13-delete-data | ++ local pod= logger.go:42: 14:25:37 | gr-demand-backup/13-delete-data | +++ get_client_pod logger.go:42: 14:25:37 | gr-demand-backup/13-delete-data | +++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:25:37 | gr-demand-backup/13-delete-data | ++ client_pod=mysql-client logger.go:42: 14:25:37 | gr-demand-backup/13-delete-data | ++ wait_pod mysql-client logger.go:42: 14:25:37 | gr-demand-backup/13-delete-data | ++ local pod=mysql-client logger.go:42: 14:25:37 | gr-demand-backup/13-delete-data | ++ set +o xtrace logger.go:42: 14:25:38 | gr-demand-backup/13-delete-data | mysql-clienttrue logger.go:42: 14:25:38 | gr-demand-backup/13-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 14:25:38 | gr-demand-backup/13-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:25:38 | gr-demand-backup/13-delete-data | ++ kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:25:39 | gr-demand-backup/13-delete-data | ++ : logger.go:42: 14:25:39 | gr-demand-backup/13-delete-data | + data= logger.go:42: 14:25:39 | gr-demand-backup/13-delete-data | + kubectl create configmap -n kuttl-test-crack-yak 12-delete-data-gcp-2 --from-literal=data= logger.go:42: 14:25:39 | gr-demand-backup/13-delete-data | configmap/12-delete-data-gcp-2 created logger.go:42: 14:25:41 | gr-demand-backup/13-delete-data | test step completed 13-delete-data logger.go:42: 14:25:41 | gr-demand-backup/14-restore-from-gcp | starting test step 14-restore-from-gcp logger.go:42: 14:25:42 | gr-demand-backup/14-restore-from-gcp | PerconaServerMySQLRestore:kuttl-test-crack-yak/gr-demand-backup-restore-gcp created logger.go:42: 14:30:35 | gr-demand-backup/14-restore-from-gcp | test step completed 14-restore-from-gcp logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | starting test step 15-read-data logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions cluster_name=$(get_cluster_name) for i in 0 1 2; do data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 06-read-data-gcp-${i} --from-literal=data="${data}" done] logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | + source ../../functions logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ realpath ../../.. logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | ++++ pwd logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-demand-backup logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | ++ test_name=gr-demand-backup logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ export GIT_BRANCH=PR-523 logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ GIT_BRANCH=PR-523 logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ export VERSION=PR-523-f00253e logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ VERSION=PR-523-f00253e logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | ++++ which gdate logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | ++++ which date logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ date=/usr/bin/date logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ command -v oc logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ kubectl get nodes logger.go:42: 14:30:35 | gr-demand-backup/15-read-data | +++ grep '^minikube' logger.go:42: 14:30:36 | gr-demand-backup/15-read-data | ++ get_cluster_name logger.go:42: 14:30:36 | gr-demand-backup/15-read-data | ++ kubectl -n kuttl-test-crack-yak get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:30:36 | gr-demand-backup/15-read-data | + cluster_name=gr-demand-backup logger.go:42: 14:30:36 | gr-demand-backup/15-read-data | + for i in 0 1 2 logger.go:42: 14:30:36 | gr-demand-backup/15-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:30:36 | gr-demand-backup/15-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:30:36 | gr-demand-backup/15-read-data | ++ local 'uri=-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:30:36 | gr-demand-backup/15-read-data | ++ local pod= logger.go:42: 14:30:36 | gr-demand-backup/15-read-data | +++ get_client_pod logger.go:42: 14:30:36 | gr-demand-backup/15-read-data | +++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:30:37 | gr-demand-backup/15-read-data | ++ client_pod=mysql-client logger.go:42: 14:30:37 | gr-demand-backup/15-read-data | ++ wait_pod mysql-client logger.go:42: 14:30:37 | gr-demand-backup/15-read-data | ++ local pod=mysql-client logger.go:42: 14:30:37 | gr-demand-backup/15-read-data | ++ set +o xtrace logger.go:42: 14:30:37 | gr-demand-backup/15-read-data | mysql-clienttrue logger.go:42: 14:30:37 | gr-demand-backup/15-read-data | ++ kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:30:37 | gr-demand-backup/15-read-data | ++ sed -e 's/mysql: //' logger.go:42: 14:30:37 | gr-demand-backup/15-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:30:38 | gr-demand-backup/15-read-data | + data=100500 logger.go:42: 14:30:38 | gr-demand-backup/15-read-data | + kubectl create configmap -n kuttl-test-crack-yak 06-read-data-gcp-0 --from-literal=data=100500 logger.go:42: 14:30:38 | gr-demand-backup/15-read-data | configmap/06-read-data-gcp-0 created logger.go:42: 14:30:38 | gr-demand-backup/15-read-data | + for i in 0 1 2 logger.go:42: 14:30:38 | gr-demand-backup/15-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:30:38 | gr-demand-backup/15-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:30:38 | gr-demand-backup/15-read-data | ++ local 'uri=-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:30:38 | gr-demand-backup/15-read-data | ++ local pod= logger.go:42: 14:30:38 | gr-demand-backup/15-read-data | +++ get_client_pod logger.go:42: 14:30:38 | gr-demand-backup/15-read-data | +++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:30:39 | gr-demand-backup/15-read-data | ++ client_pod=mysql-client logger.go:42: 14:30:39 | gr-demand-backup/15-read-data | ++ wait_pod mysql-client logger.go:42: 14:30:39 | gr-demand-backup/15-read-data | ++ local pod=mysql-client logger.go:42: 14:30:39 | gr-demand-backup/15-read-data | ++ set +o xtrace logger.go:42: 14:30:39 | gr-demand-backup/15-read-data | mysql-clienttrue logger.go:42: 14:30:39 | gr-demand-backup/15-read-data | ++ kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:30:39 | gr-demand-backup/15-read-data | ++ sed -e 's/mysql: //' logger.go:42: 14:30:39 | gr-demand-backup/15-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:30:41 | gr-demand-backup/15-read-data | + data=100500 logger.go:42: 14:30:41 | gr-demand-backup/15-read-data | + kubectl create configmap -n kuttl-test-crack-yak 06-read-data-gcp-1 --from-literal=data=100500 logger.go:42: 14:30:41 | gr-demand-backup/15-read-data | configmap/06-read-data-gcp-1 created logger.go:42: 14:30:41 | gr-demand-backup/15-read-data | + for i in 0 1 2 logger.go:42: 14:30:41 | gr-demand-backup/15-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:30:41 | gr-demand-backup/15-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:30:41 | gr-demand-backup/15-read-data | ++ local 'uri=-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:30:41 | gr-demand-backup/15-read-data | ++ local pod= logger.go:42: 14:30:41 | gr-demand-backup/15-read-data | +++ get_client_pod logger.go:42: 14:30:41 | gr-demand-backup/15-read-data | +++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:30:41 | gr-demand-backup/15-read-data | ++ client_pod=mysql-client logger.go:42: 14:30:41 | gr-demand-backup/15-read-data | ++ wait_pod mysql-client logger.go:42: 14:30:41 | gr-demand-backup/15-read-data | ++ local pod=mysql-client logger.go:42: 14:30:41 | gr-demand-backup/15-read-data | ++ set +o xtrace logger.go:42: 14:30:42 | gr-demand-backup/15-read-data | mysql-clienttrue logger.go:42: 14:30:42 | gr-demand-backup/15-read-data | ++ kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:30:42 | gr-demand-backup/15-read-data | ++ sed -e 's/mysql: //' logger.go:42: 14:30:42 | gr-demand-backup/15-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:30:43 | gr-demand-backup/15-read-data | + data=100500 logger.go:42: 14:30:43 | gr-demand-backup/15-read-data | + kubectl create configmap -n kuttl-test-crack-yak 06-read-data-gcp-2 --from-literal=data=100500 logger.go:42: 14:30:43 | gr-demand-backup/15-read-data | configmap/06-read-data-gcp-2 created logger.go:42: 14:30:45 | gr-demand-backup/15-read-data | test step completed 15-read-data logger.go:42: 14:30:45 | gr-demand-backup/16-create-backup-azure | starting test step 16-create-backup-azure logger.go:42: 14:30:46 | gr-demand-backup/16-create-backup-azure | PerconaServerMySQLBackup:kuttl-test-crack-yak/gr-demand-backup-azure created logger.go:42: 14:30:57 | gr-demand-backup/16-create-backup-azure | test step completed 16-create-backup-azure logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | starting test step 17-delete-data logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "TRUNCATE TABLE myDB.myTable" \ "-h $(get_mysql_router_service $(get_cluster_name)) -uroot -proot_password" cluster_name=$(get_cluster_name) for i in 0 1 2; do data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 16-delete-data-azure-${i} --from-literal=data="${data}" done] logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | + source ../../functions logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ realpath ../../.. logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | ++++ pwd logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-demand-backup logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | ++ test_name=gr-demand-backup logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ export GIT_BRANCH=PR-523 logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ GIT_BRANCH=PR-523 logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ export VERSION=PR-523-f00253e logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ VERSION=PR-523-f00253e logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | ++++ which gdate logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | ++++ which date logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ date=/usr/bin/date logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ command -v oc logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ kubectl get nodes logger.go:42: 14:30:57 | gr-demand-backup/17-delete-data | +++ grep '^minikube' logger.go:42: 14:30:58 | gr-demand-backup/17-delete-data | +++ get_cluster_name logger.go:42: 14:30:58 | gr-demand-backup/17-delete-data | +++ kubectl -n kuttl-test-crack-yak get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:30:58 | gr-demand-backup/17-delete-data | ++ get_mysql_router_service gr-demand-backup logger.go:42: 14:30:58 | gr-demand-backup/17-delete-data | ++ local cluster=gr-demand-backup logger.go:42: 14:30:58 | gr-demand-backup/17-delete-data | ++ echo gr-demand-backup-router logger.go:42: 14:30:58 | gr-demand-backup/17-delete-data | + run_mysql 'TRUNCATE TABLE myDB.myTable' '-h gr-demand-backup-router -uroot -proot_password' logger.go:42: 14:30:58 | gr-demand-backup/17-delete-data | + local 'command=TRUNCATE TABLE myDB.myTable' logger.go:42: 14:30:58 | gr-demand-backup/17-delete-data | + local 'uri=-h gr-demand-backup-router -uroot -proot_password' logger.go:42: 14:30:58 | gr-demand-backup/17-delete-data | + local pod= logger.go:42: 14:30:58 | gr-demand-backup/17-delete-data | ++ get_client_pod logger.go:42: 14:30:58 | gr-demand-backup/17-delete-data | ++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:30:59 | gr-demand-backup/17-delete-data | + client_pod=mysql-client logger.go:42: 14:30:59 | gr-demand-backup/17-delete-data | + wait_pod mysql-client logger.go:42: 14:30:59 | gr-demand-backup/17-delete-data | + local pod=mysql-client logger.go:42: 14:30:59 | gr-demand-backup/17-delete-data | + set +o xtrace logger.go:42: 14:30:59 | gr-demand-backup/17-delete-data | mysql-clienttrue logger.go:42: 14:30:59 | gr-demand-backup/17-delete-data | + kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "TRUNCATE TABLE myDB.myTable" | mysql -sN -h gr-demand-backup-router -uroot -proot_password' logger.go:42: 14:30:59 | gr-demand-backup/17-delete-data | + sed -e 's/mysql: //' logger.go:42: 14:30:59 | gr-demand-backup/17-delete-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:31:00 | gr-demand-backup/17-delete-data | + : logger.go:42: 14:31:00 | gr-demand-backup/17-delete-data | ++ get_cluster_name logger.go:42: 14:31:00 | gr-demand-backup/17-delete-data | ++ kubectl -n kuttl-test-crack-yak get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:31:01 | gr-demand-backup/17-delete-data | + cluster_name=gr-demand-backup logger.go:42: 14:31:01 | gr-demand-backup/17-delete-data | + for i in 0 1 2 logger.go:42: 14:31:01 | gr-demand-backup/17-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:31:01 | gr-demand-backup/17-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:31:01 | gr-demand-backup/17-delete-data | ++ local 'uri=-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:31:01 | gr-demand-backup/17-delete-data | ++ local pod= logger.go:42: 14:31:01 | gr-demand-backup/17-delete-data | +++ get_client_pod logger.go:42: 14:31:01 | gr-demand-backup/17-delete-data | +++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:31:01 | gr-demand-backup/17-delete-data | ++ client_pod=mysql-client logger.go:42: 14:31:01 | gr-demand-backup/17-delete-data | ++ wait_pod mysql-client logger.go:42: 14:31:01 | gr-demand-backup/17-delete-data | ++ local pod=mysql-client logger.go:42: 14:31:01 | gr-demand-backup/17-delete-data | ++ set +o xtrace logger.go:42: 14:31:02 | gr-demand-backup/17-delete-data | mysql-clienttrue logger.go:42: 14:31:02 | gr-demand-backup/17-delete-data | ++ kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:31:02 | gr-demand-backup/17-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 14:31:02 | gr-demand-backup/17-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:31:03 | gr-demand-backup/17-delete-data | ++ : logger.go:42: 14:31:03 | gr-demand-backup/17-delete-data | + data= logger.go:42: 14:31:03 | gr-demand-backup/17-delete-data | + kubectl create configmap -n kuttl-test-crack-yak 16-delete-data-azure-0 --from-literal=data= logger.go:42: 14:31:03 | gr-demand-backup/17-delete-data | configmap/16-delete-data-azure-0 created logger.go:42: 14:31:03 | gr-demand-backup/17-delete-data | + for i in 0 1 2 logger.go:42: 14:31:03 | gr-demand-backup/17-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:31:03 | gr-demand-backup/17-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:31:03 | gr-demand-backup/17-delete-data | ++ local 'uri=-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:31:03 | gr-demand-backup/17-delete-data | ++ local pod= logger.go:42: 14:31:03 | gr-demand-backup/17-delete-data | +++ get_client_pod logger.go:42: 14:31:03 | gr-demand-backup/17-delete-data | +++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:31:03 | gr-demand-backup/17-delete-data | ++ client_pod=mysql-client logger.go:42: 14:31:03 | gr-demand-backup/17-delete-data | ++ wait_pod mysql-client logger.go:42: 14:31:03 | gr-demand-backup/17-delete-data | ++ local pod=mysql-client logger.go:42: 14:31:03 | gr-demand-backup/17-delete-data | ++ set +o xtrace logger.go:42: 14:31:04 | gr-demand-backup/17-delete-data | mysql-clienttrue logger.go:42: 14:31:04 | gr-demand-backup/17-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 14:31:04 | gr-demand-backup/17-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:31:04 | gr-demand-backup/17-delete-data | ++ kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:31:05 | gr-demand-backup/17-delete-data | ++ : logger.go:42: 14:31:05 | gr-demand-backup/17-delete-data | + data= logger.go:42: 14:31:05 | gr-demand-backup/17-delete-data | + kubectl create configmap -n kuttl-test-crack-yak 16-delete-data-azure-1 --from-literal=data= logger.go:42: 14:31:05 | gr-demand-backup/17-delete-data | configmap/16-delete-data-azure-1 created logger.go:42: 14:31:05 | gr-demand-backup/17-delete-data | + for i in 0 1 2 logger.go:42: 14:31:05 | gr-demand-backup/17-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:31:05 | gr-demand-backup/17-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:31:05 | gr-demand-backup/17-delete-data | ++ local 'uri=-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:31:05 | gr-demand-backup/17-delete-data | ++ local pod= logger.go:42: 14:31:05 | gr-demand-backup/17-delete-data | +++ get_client_pod logger.go:42: 14:31:05 | gr-demand-backup/17-delete-data | +++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:31:06 | gr-demand-backup/17-delete-data | ++ client_pod=mysql-client logger.go:42: 14:31:06 | gr-demand-backup/17-delete-data | ++ wait_pod mysql-client logger.go:42: 14:31:06 | gr-demand-backup/17-delete-data | ++ local pod=mysql-client logger.go:42: 14:31:06 | gr-demand-backup/17-delete-data | ++ set +o xtrace logger.go:42: 14:31:06 | gr-demand-backup/17-delete-data | mysql-clienttrue logger.go:42: 14:31:06 | gr-demand-backup/17-delete-data | ++ kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:31:06 | gr-demand-backup/17-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 14:31:06 | gr-demand-backup/17-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:31:07 | gr-demand-backup/17-delete-data | ++ : logger.go:42: 14:31:07 | gr-demand-backup/17-delete-data | + data= logger.go:42: 14:31:07 | gr-demand-backup/17-delete-data | + kubectl create configmap -n kuttl-test-crack-yak 16-delete-data-azure-2 --from-literal=data= logger.go:42: 14:31:08 | gr-demand-backup/17-delete-data | configmap/16-delete-data-azure-2 created logger.go:42: 14:31:09 | gr-demand-backup/17-delete-data | test step completed 17-delete-data logger.go:42: 14:31:09 | gr-demand-backup/18-restore-from-azure | starting test step 18-restore-from-azure logger.go:42: 14:31:10 | gr-demand-backup/18-restore-from-azure | PerconaServerMySQLRestore:kuttl-test-crack-yak/gr-demand-backup-restore-azure created logger.go:42: 14:36:04 | gr-demand-backup/18-restore-from-azure | test step completed 18-restore-from-azure logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | starting test step 19-read-data logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions cluster_name=$(get_cluster_name) for i in 0 1 2; do data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 06-read-data-azure-${i} --from-literal=data="${data}" done] logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | + source ../../functions logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ realpath ../../.. logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | ++++ pwd logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-demand-backup logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | ++ test_name=gr-demand-backup logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-demand-backup logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ export GIT_BRANCH=PR-523 logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ GIT_BRANCH=PR-523 logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ export VERSION=PR-523-f00253e logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ VERSION=PR-523-f00253e logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | ++++ which gdate logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | ++++ which date logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ date=/usr/bin/date logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ command -v oc logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ kubectl get nodes logger.go:42: 14:36:04 | gr-demand-backup/19-read-data | +++ grep '^minikube' logger.go:42: 14:36:05 | gr-demand-backup/19-read-data | ++ get_cluster_name logger.go:42: 14:36:05 | gr-demand-backup/19-read-data | ++ kubectl -n kuttl-test-crack-yak get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:36:05 | gr-demand-backup/19-read-data | + cluster_name=gr-demand-backup logger.go:42: 14:36:05 | gr-demand-backup/19-read-data | + for i in 0 1 2 logger.go:42: 14:36:05 | gr-demand-backup/19-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:36:05 | gr-demand-backup/19-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:36:05 | gr-demand-backup/19-read-data | ++ local 'uri=-h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:36:05 | gr-demand-backup/19-read-data | ++ local pod= logger.go:42: 14:36:05 | gr-demand-backup/19-read-data | +++ get_client_pod logger.go:42: 14:36:05 | gr-demand-backup/19-read-data | +++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:36:06 | gr-demand-backup/19-read-data | ++ client_pod=mysql-client logger.go:42: 14:36:06 | gr-demand-backup/19-read-data | ++ wait_pod mysql-client logger.go:42: 14:36:06 | gr-demand-backup/19-read-data | ++ local pod=mysql-client logger.go:42: 14:36:06 | gr-demand-backup/19-read-data | ++ set +o xtrace logger.go:42: 14:36:06 | gr-demand-backup/19-read-data | mysql-clienttrue logger.go:42: 14:36:06 | gr-demand-backup/19-read-data | ++ kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-0.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:36:06 | gr-demand-backup/19-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:36:06 | gr-demand-backup/19-read-data | ++ sed -e 's/mysql: //' logger.go:42: 14:36:07 | gr-demand-backup/19-read-data | + data=100500 logger.go:42: 14:36:07 | gr-demand-backup/19-read-data | + kubectl create configmap -n kuttl-test-crack-yak 06-read-data-azure-0 --from-literal=data=100500 logger.go:42: 14:36:08 | gr-demand-backup/19-read-data | configmap/06-read-data-azure-0 created logger.go:42: 14:36:08 | gr-demand-backup/19-read-data | + for i in 0 1 2 logger.go:42: 14:36:08 | gr-demand-backup/19-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:36:08 | gr-demand-backup/19-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:36:08 | gr-demand-backup/19-read-data | ++ local 'uri=-h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:36:08 | gr-demand-backup/19-read-data | ++ local pod= logger.go:42: 14:36:08 | gr-demand-backup/19-read-data | +++ get_client_pod logger.go:42: 14:36:08 | gr-demand-backup/19-read-data | +++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:36:08 | gr-demand-backup/19-read-data | ++ client_pod=mysql-client logger.go:42: 14:36:08 | gr-demand-backup/19-read-data | ++ wait_pod mysql-client logger.go:42: 14:36:08 | gr-demand-backup/19-read-data | ++ local pod=mysql-client logger.go:42: 14:36:08 | gr-demand-backup/19-read-data | ++ set +o xtrace logger.go:42: 14:36:09 | gr-demand-backup/19-read-data | mysql-clienttrue logger.go:42: 14:36:09 | gr-demand-backup/19-read-data | ++ sed -e 's/mysql: //' logger.go:42: 14:36:09 | gr-demand-backup/19-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:36:09 | gr-demand-backup/19-read-data | ++ kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-1.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:36:10 | gr-demand-backup/19-read-data | + data=100500 logger.go:42: 14:36:10 | gr-demand-backup/19-read-data | + kubectl create configmap -n kuttl-test-crack-yak 06-read-data-azure-1 --from-literal=data=100500 logger.go:42: 14:36:10 | gr-demand-backup/19-read-data | configmap/06-read-data-azure-1 created logger.go:42: 14:36:10 | gr-demand-backup/19-read-data | + for i in 0 1 2 logger.go:42: 14:36:10 | gr-demand-backup/19-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:36:10 | gr-demand-backup/19-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:36:10 | gr-demand-backup/19-read-data | ++ local 'uri=-h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:36:10 | gr-demand-backup/19-read-data | ++ local pod= logger.go:42: 14:36:10 | gr-demand-backup/19-read-data | +++ get_client_pod logger.go:42: 14:36:10 | gr-demand-backup/19-read-data | +++ kubectl -n kuttl-test-crack-yak get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:36:11 | gr-demand-backup/19-read-data | ++ client_pod=mysql-client logger.go:42: 14:36:11 | gr-demand-backup/19-read-data | ++ wait_pod mysql-client logger.go:42: 14:36:11 | gr-demand-backup/19-read-data | ++ local pod=mysql-client logger.go:42: 14:36:11 | gr-demand-backup/19-read-data | ++ set +o xtrace logger.go:42: 14:36:11 | gr-demand-backup/19-read-data | mysql-clienttrue logger.go:42: 14:36:11 | gr-demand-backup/19-read-data | ++ kubectl -n kuttl-test-crack-yak exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-demand-backup-mysql-2.gr-demand-backup-mysql -uroot -proot_password' logger.go:42: 14:36:11 | gr-demand-backup/19-read-data | ++ sed -e 's/mysql: //' logger.go:42: 14:36:11 | gr-demand-backup/19-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:36:12 | gr-demand-backup/19-read-data | + data=100500 logger.go:42: 14:36:12 | gr-demand-backup/19-read-data | + kubectl create configmap -n kuttl-test-crack-yak 06-read-data-azure-2 --from-literal=data=100500 logger.go:42: 14:36:13 | gr-demand-backup/19-read-data | configmap/06-read-data-azure-2 created logger.go:42: 14:36:14 | gr-demand-backup/19-read-data | test step completed 19-read-data logger.go:42: 14:36:14 | gr-demand-backup/99-drop-finalizer | starting test step 99-drop-finalizer logger.go:42: 14:36:15 | gr-demand-backup/99-drop-finalizer | PerconaServerMySQL:kuttl-test-crack-yak/gr-demand-backup updated logger.go:42: 14:36:15 | gr-demand-backup/99-drop-finalizer | test step completed 99-drop-finalizer logger.go:42: 14:36:15 | gr-demand-backup | gr-demand-backup events from ns kuttl-test-crack-yak: logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:06:35 +0000 UTC Normal Pod percona-server-mysql-operator-b4c599bbb-847px Scheduled Successfully assigned kuttl-test-crack-yak/percona-server-mysql-operator-b4c599bbb-847px to gke-jen-ps-523-f00253e-4-default-pool-e964acce-7189 default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:06:35 +0000 UTC Normal ReplicaSet.apps percona-server-mysql-operator-b4c599bbb SuccessfulCreate Created pod: percona-server-mysql-operator-b4c599bbb-847px replicaset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:06:35 +0000 UTC Normal Deployment.apps percona-server-mysql-operator ScalingReplicaSet Scaled up replica set percona-server-mysql-operator-b4c599bbb to 1 deployment-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:06:37 +0000 UTC Normal Pod percona-server-mysql-operator-b4c599bbb-847px.spec.containers{manager} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:06:39 +0000 UTC Normal Pod mysql-client Scheduled Successfully assigned kuttl-test-crack-yak/mysql-client to gke-jen-ps-523-f00253e-4-default-pool-e964acce-7189 default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:06:40 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Pulling Pulling image "percona/percona-server:8.0.33" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:06:50 +0000 UTC Normal PersistentVolumeClaim minio-service WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:06:51 +0000 UTC Normal ReplicaSet.apps minio-service-85cdcd4d44 SuccessfulCreate Created pod: minio-service-85cdcd4d44-v7kmg replicaset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:06:51 +0000 UTC Normal Deployment.apps minio-service ScalingReplicaSet Scaled up replica set minio-service-85cdcd4d44 to 1 deployment-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:06:51 +0000 UTC Normal PersistentVolumeClaim minio-service ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:06:51 +0000 UTC Normal PersistentVolumeClaim minio-service Provisioning External provisioner is provisioning volume for claim "kuttl-test-crack-yak/minio-service" pd.csi.storage.gke.io_gke-77d1c4fac69c4c2a896d-0b77-45e4-vm_8a268d8f-c2f3-4090-b361-6efbcaeaaace logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:06:51 +0000 UTC Normal Pod percona-server-mysql-operator-b4c599bbb-847px.spec.containers{manager} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 14.334443827s (14.334492172s including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:06:51 +0000 UTC Normal Pod percona-server-mysql-operator-b4c599bbb-847px.spec.containers{manager} Created Created container manager kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:06:51 +0000 UTC Normal Pod percona-server-mysql-operator-b4c599bbb-847px.spec.containers{manager} Started Started container manager kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:06:52 +0000 UTC Normal Lease.coordination.k8s.io 08db2feb.percona.com LeaderElection percona-server-mysql-operator-b4c599bbb-847px_b1318720-dbf0-46eb-bc25-fbafc7aa096d became leader percona-server-mysql-operator-b4c599bbb-847px_b1318720-dbf0-46eb-bc25-fbafc7aa096d logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:06:52 +0000 UTC Normal Pod minio-service-post-job-nc6m7 Scheduled Successfully assigned kuttl-test-crack-yak/minio-service-post-job-nc6m7 to gke-jen-ps-523-f00253e-4-default-pool-e964acce-7189 default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:06:52 +0000 UTC Normal Job.batch minio-service-post-job SuccessfulCreate Created pod: minio-service-post-job-nc6m7 job-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:06:53 +0000 UTC Normal Pod minio-service-post-job-nc6m7.spec.containers{minio-make-user} Pulling Pulling image "quay.io/minio/mc:RELEASE.2023-09-29T16-41-22Z" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:06:55 +0000 UTC Normal PersistentVolumeClaim minio-service ProvisioningSucceeded Successfully provisioned volume pvc-6f876da2-0476-485a-acb2-ec2b3a578af6 pd.csi.storage.gke.io_gke-77d1c4fac69c4c2a896d-0b77-45e4-vm_8a268d8f-c2f3-4090-b361-6efbcaeaaace logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:06:56 +0000 UTC Normal Pod minio-service-85cdcd4d44-v7kmg Scheduled Successfully assigned kuttl-test-crack-yak/minio-service-85cdcd4d44-v7kmg to gke-jen-ps-523-f00253e-4-default-pool-e964acce-7189 default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:07:03 +0000 UTC Normal Pod minio-service-85cdcd4d44-v7kmg SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-6f876da2-0476-485a-acb2-ec2b3a578af6" attachdetach-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:07:06 +0000 UTC Normal Pod minio-service-85cdcd4d44-v7kmg.spec.containers{minio} Pulling Pulling image "quay.io/minio/minio:RELEASE.2023-09-30T07-02-29Z" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:07:16 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Pulled Successfully pulled image "percona/percona-server:8.0.33" in 25.208471347s (36.218190361s including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:07:16 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Created Created container mysql-client kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:07:17 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Started Started container mysql-client kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:07:20 +0000 UTC Normal Pod minio-service-post-job-nc6m7.spec.containers{minio-make-user} Pulled Successfully pulled image "quay.io/minio/mc:RELEASE.2023-09-29T16-41-22Z" in 3.506588261s (26.670982291s including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:07:20 +0000 UTC Normal Pod minio-service-post-job-nc6m7.spec.containers{minio-make-user} Created Created container minio-make-user kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:07:20 +0000 UTC Normal Pod minio-service-post-job-nc6m7.spec.containers{minio-make-user} Started Started container minio-make-user kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:07:24 +0000 UTC Normal Pod minio-service-85cdcd4d44-v7kmg.spec.containers{minio} Pulled Successfully pulled image "quay.io/minio/minio:RELEASE.2023-09-30T07-02-29Z" in 3.580517793s (17.866920452s including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:07:24 +0000 UTC Normal Pod minio-service-85cdcd4d44-v7kmg.spec.containers{minio} Created Created container minio kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:07:24 +0000 UTC Normal Pod minio-service-85cdcd4d44-v7kmg.spec.containers{minio} Started Started container minio kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:07:37 +0000 UTC Normal Job.batch minio-service-post-job Completed Job completed job-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:07:40 +0000 UTC Normal Pod aws-cli Scheduled Successfully assigned kuttl-test-crack-yak/aws-cli to gke-jen-ps-523-f00253e-4-default-pool-e964acce-7189 default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:07:41 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Pulling Pulling image "perconalab/awscli" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:07:44 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Pulled Successfully pulled image "perconalab/awscli" in 3.13736042s (3.137375778s including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:07:44 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Created Created container aws-cli kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:07:44 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Started Started container aws-cli kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:07:57 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:07:57 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-0 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:07:57 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-crack-yak/datadir-gr-demand-backup-mysql-0" pd.csi.storage.gke.io_gke-77d1c4fac69c4c2a896d-0b77-45e4-vm_8a268d8f-c2f3-4090-b361-6efbcaeaaace logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:07:57 +0000 UTC Normal StatefulSet.apps gr-demand-backup-mysql SuccessfulCreate create Claim datadir-gr-demand-backup-mysql-0 Pod gr-demand-backup-mysql-0 in StatefulSet gr-demand-backup-mysql success statefulset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:07:57 +0000 UTC Normal StatefulSet.apps gr-demand-backup-mysql SuccessfulCreate create Pod gr-demand-backup-mysql-0 in StatefulSet gr-demand-backup-mysql successful statefulset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:08:01 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-0 ProvisioningSucceeded Successfully provisioned volume pvc-b0d7a2e1-08de-43c6-a5c6-18a6360e6a75 pd.csi.storage.gke.io_gke-77d1c4fac69c4c2a896d-0b77-45e4-vm_8a268d8f-c2f3-4090-b361-6efbcaeaaace logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:08:01 +0000 UTC Normal Pod gr-demand-backup-mysql-0 Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-mysql-0 to gke-jen-ps-523-f00253e-4-default-pool-e964acce-bq8g default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:08:06 +0000 UTC Normal Pod gr-demand-backup-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-b0d7a2e1-08de-43c6-a5c6-18a6360e6a75" attachdetach-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:08:08 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:08:21 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 13.488763081s (13.488779031s including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:08:21 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:08:21 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:08:28 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:08:48 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 20.106535821s (20.106550947s including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:08:48 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:08:48 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:08:48 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:08:57 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 8.36826501s (8.368288947s including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:08:58 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:08:58 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:09:23 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-1 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:09:23 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-1 Provisioning External provisioner is provisioning volume for claim "kuttl-test-crack-yak/datadir-gr-demand-backup-mysql-1" pd.csi.storage.gke.io_gke-77d1c4fac69c4c2a896d-0b77-45e4-vm_8a268d8f-c2f3-4090-b361-6efbcaeaaace logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:09:23 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-1 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:09:23 +0000 UTC Normal StatefulSet.apps gr-demand-backup-mysql SuccessfulCreate create Claim datadir-gr-demand-backup-mysql-1 Pod gr-demand-backup-mysql-1 in StatefulSet gr-demand-backup-mysql success statefulset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:09:23 +0000 UTC Normal StatefulSet.apps gr-demand-backup-mysql SuccessfulCreate create Pod gr-demand-backup-mysql-1 in StatefulSet gr-demand-backup-mysql successful statefulset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:09:27 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-1 ProvisioningSucceeded Successfully provisioned volume pvc-1d935bf6-3215-4af9-9c3f-84b9b844d646 pd.csi.storage.gke.io_gke-77d1c4fac69c4c2a896d-0b77-45e4-vm_8a268d8f-c2f3-4090-b361-6efbcaeaaace logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:09:28 +0000 UTC Normal Pod gr-demand-backup-mysql-1 Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-mysql-1 to gke-jen-ps-523-f00253e-4-default-pool-e964acce-7189 default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:09:35 +0000 UTC Normal Pod gr-demand-backup-mysql-1 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-1d935bf6-3215-4af9-9c3f-84b9b844d646" attachdetach-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:09:38 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:09:38 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 121.752355ms (121.761984ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:09:38 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:09:39 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:09:41 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:10:01 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 20.821199322s (20.821222266s including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:10:01 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:10:01 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:10:01 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:10:12 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 10.408869148s (10.408880895s including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:10:12 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:10:12 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:10:36 +0000 UTC Warning Pod gr-demand-backup-mysql-1.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:10:36 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:10:37 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 79.416207ms (79.429716ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:11:08 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-2 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:11:08 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-2 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:11:08 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-2 Provisioning External provisioner is provisioning volume for claim "kuttl-test-crack-yak/datadir-gr-demand-backup-mysql-2" pd.csi.storage.gke.io_gke-77d1c4fac69c4c2a896d-0b77-45e4-vm_8a268d8f-c2f3-4090-b361-6efbcaeaaace logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:11:08 +0000 UTC Normal StatefulSet.apps gr-demand-backup-mysql SuccessfulCreate create Claim datadir-gr-demand-backup-mysql-2 Pod gr-demand-backup-mysql-2 in StatefulSet gr-demand-backup-mysql success statefulset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:11:08 +0000 UTC Normal StatefulSet.apps gr-demand-backup-mysql SuccessfulCreate create Pod gr-demand-backup-mysql-2 in StatefulSet gr-demand-backup-mysql successful statefulset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:11:12 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-2 ProvisioningSucceeded Successfully provisioned volume pvc-a6d60621-4341-4c07-a834-34625219bf69 pd.csi.storage.gke.io_gke-77d1c4fac69c4c2a896d-0b77-45e4-vm_8a268d8f-c2f3-4090-b361-6efbcaeaaace logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:11:12 +0000 UTC Normal Pod gr-demand-backup-mysql-2 Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-mysql-2 to gke-jen-ps-523-f00253e-4-default-pool-e964acce-w8zz default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:11:20 +0000 UTC Normal Pod gr-demand-backup-mysql-2 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-a6d60621-4341-4c07-a834-34625219bf69" attachdetach-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:11:21 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:11:35 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 14.254909752s (14.254932076s including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:11:35 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:11:35 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:11:42 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:12:03 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 20.82883589s (20.828852481s including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:12:03 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:12:03 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:12:03 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:12:14 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 10.894554512s (10.894572578s including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:12:14 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:12:15 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:12:37 +0000 UTC Warning Pod gr-demand-backup-mysql-2.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:12:37 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:12:40 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 120.779058ms (120.835723ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:16 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-s2krv Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-router-6498c5b898-s2krv to gke-jen-ps-523-f00253e-4-default-pool-e964acce-bq8g default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:16 +0000 UTC Normal ReplicaSet.apps gr-demand-backup-router-6498c5b898 SuccessfulCreate Created pod: gr-demand-backup-router-6498c5b898-s2krv replicaset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:16 +0000 UTC Normal Deployment.apps gr-demand-backup-router ScalingReplicaSet Scaled up replica set gr-demand-backup-router-6498c5b898 to 3 deployment-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:17 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-7tzrh Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-router-6498c5b898-7tzrh to gke-jen-ps-523-f00253e-4-default-pool-e964acce-7189 default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:17 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-7tzrh.spec.initContainers{router-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:17 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-7tzrh.spec.initContainers{router-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 111.553357ms (111.563797ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:17 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-7tzrh.spec.initContainers{router-init} Created Created container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:17 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-s2krv.spec.initContainers{router-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:17 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-s2krv.spec.initContainers{router-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 107.477229ms (107.484618ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:17 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-s2krv.spec.initContainers{router-init} Created Created container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:17 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-s2krv.spec.initContainers{router-init} Started Started container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:17 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-szmbj Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-router-6498c5b898-szmbj to gke-jen-ps-523-f00253e-4-default-pool-e964acce-w8zz default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:17 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-szmbj.spec.initContainers{router-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:17 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-szmbj.spec.initContainers{router-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 106.927219ms (106.941918ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:17 +0000 UTC Normal ReplicaSet.apps gr-demand-backup-router-6498c5b898 SuccessfulCreate Created pod: gr-demand-backup-router-6498c5b898-7tzrh replicaset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:17 +0000 UTC Normal ReplicaSet.apps gr-demand-backup-router-6498c5b898 SuccessfulCreate Created pod: gr-demand-backup-router-6498c5b898-szmbj replicaset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:18 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-7tzrh.spec.initContainers{router-init} Started Started container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:18 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-szmbj.spec.initContainers{router-init} Created Created container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:18 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-szmbj.spec.initContainers{router-init} Started Started container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:19 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-7tzrh.spec.containers{router} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-router" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:19 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-s2krv.spec.containers{router} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-router" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:19 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-szmbj.spec.containers{router} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-router" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:34 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-7tzrh.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 15.258604517s (15.258619967s including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:34 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-7tzrh.spec.containers{router} Created Created container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:34 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-7tzrh.spec.containers{router} Started Started container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:34 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-s2krv.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 15.591400389s (15.59141756s including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:34 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-s2krv.spec.containers{router} Created Created container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:34 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-s2krv.spec.containers{router} Started Started container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:34 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-szmbj.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 14.897010407s (14.897017902s including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:34 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-szmbj.spec.containers{router} Created Created container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:34 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-szmbj.spec.containers{router} Started Started container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:54 +0000 UTC Normal Pod xb-gr-demand-backup-minio-minio-nxtt7 Scheduled Successfully assigned kuttl-test-crack-yak/xb-gr-demand-backup-minio-minio-nxtt7 to gke-jen-ps-523-f00253e-4-default-pool-e964acce-bq8g default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:54 +0000 UTC Normal Job.batch xb-gr-demand-backup-minio-minio SuccessfulCreate Created pod: xb-gr-demand-backup-minio-minio-nxtt7 job-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:55 +0000 UTC Normal Pod xb-gr-demand-backup-minio-minio-nxtt7.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:56 +0000 UTC Normal Pod xb-gr-demand-backup-minio-minio-nxtt7.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 924.459871ms (924.47506ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:56 +0000 UTC Normal Pod xb-gr-demand-backup-minio-minio-nxtt7.spec.initContainers{xtrabackup-init} Created Created container xtrabackup-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:13:56 +0000 UTC Normal Pod xb-gr-demand-backup-minio-minio-nxtt7.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:00 +0000 UTC Normal Pod xb-gr-demand-backup-minio-minio-nxtt7.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:00 +0000 UTC Normal Pod xb-gr-demand-backup-minio-minio-nxtt7.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 82.520037ms (82.535161ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:00 +0000 UTC Normal Pod xb-gr-demand-backup-minio-minio-nxtt7.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:00 +0000 UTC Normal Pod xb-gr-demand-backup-minio-minio-nxtt7.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:07 +0000 UTC Normal Job.batch xb-gr-demand-backup-minio-minio Completed Job completed job-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:26 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup} Killing Stopping container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:26 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:26 +0000 UTC Normal StatefulSet.apps gr-demand-backup-mysql SuccessfulDelete delete Pod gr-demand-backup-mysql-2 in StatefulSet gr-demand-backup-mysql successful statefulset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:27 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-7tzrh.spec.containers{router} Killing Stopping container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:27 +0000 UTC Warning Pod gr-demand-backup-router-6498c5b898-7tzrh.spec.containers{router} Unhealthy Readiness probe errored: rpc error: code = Unknown desc = failed to exec in container: container is in CONTAINER_EXITED state kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:27 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-s2krv.spec.containers{router} Killing Stopping container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:27 +0000 UTC Warning Pod gr-demand-backup-router-6498c5b898-s2krv.spec.containers{router} Unhealthy Readiness probe failed: kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:27 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-szmbj.spec.containers{router} Killing Stopping container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:27 +0000 UTC Warning Pod gr-demand-backup-router-6498c5b898-szmbj.spec.containers{router} Unhealthy Readiness probe errored: rpc error: code = Unknown desc = failed to exec in container: container is in CONTAINER_EXITED state kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:27 +0000 UTC Normal ReplicaSet.apps gr-demand-backup-router-6498c5b898 SuccessfulDelete Deleted pod: gr-demand-backup-router-6498c5b898-szmbj replicaset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:27 +0000 UTC Normal ReplicaSet.apps gr-demand-backup-router-6498c5b898 SuccessfulDelete Deleted pod: gr-demand-backup-router-6498c5b898-s2krv replicaset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:27 +0000 UTC Normal ReplicaSet.apps gr-demand-backup-router-6498c5b898 SuccessfulDelete Deleted pod: gr-demand-backup-router-6498c5b898-7tzrh replicaset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:27 +0000 UTC Normal Deployment.apps gr-demand-backup-router ScalingReplicaSet Scaled down replica set gr-demand-backup-router-6498c5b898 to 0 from 3 deployment-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:27 +0000 UTC Warning Endpoints gr-demand-backup-router FailedToUpdateEndpoint Failed to update endpoint kuttl-test-crack-yak/gr-demand-backup-router: Operation cannot be fulfilled on endpoints "gr-demand-backup-router": the object has been modified; please apply your changes to the latest version and try again endpoint-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:31 +0000 UTC Warning Pod gr-demand-backup-mysql-2.spec.containers{mysql} Unhealthy Readiness probe failed: 2024/02/05 14:14:31 readiness check failed: connect to db: ping DB: dial tcp 10.208.42.9:33062: connect: connection refused kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:33 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup} Killing Stopping container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:33 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:33 +0000 UTC Normal StatefulSet.apps gr-demand-backup-mysql SuccessfulDelete delete Pod gr-demand-backup-mysql-1 in StatefulSet gr-demand-backup-mysql successful statefulset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:38 +0000 UTC Warning Pod gr-demand-backup-mysql-1.spec.containers{mysql} Unhealthy Readiness probe failed: 2024/02/05 14:14:38 readiness check failed: connect to db: ping DB: dial tcp 10.208.40.9:33062: connect: connection refused kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:41 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:41 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup} Killing Stopping container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:41 +0000 UTC Normal StatefulSet.apps gr-demand-backup-mysql SuccessfulDelete delete Pod gr-demand-backup-mysql-0 in StatefulSet gr-demand-backup-mysql successful statefulset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:44 +0000 UTC Warning Pod gr-demand-backup-mysql-0.spec.containers{mysql} Unhealthy Readiness probe failed: [mysql] 2024/02/05 14:14:44 packets.go:37: read tcp 10.208.41.4:56650->10.208.41.4:33062: read: connection reset by peer 2024/02/05 14:14:44 readiness check failed: connect to db: ping DB: dial tcp 10.208.41.4:33062: connect: connection refused kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:53 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-minio-b7fjk Scheduled Successfully assigned kuttl-test-crack-yak/xb-restore-gr-demand-backup-restore-minio-b7fjk to gke-jen-ps-523-f00253e-4-default-pool-e964acce-7189 default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:14:53 +0000 UTC Normal Job.batch xb-restore-gr-demand-backup-restore-minio SuccessfulCreate Created pod: xb-restore-gr-demand-backup-restore-minio-b7fjk job-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:15:06 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-minio-b7fjk SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-b0d7a2e1-08de-43c6-a5c6-18a6360e6a75" attachdetach-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:15:07 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-minio-b7fjk.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:15:07 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-minio-b7fjk.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 104.821654ms (104.836824ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:15:07 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-minio-b7fjk.spec.initContainers{xtrabackup-init} Created Created container xtrabackup-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:15:07 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-minio-b7fjk.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:15:08 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-minio-b7fjk.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:15:08 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-minio-b7fjk.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 80.974172ms (80.982047ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:15:08 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-minio-b7fjk.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:15:09 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-minio-b7fjk.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:15:18 +0000 UTC Normal Job.batch xb-restore-gr-demand-backup-restore-minio Completed Job completed job-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:15:22 +0000 UTC Normal Pod gr-demand-backup-mysql-0 Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-mysql-0 to gke-jen-ps-523-f00253e-4-default-pool-e964acce-bq8g default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:15:22 +0000 UTC Warning Pod gr-demand-backup-mysql-0 FailedAttachVolume Multi-Attach error for volume "pvc-b0d7a2e1-08de-43c6-a5c6-18a6360e6a75" Volume is already exclusively attached to one node and can't be attached to another attachdetach-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:15:39 +0000 UTC Normal Pod gr-demand-backup-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-b0d7a2e1-08de-43c6-a5c6-18a6360e6a75" attachdetach-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:15:40 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:15:40 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 125.416875ms (125.425788ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:15:40 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:15:41 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:15:42 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:15:42 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 84.38427ms (84.399035ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:15:42 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:15:42 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:15:42 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:15:42 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 81.860813ms (81.872851ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:15:42 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:15:43 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:16:15 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-1 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:16:15 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-1 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:16:15 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-1 Provisioning External provisioner is provisioning volume for claim "kuttl-test-crack-yak/datadir-gr-demand-backup-mysql-1" pd.csi.storage.gke.io_gke-77d1c4fac69c4c2a896d-0b77-45e4-vm_8a268d8f-c2f3-4090-b361-6efbcaeaaace logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:16:19 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-1 ProvisioningSucceeded Successfully provisioned volume pvc-aaf7b7e3-62c9-477d-b8ca-df02166d260a pd.csi.storage.gke.io_gke-77d1c4fac69c4c2a896d-0b77-45e4-vm_8a268d8f-c2f3-4090-b361-6efbcaeaaace logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:16:19 +0000 UTC Normal Pod gr-demand-backup-mysql-1 Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-mysql-1 to gke-jen-ps-523-f00253e-4-default-pool-e964acce-w8zz default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:16:27 +0000 UTC Normal Pod gr-demand-backup-mysql-1 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-aaf7b7e3-62c9-477d-b8ca-df02166d260a" attachdetach-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:16:28 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:16:28 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 110.128987ms (110.150943ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:16:28 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:16:28 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:16:30 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:16:30 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 98.737822ms (98.755046ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:16:30 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:16:30 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:16:30 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:16:30 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 92.425723ms (92.433047ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:16:30 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:16:30 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:17:01 +0000 UTC Warning Pod gr-demand-backup-mysql-1.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:17:01 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:17:01 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 90.377399ms (90.395285ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:17:33 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-2 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:17:33 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-2 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:17:33 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-2 Provisioning External provisioner is provisioning volume for claim "kuttl-test-crack-yak/datadir-gr-demand-backup-mysql-2" pd.csi.storage.gke.io_gke-77d1c4fac69c4c2a896d-0b77-45e4-vm_8a268d8f-c2f3-4090-b361-6efbcaeaaace logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:17:36 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-2 ProvisioningSucceeded Successfully provisioned volume pvc-bfe1db87-789c-4a55-a1b4-de3220ed5467 pd.csi.storage.gke.io_gke-77d1c4fac69c4c2a896d-0b77-45e4-vm_8a268d8f-c2f3-4090-b361-6efbcaeaaace logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:17:37 +0000 UTC Normal Pod gr-demand-backup-mysql-2 Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-mysql-2 to gke-jen-ps-523-f00253e-4-default-pool-e964acce-7189 default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:17:44 +0000 UTC Normal Pod gr-demand-backup-mysql-2 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-bfe1db87-789c-4a55-a1b4-de3220ed5467" attachdetach-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:17:48 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:17:48 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 100.577577ms (100.595127ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:17:48 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:17:48 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:17:50 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:17:50 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 79.850436ms (79.875451ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:17:50 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:17:50 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:17:50 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:17:50 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 78.228427ms (78.250646ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:17:50 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:17:50 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:18:20 +0000 UTC Warning Pod gr-demand-backup-mysql-2.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:18:20 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:18:20 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 96.851249ms (96.865673ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:02 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-lqsxw Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-router-6498c5b898-lqsxw to gke-jen-ps-523-f00253e-4-default-pool-e964acce-bq8g default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:02 +0000 UTC Normal ReplicaSet.apps gr-demand-backup-router-6498c5b898 SuccessfulCreate Created pod: gr-demand-backup-router-6498c5b898-lqsxw replicaset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:02 +0000 UTC Normal ReplicaSet.apps gr-demand-backup-router-6498c5b898 SuccessfulCreate Created pod: gr-demand-backup-router-6498c5b898-qsg8r replicaset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:02 +0000 UTC Normal ReplicaSet.apps gr-demand-backup-router-6498c5b898 SuccessfulCreate Created pod: gr-demand-backup-router-6498c5b898-ndll6 replicaset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:02 +0000 UTC Normal Deployment.apps gr-demand-backup-router ScalingReplicaSet Scaled up replica set gr-demand-backup-router-6498c5b898 to 3 from 0 deployment-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:03 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-lqsxw.spec.initContainers{router-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:03 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-lqsxw.spec.initContainers{router-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 92.223746ms (92.233943ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:03 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-lqsxw.spec.initContainers{router-init} Created Created container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:03 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-lqsxw.spec.initContainers{router-init} Started Started container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:03 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-ndll6 Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-router-6498c5b898-ndll6 to gke-jen-ps-523-f00253e-4-default-pool-e964acce-w8zz default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:03 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-ndll6.spec.initContainers{router-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:03 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-ndll6.spec.initContainers{router-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 98.837318ms (98.85396ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:03 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-ndll6.spec.initContainers{router-init} Created Created container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:03 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-ndll6.spec.initContainers{router-init} Started Started container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:03 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-qsg8r Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-router-6498c5b898-qsg8r to gke-jen-ps-523-f00253e-4-default-pool-e964acce-7189 default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:03 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-qsg8r.spec.initContainers{router-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:03 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-qsg8r.spec.initContainers{router-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 81.825441ms (81.833882ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:03 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-qsg8r.spec.initContainers{router-init} Created Created container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:03 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-qsg8r.spec.initContainers{router-init} Started Started container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:05 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-lqsxw.spec.containers{router} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-router" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:05 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-lqsxw.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 88.048828ms (88.056297ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:05 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-lqsxw.spec.containers{router} Created Created container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:05 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-lqsxw.spec.containers{router} Started Started container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:05 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-ndll6.spec.containers{router} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-router" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:05 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-qsg8r.spec.containers{router} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-router" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:05 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-qsg8r.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 73.993429ms (74.009533ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:05 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-qsg8r.spec.containers{router} Created Created container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:05 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-qsg8r.spec.containers{router} Started Started container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:06 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-ndll6.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 107.929738ms (107.942983ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:06 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-ndll6.spec.containers{router} Created Created container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:06 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-ndll6.spec.containers{router} Started Started container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:32 +0000 UTC Normal Pod xb-gr-demand-backup-s3-aws-s3-tj9pk Scheduled Successfully assigned kuttl-test-crack-yak/xb-gr-demand-backup-s3-aws-s3-tj9pk to gke-jen-ps-523-f00253e-4-default-pool-e964acce-bq8g default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:32 +0000 UTC Normal Job.batch xb-gr-demand-backup-s3-aws-s3 SuccessfulCreate Created pod: xb-gr-demand-backup-s3-aws-s3-tj9pk job-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:33 +0000 UTC Normal Pod xb-gr-demand-backup-s3-aws-s3-tj9pk.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:33 +0000 UTC Normal Pod xb-gr-demand-backup-s3-aws-s3-tj9pk.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 80.241081ms (80.256703ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:33 +0000 UTC Normal Pod xb-gr-demand-backup-s3-aws-s3-tj9pk.spec.initContainers{xtrabackup-init} Created Created container xtrabackup-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:33 +0000 UTC Normal Pod xb-gr-demand-backup-s3-aws-s3-tj9pk.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:35 +0000 UTC Normal Pod xb-gr-demand-backup-s3-aws-s3-tj9pk.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:35 +0000 UTC Normal Pod xb-gr-demand-backup-s3-aws-s3-tj9pk.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 80.991511ms (81.007611ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:35 +0000 UTC Normal Pod xb-gr-demand-backup-s3-aws-s3-tj9pk.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:35 +0000 UTC Normal Pod xb-gr-demand-backup-s3-aws-s3-tj9pk.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:19:42 +0000 UTC Normal Job.batch xb-gr-demand-backup-s3-aws-s3 Completed Job completed job-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:07 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup} Killing Stopping container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:07 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:07 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-lqsxw.spec.containers{router} Killing Stopping container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:07 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-ndll6.spec.containers{router} Killing Stopping container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:07 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-qsg8r.spec.containers{router} Killing Stopping container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:07 +0000 UTC Normal ReplicaSet.apps gr-demand-backup-router-6498c5b898 SuccessfulDelete Deleted pod: gr-demand-backup-router-6498c5b898-lqsxw replicaset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:07 +0000 UTC Normal ReplicaSet.apps gr-demand-backup-router-6498c5b898 SuccessfulDelete Deleted pod: gr-demand-backup-router-6498c5b898-qsg8r replicaset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:07 +0000 UTC Normal ReplicaSet.apps gr-demand-backup-router-6498c5b898 SuccessfulDelete Deleted pod: gr-demand-backup-router-6498c5b898-ndll6 replicaset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:10 +0000 UTC Warning Pod gr-demand-backup-mysql-2.spec.containers{mysql} Unhealthy Readiness probe failed: [mysql] 2024/02/05 14:20:10 packets.go:37: read tcp 10.208.40.12:34804->10.208.40.12:33062: read: connection reset by peer 2024/02/05 14:20:10 readiness check failed: connect to db: ping DB: dial tcp 10.208.40.12:33062: connect: connection refused kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:13 +0000 UTC Warning Pod gr-demand-backup-mysql-2.spec.containers{mysql} Unhealthy Readiness probe failed: 2024/02/05 14:20:12 readiness check failed: connect to db: ping DB: dial tcp 10.208.40.12:33062: connect: connection refused kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:14 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup} Killing Stopping container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:14 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:18 +0000 UTC Warning Pod gr-demand-backup-mysql-1.spec.containers{mysql} Unhealthy Readiness probe failed: [mysql] 2024/02/05 14:20:18 packets.go:37: read tcp 10.208.42.11:52452->10.208.42.11:33062: read: connection reset by peer 2024/02/05 14:20:18 readiness check failed: connect to db: ping DB: dial tcp 10.208.42.11:33062: connect: connection refused kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:22 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:22 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup} Killing Stopping container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:25 +0000 UTC Warning Pod gr-demand-backup-mysql-0.spec.containers{mysql} Unhealthy Readiness probe failed: [mysql] 2024/02/05 14:20:25 packets.go:37: read tcp 10.208.41.7:60478->10.208.41.7:33062: read: connection reset by peer 2024/02/05 14:20:25 readiness check failed: connect to db: ping DB: dial tcp 10.208.41.7:33062: connect: connection refused kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:32 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-s3-rb6v7 Scheduled Successfully assigned kuttl-test-crack-yak/xb-restore-gr-demand-backup-restore-s3-rb6v7 to gke-jen-ps-523-f00253e-4-default-pool-e964acce-7189 default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:32 +0000 UTC Warning Pod xb-restore-gr-demand-backup-restore-s3-rb6v7 FailedAttachVolume Multi-Attach error for volume "pvc-b0d7a2e1-08de-43c6-a5c6-18a6360e6a75" Volume is already exclusively attached to one node and can't be attached to another attachdetach-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:32 +0000 UTC Normal Job.batch xb-restore-gr-demand-backup-restore-s3 SuccessfulCreate Created pod: xb-restore-gr-demand-backup-restore-s3-rb6v7 job-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:52 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-s3-rb6v7 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-b0d7a2e1-08de-43c6-a5c6-18a6360e6a75" attachdetach-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:53 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-s3-rb6v7.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:53 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-s3-rb6v7.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 97.670564ms (97.682107ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:53 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-s3-rb6v7.spec.initContainers{xtrabackup-init} Created Created container xtrabackup-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:53 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-s3-rb6v7.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:55 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-s3-rb6v7.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:55 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-s3-rb6v7.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 68.754109ms (68.769888ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:56 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-s3-rb6v7.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:20:56 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-s3-rb6v7.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:21:08 +0000 UTC Normal Job.batch xb-restore-gr-demand-backup-restore-s3 Completed Job completed job-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:21:14 +0000 UTC Normal Pod gr-demand-backup-mysql-0 Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-mysql-0 to gke-jen-ps-523-f00253e-4-default-pool-e964acce-bq8g default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:21:25 +0000 UTC Normal Pod gr-demand-backup-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-b0d7a2e1-08de-43c6-a5c6-18a6360e6a75" attachdetach-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:21:27 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:21:27 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 92.21371ms (92.230016ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:21:27 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:21:27 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:21:29 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:21:29 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 69.672101ms (69.68677ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:21:29 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:21:29 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:21:29 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:21:29 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 81.508356ms (81.525344ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:21:29 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:21:29 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:22:02 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-1 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:22:02 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-1 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:22:02 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-1 Provisioning External provisioner is provisioning volume for claim "kuttl-test-crack-yak/datadir-gr-demand-backup-mysql-1" pd.csi.storage.gke.io_gke-77d1c4fac69c4c2a896d-0b77-45e4-vm_8a268d8f-c2f3-4090-b361-6efbcaeaaace logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:22:06 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-1 ProvisioningSucceeded Successfully provisioned volume pvc-5c3847a8-3699-4633-a5c1-4553ece36020 pd.csi.storage.gke.io_gke-77d1c4fac69c4c2a896d-0b77-45e4-vm_8a268d8f-c2f3-4090-b361-6efbcaeaaace logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:22:06 +0000 UTC Normal Pod gr-demand-backup-mysql-1 Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-mysql-1 to gke-jen-ps-523-f00253e-4-default-pool-e964acce-7189 default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:22:11 +0000 UTC Normal Pod gr-demand-backup-mysql-1 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-5c3847a8-3699-4633-a5c1-4553ece36020" attachdetach-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:22:13 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:22:13 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 115.898549ms (115.919455ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:22:13 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:22:13 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:22:14 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:22:15 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 92.920166ms (92.927585ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:22:15 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:22:15 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:22:15 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:22:15 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 84.256875ms (84.275973ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:22:15 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:22:15 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:22:45 +0000 UTC Warning Pod gr-demand-backup-mysql-1.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:22:45 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:22:45 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 118.821184ms (118.831613ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:23:17 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-2 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:23:18 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-2 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:23:18 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-2 Provisioning External provisioner is provisioning volume for claim "kuttl-test-crack-yak/datadir-gr-demand-backup-mysql-2" pd.csi.storage.gke.io_gke-77d1c4fac69c4c2a896d-0b77-45e4-vm_8a268d8f-c2f3-4090-b361-6efbcaeaaace logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:23:21 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-2 ProvisioningSucceeded Successfully provisioned volume pvc-92a84e98-ed7d-48f3-86c5-a42d98b84cc4 pd.csi.storage.gke.io_gke-77d1c4fac69c4c2a896d-0b77-45e4-vm_8a268d8f-c2f3-4090-b361-6efbcaeaaace logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:23:22 +0000 UTC Normal Pod gr-demand-backup-mysql-2 Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-mysql-2 to gke-jen-ps-523-f00253e-4-default-pool-e964acce-w8zz default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:23:29 +0000 UTC Normal Pod gr-demand-backup-mysql-2 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-92a84e98-ed7d-48f3-86c5-a42d98b84cc4" attachdetach-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:23:30 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:23:31 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 138.391268ms (138.406743ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:23:31 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:23:31 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:23:32 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:23:32 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 94.39306ms (94.408307ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:23:32 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:23:32 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:23:32 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:23:32 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 86.939041ms (86.969664ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:23:32 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:23:33 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:03 +0000 UTC Warning Pod gr-demand-backup-mysql-2.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:03 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:03 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 97.275689ms (97.306255ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:47 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-2jc8q Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-router-6498c5b898-2jc8q to gke-jen-ps-523-f00253e-4-default-pool-e964acce-w8zz default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:47 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-dvbl9 Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-router-6498c5b898-dvbl9 to gke-jen-ps-523-f00253e-4-default-pool-e964acce-7189 default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:47 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-hdjbd Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-router-6498c5b898-hdjbd to gke-jen-ps-523-f00253e-4-default-pool-e964acce-bq8g default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:47 +0000 UTC Normal ReplicaSet.apps gr-demand-backup-router-6498c5b898 SuccessfulCreate Created pod: gr-demand-backup-router-6498c5b898-hdjbd replicaset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:47 +0000 UTC Normal ReplicaSet.apps gr-demand-backup-router-6498c5b898 SuccessfulCreate Created pod: gr-demand-backup-router-6498c5b898-dvbl9 replicaset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:47 +0000 UTC Normal ReplicaSet.apps gr-demand-backup-router-6498c5b898 SuccessfulCreate Created pod: gr-demand-backup-router-6498c5b898-2jc8q replicaset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:48 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-2jc8q.spec.initContainers{router-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:48 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-2jc8q.spec.initContainers{router-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 98.08021ms (98.09646ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:48 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-2jc8q.spec.initContainers{router-init} Created Created container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:48 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-2jc8q.spec.initContainers{router-init} Started Started container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:48 +0000 UTC Warning Pod gr-demand-backup-router-6498c5b898-dvbl9 FailedMount MountVolume.SetUp failed for volume "config" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:48 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-hdjbd.spec.initContainers{router-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:48 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-hdjbd.spec.initContainers{router-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 96.82764ms (96.844635ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:48 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-hdjbd.spec.initContainers{router-init} Created Created container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:48 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-hdjbd.spec.initContainers{router-init} Started Started container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:49 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-2jc8q.spec.containers{router} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-router" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:49 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-2jc8q.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 94.982154ms (94.997728ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:49 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-dvbl9.spec.initContainers{router-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:49 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-dvbl9.spec.initContainers{router-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 85.421646ms (85.435733ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:50 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-2jc8q.spec.containers{router} Created Created container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:50 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-2jc8q.spec.containers{router} Started Started container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:50 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-dvbl9.spec.initContainers{router-init} Created Created container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:50 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-dvbl9.spec.initContainers{router-init} Started Started container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:50 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-hdjbd.spec.containers{router} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-router" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:50 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-hdjbd.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 88.216046ms (88.22986ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:50 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-hdjbd.spec.containers{router} Created Created container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:50 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-hdjbd.spec.containers{router} Started Started container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:51 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-dvbl9.spec.containers{router} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-router" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:51 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-dvbl9.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 77.474946ms (77.494815ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:51 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-dvbl9.spec.containers{router} Created Created container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:24:51 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-dvbl9.spec.containers{router} Started Started container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:25:13 +0000 UTC Normal Pod xb-gr-demand-backup-gcp-gcp-cs-q7qd4 Scheduled Successfully assigned kuttl-test-crack-yak/xb-gr-demand-backup-gcp-gcp-cs-q7qd4 to gke-jen-ps-523-f00253e-4-default-pool-e964acce-bq8g default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:25:13 +0000 UTC Normal Pod xb-gr-demand-backup-gcp-gcp-cs-q7qd4.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:25:13 +0000 UTC Normal Job.batch xb-gr-demand-backup-gcp-gcp-cs SuccessfulCreate Created pod: xb-gr-demand-backup-gcp-gcp-cs-q7qd4 job-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:25:14 +0000 UTC Normal Pod xb-gr-demand-backup-gcp-gcp-cs-q7qd4.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 100.814218ms (100.832634ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:25:14 +0000 UTC Normal Pod xb-gr-demand-backup-gcp-gcp-cs-q7qd4.spec.initContainers{xtrabackup-init} Created Created container xtrabackup-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:25:14 +0000 UTC Normal Pod xb-gr-demand-backup-gcp-gcp-cs-q7qd4.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:25:16 +0000 UTC Normal Pod xb-gr-demand-backup-gcp-gcp-cs-q7qd4.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:25:16 +0000 UTC Normal Pod xb-gr-demand-backup-gcp-gcp-cs-q7qd4.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 95.254729ms (95.285958ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:25:16 +0000 UTC Normal Pod xb-gr-demand-backup-gcp-gcp-cs-q7qd4.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:25:16 +0000 UTC Normal Pod xb-gr-demand-backup-gcp-gcp-cs-q7qd4.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:25:23 +0000 UTC Normal Job.batch xb-gr-demand-backup-gcp-gcp-cs Completed Job completed job-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:25:45 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup} Killing Stopping container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:25:45 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:25:46 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-2jc8q.spec.containers{router} Killing Stopping container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:25:46 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-dvbl9.spec.containers{router} Killing Stopping container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:25:46 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-hdjbd.spec.containers{router} Killing Stopping container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:25:46 +0000 UTC Normal ReplicaSet.apps gr-demand-backup-router-6498c5b898 SuccessfulDelete Deleted pod: gr-demand-backup-router-6498c5b898-dvbl9 replicaset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:25:46 +0000 UTC Normal ReplicaSet.apps gr-demand-backup-router-6498c5b898 SuccessfulDelete Deleted pod: gr-demand-backup-router-6498c5b898-hdjbd replicaset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:25:46 +0000 UTC Normal ReplicaSet.apps gr-demand-backup-router-6498c5b898 SuccessfulDelete Deleted pod: gr-demand-backup-router-6498c5b898-2jc8q replicaset-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:25:50 +0000 UTC Warning Pod gr-demand-backup-mysql-2.spec.containers{mysql} Unhealthy Readiness probe failed: 2024/02/05 14:25:50 readiness check failed: connect to db: ping DB: dial tcp 10.208.42.13:33062: connect: connection refused kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:25:53 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup} Killing Stopping container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:25:53 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:25:57 +0000 UTC Warning Pod gr-demand-backup-mysql-1.spec.containers{mysql} Unhealthy Readiness probe failed: 2024/02/05 14:25:57 readiness check failed: connect to db: ping DB: dial tcp 10.208.40.15:33062: connect: connection refused kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:26:00 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:26:00 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup} Killing Stopping container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:26:03 +0000 UTC Warning Pod gr-demand-backup-mysql-0.spec.containers{mysql} Unhealthy Readiness probe failed: [mysql] 2024/02/05 14:26:03 packets.go:37: read tcp 10.208.41.10:59344->10.208.41.10:33062: read: connection reset by peer 2024/02/05 14:26:03 readiness check failed: connect to db: ping DB: dial tcp 10.208.41.10:33062: connect: connection refused kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:26:07 +0000 UTC Warning Pod gr-demand-backup-mysql-0.spec.containers{mysql} Unhealthy Readiness probe failed: 2024/02/05 14:26:07 readiness check failed: connect to db: ping DB: dial tcp 10.208.41.10:33062: connect: connection refused kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:26:13 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-gcp-mvl6s Scheduled Successfully assigned kuttl-test-crack-yak/xb-restore-gr-demand-backup-restore-gcp-mvl6s to gke-jen-ps-523-f00253e-4-default-pool-e964acce-7189 default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:26:13 +0000 UTC Warning Pod xb-restore-gr-demand-backup-restore-gcp-mvl6s FailedAttachVolume Multi-Attach error for volume "pvc-b0d7a2e1-08de-43c6-a5c6-18a6360e6a75" Volume is already exclusively attached to one node and can't be attached to another attachdetach-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:26:13 +0000 UTC Normal Job.batch xb-restore-gr-demand-backup-restore-gcp SuccessfulCreate Created pod: xb-restore-gr-demand-backup-restore-gcp-mvl6s job-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:26:28 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-gcp-mvl6s SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-b0d7a2e1-08de-43c6-a5c6-18a6360e6a75" attachdetach-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:26:30 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-gcp-mvl6s.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:26:30 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-gcp-mvl6s.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 100.238344ms (100.255412ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:26:30 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-gcp-mvl6s.spec.initContainers{xtrabackup-init} Created Created container xtrabackup-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:26:30 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-gcp-mvl6s.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:26:31 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-gcp-mvl6s.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:26:31 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-gcp-mvl6s.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 80.865336ms (80.872995ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:26:31 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-gcp-mvl6s.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:26:31 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-gcp-mvl6s.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:26:45 +0000 UTC Normal Job.batch xb-restore-gr-demand-backup-restore-gcp Completed Job completed job-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:26:48 +0000 UTC Normal Pod gr-demand-backup-mysql-0 Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-mysql-0 to gke-jen-ps-523-f00253e-4-default-pool-e964acce-bq8g default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:02 +0000 UTC Normal Pod gr-demand-backup-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-b0d7a2e1-08de-43c6-a5c6-18a6360e6a75" attachdetach-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:03 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:03 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 98.485709ms (98.50603ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:03 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:03 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:05 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:05 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 77.725121ms (77.753124ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:05 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:05 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:05 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:05 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 76.323505ms (76.330498ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:05 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:05 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:38 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-1 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:38 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-1 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:38 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-1 Provisioning External provisioner is provisioning volume for claim "kuttl-test-crack-yak/datadir-gr-demand-backup-mysql-1" pd.csi.storage.gke.io_gke-77d1c4fac69c4c2a896d-0b77-45e4-vm_8a268d8f-c2f3-4090-b361-6efbcaeaaace logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:42 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-1 ProvisioningSucceeded Successfully provisioned volume pvc-c537ca90-ebef-405d-86e9-9301939a5b4f pd.csi.storage.gke.io_gke-77d1c4fac69c4c2a896d-0b77-45e4-vm_8a268d8f-c2f3-4090-b361-6efbcaeaaace logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:42 +0000 UTC Normal Pod gr-demand-backup-mysql-1 Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-mysql-1 to gke-jen-ps-523-f00253e-4-default-pool-e964acce-w8zz default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:43 +0000 UTC Warning Pod gr-demand-backup-mysql-1 FailedMount MountVolume.SetUp failed for volume "users" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:43 +0000 UTC Warning Pod gr-demand-backup-mysql-1 FailedMount MountVolume.SetUp failed for volume "tls" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:46 +0000 UTC Normal Pod gr-demand-backup-mysql-1 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-c537ca90-ebef-405d-86e9-9301939a5b4f" attachdetach-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:49 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:49 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 156.625155ms (156.64566ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:49 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:49 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:51 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:51 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 81.84462ms (81.853221ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:51 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:51 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:51 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:51 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 87.592703ms (87.60745ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:51 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:27:51 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:28:21 +0000 UTC Warning Pod gr-demand-backup-mysql-1.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:28:21 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:28:21 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 90.483086ms (90.49714ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:28:54 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-2 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:28:54 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-2 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:28:54 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-2 Provisioning External provisioner is provisioning volume for claim "kuttl-test-crack-yak/datadir-gr-demand-backup-mysql-2" pd.csi.storage.gke.io_gke-77d1c4fac69c4c2a896d-0b77-45e4-vm_8a268d8f-c2f3-4090-b361-6efbcaeaaace logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:28:58 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-2 ProvisioningSucceeded Successfully provisioned volume pvc-6aecaf84-9944-420e-95b1-d4c7e653332e pd.csi.storage.gke.io_gke-77d1c4fac69c4c2a896d-0b77-45e4-vm_8a268d8f-c2f3-4090-b361-6efbcaeaaace logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:28:58 +0000 UTC Normal Pod gr-demand-backup-mysql-2 Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-mysql-2 to gke-jen-ps-523-f00253e-4-default-pool-e964acce-7189 default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:29:05 +0000 UTC Normal Pod gr-demand-backup-mysql-2 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-6aecaf84-9944-420e-95b1-d4c7e653332e" attachdetach-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:29:07 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:29:07 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 103.320386ms (103.33449ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:29:07 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:29:07 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:29:09 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:29:09 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 81.700481ms (81.713468ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:29:09 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:29:09 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:29:09 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:29:09 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 89.145067ms (89.152197ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:29:09 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:29:09 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:29:39 +0000 UTC Warning Pod gr-demand-backup-mysql-2.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:29:39 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:29:39 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 105.060062ms (105.083012ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:20 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-8vhs6 Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-router-6498c5b898-8vhs6 to gke-jen-ps-523-f00253e-4-default-pool-e964acce-7189 default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:20 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-bqpkq Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-router-6498c5b898-bqpkq to gke-jen-ps-523-f00253e-4-default-pool-e964acce-w8zz default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:20 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-v46fl Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-router-6498c5b898-v46fl to gke-jen-ps-523-f00253e-4-default-pool-e964acce-bq8g default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:20 +0000 UTC Normal ReplicaSet.apps gr-demand-backup-router-6498c5b898 SuccessfulCreate (combined from similar events): Created pod: gr-demand-backup-router-6498c5b898-c6kjq logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:21 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-8vhs6.spec.initContainers{router-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:21 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-8vhs6.spec.initContainers{router-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 93.073001ms (93.088502ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:21 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-8vhs6.spec.initContainers{router-init} Created Created container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:21 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-8vhs6.spec.initContainers{router-init} Started Started container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:21 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-bqpkq.spec.initContainers{router-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:21 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-bqpkq.spec.initContainers{router-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 105.698823ms (105.712534ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:21 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-bqpkq.spec.initContainers{router-init} Created Created container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:21 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-bqpkq.spec.initContainers{router-init} Started Started container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:21 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-v46fl.spec.initContainers{router-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:21 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-v46fl.spec.initContainers{router-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 123.363015ms (123.370617ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:21 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-v46fl.spec.initContainers{router-init} Created Created container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:21 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-v46fl.spec.initContainers{router-init} Started Started container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:22 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-bqpkq.spec.containers{router} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-router" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:22 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-bqpkq.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 78.639271ms (78.652936ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:22 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-bqpkq.spec.containers{router} Created Created container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:22 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-bqpkq.spec.containers{router} Started Started container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:22 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-v46fl.spec.containers{router} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-router" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:23 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-8vhs6.spec.containers{router} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-router" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:23 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-8vhs6.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 107.631368ms (107.650396ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:23 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-8vhs6.spec.containers{router} Created Created container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:23 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-8vhs6.spec.containers{router} Started Started container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:23 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-v46fl.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 123.414495ms (123.433513ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:23 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-v46fl.spec.containers{router} Created Created container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:23 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-v46fl.spec.containers{router} Started Started container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:46 +0000 UTC Normal Pod xb-gr-demand-backup-azure-azure-blob-qp5ng Scheduled Successfully assigned kuttl-test-crack-yak/xb-gr-demand-backup-azure-azure-blob-qp5ng to gke-jen-ps-523-f00253e-4-default-pool-e964acce-bq8g default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:46 +0000 UTC Normal Job.batch xb-gr-demand-backup-azure-azure-blob SuccessfulCreate Created pod: xb-gr-demand-backup-azure-azure-blob-qp5ng job-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:47 +0000 UTC Normal Pod xb-gr-demand-backup-azure-azure-blob-qp5ng.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:47 +0000 UTC Normal Pod xb-gr-demand-backup-azure-azure-blob-qp5ng.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 101.870259ms (101.884265ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:47 +0000 UTC Normal Pod xb-gr-demand-backup-azure-azure-blob-qp5ng.spec.initContainers{xtrabackup-init} Created Created container xtrabackup-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:47 +0000 UTC Normal Pod xb-gr-demand-backup-azure-azure-blob-qp5ng.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:49 +0000 UTC Normal Pod xb-gr-demand-backup-azure-azure-blob-qp5ng.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:49 +0000 UTC Normal Pod xb-gr-demand-backup-azure-azure-blob-qp5ng.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 103.632765ms (103.648539ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:49 +0000 UTC Normal Pod xb-gr-demand-backup-azure-azure-blob-qp5ng.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:49 +0000 UTC Normal Pod xb-gr-demand-backup-azure-azure-blob-qp5ng.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:30:56 +0000 UTC Normal Job.batch xb-gr-demand-backup-azure-azure-blob Completed Job completed job-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:15 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup} Killing Stopping container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:15 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:16 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-8vhs6.spec.containers{router} Killing Stopping container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:16 +0000 UTC Warning Pod gr-demand-backup-router-6498c5b898-8vhs6.spec.containers{router} Unhealthy Readiness probe failed: {"isAlive":true} kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:16 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-bqpkq.spec.containers{router} Killing Stopping container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:16 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-v46fl.spec.containers{router} Killing Stopping container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:16 +0000 UTC Normal ReplicaSet.apps gr-demand-backup-router-6498c5b898 SuccessfulDelete (combined from similar events): Deleted pod: gr-demand-backup-router-6498c5b898-bqpkq logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:18 +0000 UTC Warning Pod gr-demand-backup-mysql-2.spec.containers{mysql} Unhealthy Readiness probe failed: [mysql] 2024/02/05 14:31:18 packets.go:37: read tcp 10.208.40.18:57906->10.208.40.18:33062: read: connection reset by peer 2024/02/05 14:31:18 readiness check failed: connect to db: ping DB: dial tcp 10.208.40.18:33062: connect: connection refused kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:21 +0000 UTC Warning Pod gr-demand-backup-mysql-2.spec.containers{mysql} Unhealthy Readiness probe failed: 2024/02/05 14:31:21 readiness check failed: connect to db: ping DB: dial tcp 10.208.40.18:33062: connect: connection refused kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:23 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup} Killing Stopping container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:23 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:26 +0000 UTC Warning Pod gr-demand-backup-mysql-1.spec.containers{mysql} Unhealthy Readiness probe failed: [mysql] 2024/02/05 14:31:26 packets.go:37: read tcp 10.208.42.15:35738->10.208.42.15:33062: read: connection reset by peer 2024/02/05 14:31:26 readiness check failed: connect to db: ping DB: dial tcp 10.208.42.15:33062: connect: connection refused kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:28 +0000 UTC Warning Pod gr-demand-backup-mysql-1.spec.containers{mysql} Unhealthy Readiness probe failed: 2024/02/05 14:31:28 readiness check failed: connect to db: ping DB: dial tcp 10.208.42.15:33062: connect: connection refused kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:30 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:30 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup} Killing Stopping container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:33 +0000 UTC Warning Pod gr-demand-backup-mysql-0.spec.containers{mysql} Unhealthy Readiness probe failed: [mysql] 2024/02/05 14:31:33 packets.go:37: read tcp 10.208.41.13:36010->10.208.41.13:33062: read: connection reset by peer 2024/02/05 14:31:33 readiness check failed: connect to db: ping DB: dial tcp 10.208.41.13:33062: connect: connection refused kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:41 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-azure-q445g Scheduled Successfully assigned kuttl-test-crack-yak/xb-restore-gr-demand-backup-restore-azure-q445g to gke-jen-ps-523-f00253e-4-default-pool-e964acce-7189 default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:41 +0000 UTC Normal Job.batch xb-restore-gr-demand-backup-restore-azure SuccessfulCreate Created pod: xb-restore-gr-demand-backup-restore-azure-q445g job-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:54 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-azure-q445g SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-b0d7a2e1-08de-43c6-a5c6-18a6360e6a75" attachdetach-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:56 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-azure-q445g.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:56 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-azure-q445g.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 94.986739ms (95.001658ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:56 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-azure-q445g.spec.initContainers{xtrabackup-init} Created Created container xtrabackup-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:56 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-azure-q445g.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:57 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-azure-q445g.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:57 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-azure-q445g.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 95.951194ms (95.966849ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:57 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-azure-q445g.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:31:57 +0000 UTC Normal Pod xb-restore-gr-demand-backup-restore-azure-q445g.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:32:09 +0000 UTC Normal Job.batch xb-restore-gr-demand-backup-restore-azure Completed Job completed job-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:32:17 +0000 UTC Normal Pod gr-demand-backup-mysql-0 Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-mysql-0 to gke-jen-ps-523-f00253e-4-default-pool-e964acce-bq8g default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:32:29 +0000 UTC Normal Pod gr-demand-backup-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-b0d7a2e1-08de-43c6-a5c6-18a6360e6a75" attachdetach-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:32:31 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:32:31 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 128.628105ms (128.645266ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:32:31 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:32:31 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:32:33 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:32:33 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 80.195231ms (80.21385ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:32:33 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:32:33 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:32:33 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:32:33 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 79.985629ms (79.993975ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:32:33 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:32:33 +0000 UTC Normal Pod gr-demand-backup-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:33:05 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-1 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:33:06 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-1 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:33:06 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-1 Provisioning External provisioner is provisioning volume for claim "kuttl-test-crack-yak/datadir-gr-demand-backup-mysql-1" pd.csi.storage.gke.io_gke-77d1c4fac69c4c2a896d-0b77-45e4-vm_8a268d8f-c2f3-4090-b361-6efbcaeaaace logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:33:09 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-1 ProvisioningSucceeded Successfully provisioned volume pvc-9a998a65-3060-491e-872f-805f44f3a752 pd.csi.storage.gke.io_gke-77d1c4fac69c4c2a896d-0b77-45e4-vm_8a268d8f-c2f3-4090-b361-6efbcaeaaace logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:33:10 +0000 UTC Normal Pod gr-demand-backup-mysql-1 Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-mysql-1 to gke-jen-ps-523-f00253e-4-default-pool-e964acce-w8zz default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:33:17 +0000 UTC Normal Pod gr-demand-backup-mysql-1 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-9a998a65-3060-491e-872f-805f44f3a752" attachdetach-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:33:18 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:33:18 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 105.630431ms (105.64651ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:33:18 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:33:19 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:33:20 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:33:20 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 77.936255ms (77.949894ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:33:20 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:33:20 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:33:20 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:33:20 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 89.487385ms (89.494102ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:33:20 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:33:20 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:33:51 +0000 UTC Warning Pod gr-demand-backup-mysql-1.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:33:51 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:33:51 +0000 UTC Normal Pod gr-demand-backup-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 82.768629ms (82.777223ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:34:23 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-2 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:34:23 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-2 Provisioning External provisioner is provisioning volume for claim "kuttl-test-crack-yak/datadir-gr-demand-backup-mysql-2" pd.csi.storage.gke.io_gke-77d1c4fac69c4c2a896d-0b77-45e4-vm_8a268d8f-c2f3-4090-b361-6efbcaeaaace logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:34:23 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-2 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:34:27 +0000 UTC Normal PersistentVolumeClaim datadir-gr-demand-backup-mysql-2 ProvisioningSucceeded Successfully provisioned volume pvc-a269d837-9910-4b13-bbc3-46357516be83 pd.csi.storage.gke.io_gke-77d1c4fac69c4c2a896d-0b77-45e4-vm_8a268d8f-c2f3-4090-b361-6efbcaeaaace logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:34:27 +0000 UTC Normal Pod gr-demand-backup-mysql-2 Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-mysql-2 to gke-jen-ps-523-f00253e-4-default-pool-e964acce-7189 default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:34:35 +0000 UTC Normal Pod gr-demand-backup-mysql-2 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-a269d837-9910-4b13-bbc3-46357516be83" attachdetach-controller logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:34:38 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:34:38 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 96.474392ms (96.488874ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:34:38 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:34:38 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:34:40 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:34:40 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 77.477278ms (77.494218ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:34:40 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:34:40 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:34:40 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:34:40 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 82.689317ms (82.699412ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:34:40 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:34:40 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:10 +0000 UTC Warning Pod gr-demand-backup-mysql-2.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:10 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:10 +0000 UTC Normal Pod gr-demand-backup-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 81.864999ms (81.880433ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:51 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-cwjll Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-router-6498c5b898-cwjll to gke-jen-ps-523-f00253e-4-default-pool-e964acce-bq8g default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:52 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-c6kjq Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-router-6498c5b898-c6kjq to gke-jen-ps-523-f00253e-4-default-pool-e964acce-w8zz default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:52 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-cwjll.spec.initContainers{router-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:52 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-cwjll.spec.initContainers{router-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 97.541952ms (97.551272ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:52 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-cwjll.spec.initContainers{router-init} Created Created container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:52 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-cwjll.spec.initContainers{router-init} Started Started container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:52 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-n9glp Scheduled Successfully assigned kuttl-test-crack-yak/gr-demand-backup-router-6498c5b898-n9glp to gke-jen-ps-523-f00253e-4-default-pool-e964acce-7189 default-scheduler logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:52 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-n9glp.spec.initContainers{router-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:52 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-n9glp.spec.initContainers{router-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 100.460314ms (100.476467ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:53 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-c6kjq.spec.initContainers{router-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:53 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-c6kjq.spec.initContainers{router-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 97.259658ms (97.278019ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:53 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-c6kjq.spec.initContainers{router-init} Created Created container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:53 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-c6kjq.spec.initContainers{router-init} Started Started container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:53 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-n9glp.spec.initContainers{router-init} Created Created container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:53 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-n9glp.spec.initContainers{router-init} Started Started container router-init kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:54 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-c6kjq.spec.containers{router} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-router" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:54 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-c6kjq.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 88.762068ms (88.778275ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:54 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-c6kjq.spec.containers{router} Created Created container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:54 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-c6kjq.spec.containers{router} Started Started container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:54 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-cwjll.spec.containers{router} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-router" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:54 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-cwjll.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 86.44211ms (86.450906ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:54 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-cwjll.spec.containers{router} Created Created container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:54 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-n9glp.spec.containers{router} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-router" kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:54 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-n9glp.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 88.084233ms (88.09965ms including waiting) kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:54 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-n9glp.spec.containers{router} Created Created container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:54 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-n9glp.spec.containers{router} Started Started container router kubelet logger.go:42: 14:36:15 | gr-demand-backup | 2024-02-05 14:35:55 +0000 UTC Normal Pod gr-demand-backup-router-6498c5b898-cwjll.spec.containers{router} Started Started container router kubelet logger.go:42: 14:36:16 | gr-demand-backup | Deleting namespace: kuttl-test-crack-yak === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (1850.34s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/gr-demand-backup (1841.75s) PASS