=== RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. I0809 12:05:39.510632 12878 request.go:682] Waited for 1.041594787s due to client-side throttling, not priority and fairness, request: GET:https://34.70.140.115/apis/internal.autoscaling.gke.io/v1alpha1?timeout=32s harness.go:275: Successful connection to cluster at: https://34.70.140.115 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 180 seconds for each step harness.go:372: testsuite: e2e-tests/tests has 26 tests === RUN kuttl/harness === RUN kuttl/harness/demand-backup === PAUSE kuttl/harness/demand-backup === CONT kuttl/harness/demand-backup logger.go:42: 12:05:46 | demand-backup | Creating namespace: kuttl-test-cosmic-ghoul logger.go:42: 12:05:47 | demand-backup/0-minio-secret | starting test step 0-minio-secret logger.go:42: 12:05:48 | demand-backup/0-minio-secret | Secret:kuttl-test-cosmic-ghoul/minio-secret created logger.go:42: 12:05:48 | demand-backup/0-minio-secret | test step completed 0-minio-secret logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | starting test step 1-deploy-operator logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | running command: [sh -c set -o errexit set -o xtrace source ../../functions kubectl -n "${NAMESPACE}" apply -f "${TESTS_CONFIG_DIR}/cloud-secret.yml" deploy_operator deploy_non_tls_cluster_secrets deploy_tls_cluster_secrets deploy_client deploy_minio] logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | + source ../../functions logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ realpath ../../.. logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | ++++ mktemp -d logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ export TEMP_DIR=/tmp/tmp.7j6W4kFVJt logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ TEMP_DIR=/tmp/tmp.7j6W4kFVJt logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ GIT_BRANCH=PR-339 logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ VERSION=PR-339-3a8483b logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | ++++ which gdate logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | ++++ which date logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ date=/usr/bin/date logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ command -v oc logger.go:42: 12:05:48 | demand-backup/1-deploy-operator | +++ oc get projects logger.go:42: 12:05:55 | demand-backup/1-deploy-operator | error: the server doesn't have a resource type "projects" logger.go:42: 12:05:55 | demand-backup/1-deploy-operator | +++ grep '^minikube' logger.go:42: 12:05:55 | demand-backup/1-deploy-operator | +++ kubectl get nodes logger.go:42: 12:05:56 | demand-backup/1-deploy-operator | ++++ pwd logger.go:42: 12:05:56 | demand-backup/1-deploy-operator | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/demand-backup logger.go:42: 12:05:56 | demand-backup/1-deploy-operator | ++ test_name=demand-backup logger.go:42: 12:05:56 | demand-backup/1-deploy-operator | + kubectl -n kuttl-test-cosmic-ghoul apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/cloud-secret.yml logger.go:42: 12:05:59 | demand-backup/1-deploy-operator | secret/aws-s3-secret created logger.go:42: 12:05:59 | demand-backup/1-deploy-operator | secret/gcp-cs-secret created logger.go:42: 12:06:00 | demand-backup/1-deploy-operator | secret/azure-secret created logger.go:42: 12:06:00 | demand-backup/1-deploy-operator | + deploy_operator logger.go:42: 12:06:00 | demand-backup/1-deploy-operator | + kubectl -n kuttl-test-cosmic-ghoul apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/crd.yaml logger.go:42: 12:06:01 | demand-backup/1-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlbackups.ps.percona.com serverside-applied logger.go:42: 12:06:01 | demand-backup/1-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlrestores.ps.percona.com serverside-applied logger.go:42: 12:06:03 | demand-backup/1-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqls.ps.percona.com serverside-applied logger.go:42: 12:06:03 | demand-backup/1-deploy-operator | + kubectl -n kuttl-test-cosmic-ghoul apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/rbac.yaml logger.go:42: 12:06:04 | demand-backup/1-deploy-operator | serviceaccount/percona-server-mysql-operator created logger.go:42: 12:06:05 | demand-backup/1-deploy-operator | serviceaccount/percona-server-mysql-operator-orchestrator created logger.go:42: 12:06:05 | demand-backup/1-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 12:06:06 | demand-backup/1-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 12:06:06 | demand-backup/1-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 12:06:06 | demand-backup/1-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 12:06:07 | demand-backup/1-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 12:06:07 | demand-backup/1-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 12:06:07 | demand-backup/1-deploy-operator | + yq eval '(select(documentIndex==1).spec.template.spec.containers[] | select(.name=="manager").env[] | select(.name=="DISABLE_TELEMETRY").value) = "true"' logger.go:42: 12:06:07 | demand-backup/1-deploy-operator | + kubectl -n kuttl-test-cosmic-ghoul apply -f - logger.go:42: 12:06:07 | demand-backup/1-deploy-operator | ++ printf 'select(documentIndex==1).spec.template.spec.containers[0].image="%s"' perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:06:07 | demand-backup/1-deploy-operator | + yq eval 'select(documentIndex==1).spec.template.spec.containers[0].image="perconalab/percona-server-mysql-operator:PR-339-3a8483b"' /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/operator.yaml logger.go:42: 12:06:10 | demand-backup/1-deploy-operator | configmap/percona-server-mysql-operator-config created logger.go:42: 12:06:11 | demand-backup/1-deploy-operator | deployment.apps/percona-server-mysql-operator created logger.go:42: 12:06:11 | demand-backup/1-deploy-operator | + deploy_non_tls_cluster_secrets logger.go:42: 12:06:11 | demand-backup/1-deploy-operator | + kubectl -n kuttl-test-cosmic-ghoul apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/secrets.yaml logger.go:42: 12:06:12 | demand-backup/1-deploy-operator | secret/test-secrets created logger.go:42: 12:06:12 | demand-backup/1-deploy-operator | + deploy_tls_cluster_secrets logger.go:42: 12:06:12 | demand-backup/1-deploy-operator | + kubectl -n kuttl-test-cosmic-ghoul apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/ssl-secret.yaml logger.go:42: 12:06:13 | demand-backup/1-deploy-operator | secret/test-ssl created logger.go:42: 12:06:13 | demand-backup/1-deploy-operator | + deploy_client logger.go:42: 12:06:13 | demand-backup/1-deploy-operator | + kubectl -n kuttl-test-cosmic-ghoul apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/client.yaml logger.go:42: 12:06:14 | demand-backup/1-deploy-operator | pod/mysql-client created logger.go:42: 12:06:14 | demand-backup/1-deploy-operator | + deploy_minio logger.go:42: 12:06:14 | demand-backup/1-deploy-operator | ++ kubectl -n kuttl-test-cosmic-ghoul get secret minio-secret -o 'jsonpath={.data.AWS_ACCESS_KEY_ID}' logger.go:42: 12:06:14 | demand-backup/1-deploy-operator | ++ base64 -d logger.go:42: 12:06:15 | demand-backup/1-deploy-operator | + accessKey='some-access$\n"-key' logger.go:42: 12:06:15 | demand-backup/1-deploy-operator | ++ kubectl -n kuttl-test-cosmic-ghoul get secret minio-secret -o 'jsonpath={.data.AWS_SECRET_ACCESS_KEY}' logger.go:42: 12:06:15 | demand-backup/1-deploy-operator | ++ base64 -d logger.go:42: 12:06:16 | demand-backup/1-deploy-operator | + secretKey='some-$\n"secret-key' logger.go:42: 12:06:16 | demand-backup/1-deploy-operator | + helm uninstall -n kuttl-test-cosmic-ghoul minio-service logger.go:42: 12:06:16 | demand-backup/1-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 12:06:16 | demand-backup/1-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 12:06:16 | demand-backup/1-deploy-operator | Error: uninstall: Release not loaded: minio-service: release: not found logger.go:42: 12:06:16 | demand-backup/1-deploy-operator | + : logger.go:42: 12:06:16 | demand-backup/1-deploy-operator | + helm repo remove minio logger.go:42: 12:06:16 | demand-backup/1-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 12:06:16 | demand-backup/1-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 12:06:16 | demand-backup/1-deploy-operator | Error: no repositories configured logger.go:42: 12:06:16 | demand-backup/1-deploy-operator | + : logger.go:42: 12:06:16 | demand-backup/1-deploy-operator | + helm repo add minio https://helm.min.io/ logger.go:42: 12:06:16 | demand-backup/1-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 12:06:16 | demand-backup/1-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 12:06:16 | demand-backup/1-deploy-operator | "minio" has been added to your repositories logger.go:42: 12:06:16 | demand-backup/1-deploy-operator | ++ printf %q 'some-access$\n"-key' logger.go:42: 12:06:16 | demand-backup/1-deploy-operator | ++ printf %q 'some-$\n"secret-key' logger.go:42: 12:06:16 | demand-backup/1-deploy-operator | + retry 10 60 helm install minio-service -n kuttl-test-cosmic-ghoul --version 8.0.5 --set 'accessKey=some-access\$\\n\"-key' --set 'secretKey=some-\$\\n\"secret-key' --set service.type=ClusterIP --set configPathmc=/tmp/.minio/ --set persistence.size=2G --set environment.MINIO_REGION=us-east-1 --set environment.MINIO_HTTP_TRACE=/tmp/trace.log --set securityContext.enabled=false minio/minio logger.go:42: 12:06:16 | demand-backup/1-deploy-operator | + local max=10 logger.go:42: 12:06:16 | demand-backup/1-deploy-operator | + local delay=60 logger.go:42: 12:06:16 | demand-backup/1-deploy-operator | + shift 2 logger.go:42: 12:06:16 | demand-backup/1-deploy-operator | + local n=1 logger.go:42: 12:06:16 | demand-backup/1-deploy-operator | + helm install minio-service -n kuttl-test-cosmic-ghoul --version 8.0.5 --set 'accessKey=some-access\$\\n\"-key' --set 'secretKey=some-\$\\n\"secret-key' --set service.type=ClusterIP --set configPathmc=/tmp/.minio/ --set persistence.size=2G --set environment.MINIO_REGION=us-east-1 --set environment.MINIO_HTTP_TRACE=/tmp/trace.log --set securityContext.enabled=false minio/minio logger.go:42: 12:06:16 | demand-backup/1-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 12:06:16 | demand-backup/1-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 12:06:17 | demand-backup/1-deploy-operator | E0809 12:06:17.783682 13586 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 12:06:17 | demand-backup/1-deploy-operator | E0809 12:06:17.901623 13586 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | NAME: minio-service logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | LAST DEPLOYED: Wed Aug 9 12:06:17 2023 logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | NAMESPACE: kuttl-test-cosmic-ghoul logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | STATUS: deployed logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | REVISION: 1 logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | TEST SUITE: None logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | NOTES: logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | Minio can be accessed via port 9000 on the following DNS name from within your cluster: logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | minio-service.kuttl-test-cosmic-ghoul.svc.cluster.local logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | To access Minio from localhost, run the below commands: logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | 1. export POD_NAME=$(kubectl get pods --namespace kuttl-test-cosmic-ghoul -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}") logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | 2. kubectl port-forward $POD_NAME 9000 --namespace kuttl-test-cosmic-ghoul logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/ logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | You can now access Minio server on http://localhost:9000. Follow the below steps to connect to Minio server with mc client: logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | 1. Download the Minio mc client - https://docs.minio.io/docs/minio-client-quickstart-guide logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | 2. Get the ACCESS_KEY=$(kubectl get secret minio-service -o jsonpath="{.data.accesskey}" | base64 --decode) and the SECRET_KEY=$(kubectl get secret minio-service -o jsonpath="{.data.secretkey}" | base64 --decode) logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | 3. mc alias set minio-service-local http://localhost:9000 "$ACCESS_KEY" "$SECRET_KEY" --api s3v4 logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | 4. mc ls minio-service-local logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | Alternately, you can use your browser or the Minio SDK to access the server - https://docs.minio.io/categories/17 logger.go:42: 12:06:21 | demand-backup/1-deploy-operator | ++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:06:22 | demand-backup/1-deploy-operator | + MINIO_POD=minio-service-6f47578c5b-2vr2g logger.go:42: 12:06:22 | demand-backup/1-deploy-operator | + wait_pod minio-service-6f47578c5b-2vr2g logger.go:42: 12:06:22 | demand-backup/1-deploy-operator | + local pod=minio-service-6f47578c5b-2vr2g logger.go:42: 12:06:22 | demand-backup/1-deploy-operator | + set +o xtrace logger.go:42: 12:06:36 | demand-backup/1-deploy-operator | minio-service-6f47578c5b-2vr2g........true logger.go:42: 12:06:36 | demand-backup/1-deploy-operator | + kubectl -n kuttl-test-cosmic-ghoul run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- bash -c 'AWS_ACCESS_KEY_ID='\''some-access$\n"-key'\'' AWS_SECRET_ACCESS_KEY='\''some-$\n"secret-key'\'' AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url http://minio-service:9000 s3 mb s3://operator-testing' logger.go:42: 12:06:42 | demand-backup/1-deploy-operator | If you don't see a command prompt, try pressing enter. logger.go:42: 12:06:51 | demand-backup/1-deploy-operator | pod "aws-cli" deleted logger.go:42: 12:06:53 | demand-backup/1-deploy-operator | test step completed 1-deploy-operator logger.go:42: 12:06:53 | demand-backup/2-create-cluster | starting test step 2-create-cluster logger.go:42: 12:06:53 | demand-backup/2-create-cluster | running command: [sh -c set -o errexit set -o xtrace source ../../functions get_cr \ | yq eval '.spec.mysql.clusterType="async"' - \ | yq eval '.spec.mysql.size=3' - \ | yq eval '.spec.proxy.haproxy.enabled=true' - \ | yq eval '.spec.proxy.haproxy.size=3' - \ | yq eval '.spec.orchestrator.enabled=true' - \ | yq eval '.spec.orchestrator.size=3' - \ | yq eval '.spec.backup.storages.minio.type="s3"' - \ | yq eval '.spec.backup.storages.minio.s3.bucket="operator-testing"' - \ | yq eval '.spec.backup.storages.minio.s3.credentialsSecret="minio-secret"' - \ | yq eval '.spec.backup.storages.minio.s3.endpointUrl="http://minio-service:9000"' - \ | yq eval '.spec.backup.storages.minio.s3.region="us-east-1"' - \ | yq eval '.spec.backup.storages.aws-s3.type="s3"' - \ | yq eval '.spec.backup.storages.aws-s3.verifyTLS=true' - \ | yq eval '.spec.backup.storages.aws-s3.s3.bucket="operator-testing"' - \ | yq eval '.spec.backup.storages.aws-s3.s3.credentialsSecret="aws-s3-secret"' - \ | yq eval '.spec.backup.storages.aws-s3.s3.region="us-east-1"' - \ | yq eval '.spec.backup.storages.aws-s3.s3.prefix="ps"' - \ | yq eval '.spec.backup.storages.gcp-cs.type="gcs"' - \ | yq eval '.spec.backup.storages.gcp-cs.verifyTLS=true' - \ | yq eval '.spec.backup.storages.gcp-cs.gcs.bucket="operator-testing"' - \ | yq eval '.spec.backup.storages.gcp-cs.gcs.credentialsSecret="gcp-cs-secret"' - \ | yq eval '.spec.backup.storages.gcp-cs.gcs.endpointUrl="https://storage.googleapis.com"' - \ | yq eval '.spec.backup.storages.gcp-cs.gcs.prefix="ps"' - \ | yq eval '.spec.backup.storages.azure-blob.type="azure"' - \ | yq eval '.spec.backup.storages.azure-blob.verifyTLS=true' - \ | yq eval '.spec.backup.storages.azure-blob.azure.containerName="operator-testing"' - \ | yq eval '.spec.backup.storages.azure-blob.azure.credentialsSecret="azure-secret"' - \ | yq eval '.spec.backup.storages.azure-blob.azure.prefix="ps"' - \ | kubectl -n "${NAMESPACE}" apply -f -] logger.go:42: 12:06:53 | demand-backup/2-create-cluster | + source ../../functions logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ realpath ../../.. logger.go:42: 12:06:53 | demand-backup/2-create-cluster | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:06:53 | demand-backup/2-create-cluster | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:06:53 | demand-backup/2-create-cluster | ++++ mktemp -d logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ export TEMP_DIR=/tmp/tmp.TIsjWciTvn logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ TEMP_DIR=/tmp/tmp.TIsjWciTvn logger.go:42: 12:06:53 | demand-backup/2-create-cluster | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ GIT_BRANCH=PR-339 logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ VERSION=PR-339-3a8483b logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:06:53 | demand-backup/2-create-cluster | ++++ which gdate logger.go:42: 12:06:53 | demand-backup/2-create-cluster | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:06:53 | demand-backup/2-create-cluster | ++++ which date logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ date=/usr/bin/date logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ command -v oc logger.go:42: 12:06:53 | demand-backup/2-create-cluster | +++ oc get projects logger.go:42: 12:06:58 | demand-backup/2-create-cluster | error: the server doesn't have a resource type "projects" logger.go:42: 12:06:58 | demand-backup/2-create-cluster | +++ grep '^minikube' logger.go:42: 12:06:58 | demand-backup/2-create-cluster | +++ kubectl get nodes logger.go:42: 12:06:59 | demand-backup/2-create-cluster | ++++ pwd logger.go:42: 12:06:59 | demand-backup/2-create-cluster | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/demand-backup logger.go:42: 12:06:59 | demand-backup/2-create-cluster | ++ test_name=demand-backup logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval .spec.mysql.size=3 - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + get_cr logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + local name_suffix= logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval .spec.backup.storages.azure-blob.verifyTLS=true - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + kubectl -n kuttl-test-cosmic-ghoul apply -f - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.azure-blob.azure.prefix="ps"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval .spec.proxy.haproxy.size=3 - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval .spec.proxy.haproxy.enabled=true - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.minio.type="s3"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval .spec.orchestrator.enabled=true - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.azure-blob.type="azure"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.minio.s3.bucket="operator-testing"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval .spec.orchestrator.size=3 - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.minio.s3.credentialsSecret="minio-secret"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.minio.s3.region="us-east-1"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.minio.s3.endpointUrl="http://minio-service:9000"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.sslSecretName="test-ssl"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.secretsName="test-secrets"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.aws-s3.s3.credentialsSecret="aws-s3-secret"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.aws-s3.type="s3"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval .spec.backup.storages.gcp-cs.verifyTLS=true - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.gcp-cs.gcs.bucket="operator-testing"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.gcp-cs.gcs.prefix="ps"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.upgradeOptions.apply="disabled"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval .spec.backup.storages.aws-s3.verifyTLS=true - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.gcp-cs.gcs.credentialsSecret="gcp-cs-secret"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.aws-s3.s3.bucket="operator-testing"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | ++ printf '.metadata.name="%s"' demand-backup logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.metadata.name="demand-backup"' /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/cr.yaml logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.aws-s3.s3.region="us-east-1"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.aws-s3.s3.prefix="ps"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.gcp-cs.type="gcs"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + '[' -n '' ']' logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | ++ printf '.spec.pmm.image="%s"' perconalab/pmm-client:dev-latest logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.pmm.image="perconalab/pmm-client:dev-latest"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | ++ printf '.spec.proxy.haproxy.image="%s"' perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:06:59 | demand-backup/2-create-cluster | ++ printf '.spec.initImage="%s"' perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.proxy.haproxy.image="perconalab/percona-server-mysql-operator:main-haproxy"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | ++ printf '.spec.orchestrator.image="%s"' perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.orchestrator.image="perconalab/percona-server-mysql-operator:main-orchestrator"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | ++ printf '.spec.proxy.router.image="%s"' perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.proxy.router.image="perconalab/percona-server-mysql-operator:main-router"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.gcp-cs.gcs.endpointUrl="https://storage.googleapis.com"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | ++ printf '.spec.backup.image="%s"' perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.backup.image="perconalab/percona-server-mysql-operator:main-backup"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | ++ printf '.spec.toolkit.image="%s"' perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.toolkit.image="perconalab/percona-server-mysql-operator:main-toolkit"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | ++ printf '.spec.mysql.image="%s"' perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.mysql.image="perconalab/percona-server-mysql-operator:main-psmysql"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.initImage="perconalab/percona-server-mysql-operator:PR-339-3a8483b"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.azure-blob.azure.containerName="operator-testing"' - logger.go:42: 12:06:59 | demand-backup/2-create-cluster | + yq eval '.spec.backup.storages.azure-blob.azure.credentialsSecret="azure-secret"' - logger.go:42: 12:07:00 | demand-backup/2-create-cluster | perconaservermysql.ps.percona.com/demand-backup created logger.go:42: 12:12:19 | demand-backup/2-create-cluster | test step completed 2-create-cluster logger.go:42: 12:12:19 | demand-backup/3-write-data | starting test step 3-write-data logger.go:42: 12:12:19 | demand-backup/3-write-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password" run_mysql \ "INSERT myDB.myTable (id) VALUES (100500)" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password"] logger.go:42: 12:12:19 | demand-backup/3-write-data | + source ../../functions logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ realpath ../../.. logger.go:42: 12:12:19 | demand-backup/3-write-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:12:19 | demand-backup/3-write-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:12:19 | demand-backup/3-write-data | ++++ mktemp -d logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ export TEMP_DIR=/tmp/tmp.QfNaRA53PL logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ TEMP_DIR=/tmp/tmp.QfNaRA53PL logger.go:42: 12:12:19 | demand-backup/3-write-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ GIT_BRANCH=PR-339 logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ VERSION=PR-339-3a8483b logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:12:19 | demand-backup/3-write-data | ++++ which gdate logger.go:42: 12:12:19 | demand-backup/3-write-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:12:19 | demand-backup/3-write-data | ++++ which date logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ date=/usr/bin/date logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ command -v oc logger.go:42: 12:12:19 | demand-backup/3-write-data | +++ oc get projects logger.go:42: 12:12:25 | demand-backup/3-write-data | error: the server doesn't have a resource type "projects" logger.go:42: 12:12:25 | demand-backup/3-write-data | +++ kubectl get nodes logger.go:42: 12:12:25 | demand-backup/3-write-data | +++ grep '^minikube' logger.go:42: 12:12:25 | demand-backup/3-write-data | ++++ pwd logger.go:42: 12:12:25 | demand-backup/3-write-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/demand-backup logger.go:42: 12:12:25 | demand-backup/3-write-data | ++ test_name=demand-backup logger.go:42: 12:12:25 | demand-backup/3-write-data | +++ get_cluster_name logger.go:42: 12:12:25 | demand-backup/3-write-data | +++ kubectl -n kuttl-test-cosmic-ghoul get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 12:12:26 | demand-backup/3-write-data | ++ get_haproxy_svc demand-backup logger.go:42: 12:12:26 | demand-backup/3-write-data | ++ local cluster=demand-backup logger.go:42: 12:12:26 | demand-backup/3-write-data | ++ echo demand-backup-haproxy logger.go:42: 12:12:26 | demand-backup/3-write-data | + run_mysql 'CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' '-h demand-backup-haproxy -uroot -proot_password' logger.go:42: 12:12:26 | demand-backup/3-write-data | + local 'command=CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' logger.go:42: 12:12:26 | demand-backup/3-write-data | + local 'uri=-h demand-backup-haproxy -uroot -proot_password' logger.go:42: 12:12:26 | demand-backup/3-write-data | + local pod= logger.go:42: 12:12:26 | demand-backup/3-write-data | ++ get_client_pod logger.go:42: 12:12:26 | demand-backup/3-write-data | ++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:12:27 | demand-backup/3-write-data | + client_pod=mysql-client logger.go:42: 12:12:27 | demand-backup/3-write-data | + wait_pod mysql-client logger.go:42: 12:12:27 | demand-backup/3-write-data | + local pod=mysql-client logger.go:42: 12:12:27 | demand-backup/3-write-data | + set +o xtrace logger.go:42: 12:12:28 | demand-backup/3-write-data | mysql-clienttrue logger.go:42: 12:12:28 | demand-backup/3-write-data | + kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" | mysql -sN -h demand-backup-haproxy -uroot -proot_password' logger.go:42: 12:12:28 | demand-backup/3-write-data | + sed -e 's/mysql: //' logger.go:42: 12:12:28 | demand-backup/3-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:12:29 | demand-backup/3-write-data | + : logger.go:42: 12:12:29 | demand-backup/3-write-data | +++ get_cluster_name logger.go:42: 12:12:29 | demand-backup/3-write-data | +++ kubectl -n kuttl-test-cosmic-ghoul get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 12:12:30 | demand-backup/3-write-data | ++ get_haproxy_svc demand-backup logger.go:42: 12:12:30 | demand-backup/3-write-data | ++ local cluster=demand-backup logger.go:42: 12:12:30 | demand-backup/3-write-data | ++ echo demand-backup-haproxy logger.go:42: 12:12:30 | demand-backup/3-write-data | + run_mysql 'INSERT myDB.myTable (id) VALUES (100500)' '-h demand-backup-haproxy -uroot -proot_password' logger.go:42: 12:12:30 | demand-backup/3-write-data | + local 'command=INSERT myDB.myTable (id) VALUES (100500)' logger.go:42: 12:12:30 | demand-backup/3-write-data | + local 'uri=-h demand-backup-haproxy -uroot -proot_password' logger.go:42: 12:12:30 | demand-backup/3-write-data | + local pod= logger.go:42: 12:12:30 | demand-backup/3-write-data | ++ get_client_pod logger.go:42: 12:12:30 | demand-backup/3-write-data | ++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:12:31 | demand-backup/3-write-data | + client_pod=mysql-client logger.go:42: 12:12:31 | demand-backup/3-write-data | + wait_pod mysql-client logger.go:42: 12:12:31 | demand-backup/3-write-data | + local pod=mysql-client logger.go:42: 12:12:31 | demand-backup/3-write-data | + set +o xtrace logger.go:42: 12:12:31 | demand-backup/3-write-data | mysql-clienttrue logger.go:42: 12:12:31 | demand-backup/3-write-data | + kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "INSERT myDB.myTable (id) VALUES (100500)" | mysql -sN -h demand-backup-haproxy -uroot -proot_password' logger.go:42: 12:12:31 | demand-backup/3-write-data | + sed -e 's/mysql: //' logger.go:42: 12:12:31 | demand-backup/3-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:12:33 | demand-backup/3-write-data | + : logger.go:42: 12:12:34 | demand-backup/3-write-data | test step completed 3-write-data logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | starting test step 4-move-primary-before-backup logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | running command: [sh -c set -o errexit set -o xtrace source ../../functions primary_pod_from_label="$(get_primary_from_label)" kubectl delete pod -n ${NAMESPACE} ${primary_pod_from_label} wait_cluster_consistency_async "${test_name}" "3" "3" new_primary_pod_from_label="$(get_primary_from_label)" if [ "${primary_pod_from_label}" == "${new_primary_pod_from_label}" ]; then echo "Old (${primary_pod_from_label}) and new (${new_primary_pod_from_label}) primary are the same (the failover didn't happen)!" exit 1 fi] logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | + source ../../functions logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ realpath ../../.. logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | ++++ mktemp -d logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ export TEMP_DIR=/tmp/tmp.ORNqGz0pv8 logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ TEMP_DIR=/tmp/tmp.ORNqGz0pv8 logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ GIT_BRANCH=PR-339 logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ VERSION=PR-339-3a8483b logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | ++++ which gdate logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | ++++ which date logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ date=/usr/bin/date logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ command -v oc logger.go:42: 12:12:34 | demand-backup/4-move-primary-before-backup | +++ oc get projects logger.go:42: 12:12:40 | demand-backup/4-move-primary-before-backup | error: the server doesn't have a resource type "projects" logger.go:42: 12:12:40 | demand-backup/4-move-primary-before-backup | +++ grep '^minikube' logger.go:42: 12:12:40 | demand-backup/4-move-primary-before-backup | +++ kubectl get nodes logger.go:42: 12:12:40 | demand-backup/4-move-primary-before-backup | ++++ pwd logger.go:42: 12:12:40 | demand-backup/4-move-primary-before-backup | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/demand-backup logger.go:42: 12:12:40 | demand-backup/4-move-primary-before-backup | ++ test_name=demand-backup logger.go:42: 12:12:40 | demand-backup/4-move-primary-before-backup | ++ get_primary_from_label logger.go:42: 12:12:40 | demand-backup/4-move-primary-before-backup | ++ kubectl -n kuttl-test-cosmic-ghoul get pods -l mysql.percona.com/primary=true '-ojsonpath={.items[0].metadata.name}' logger.go:42: 12:12:41 | demand-backup/4-move-primary-before-backup | + primary_pod_from_label=demand-backup-mysql-0 logger.go:42: 12:12:41 | demand-backup/4-move-primary-before-backup | + kubectl delete pod -n kuttl-test-cosmic-ghoul demand-backup-mysql-0 logger.go:42: 12:12:42 | demand-backup/4-move-primary-before-backup | pod "demand-backup-mysql-0" deleted logger.go:42: 12:13:03 | demand-backup/4-move-primary-before-backup | + wait_cluster_consistency_async demand-backup 3 3 logger.go:42: 12:13:03 | demand-backup/4-move-primary-before-backup | + local cluster_name=demand-backup logger.go:42: 12:13:03 | demand-backup/4-move-primary-before-backup | + local cluster_size=3 logger.go:42: 12:13:03 | demand-backup/4-move-primary-before-backup | + local orc_size=3 logger.go:42: 12:13:03 | demand-backup/4-move-primary-before-backup | + '[' -z 3 ']' logger.go:42: 12:13:03 | demand-backup/4-move-primary-before-backup | + sleep 7 logger.go:42: 12:13:10 | demand-backup/4-move-primary-before-backup | ++ kubectl get ps demand-backup -n kuttl-test-cosmic-ghoul -o 'jsonpath={.status.mysql.state}' logger.go:42: 12:13:10 | demand-backup/4-move-primary-before-backup | + [[ initializing == \r\e\a\d\y ]] logger.go:42: 12:13:10 | demand-backup/4-move-primary-before-backup | + echo 'waiting for cluster readyness (async)' logger.go:42: 12:13:10 | demand-backup/4-move-primary-before-backup | waiting for cluster readyness (async) logger.go:42: 12:13:10 | demand-backup/4-move-primary-before-backup | + sleep 15 logger.go:42: 12:13:25 | demand-backup/4-move-primary-before-backup | ++ kubectl get ps demand-backup -n kuttl-test-cosmic-ghoul -o 'jsonpath={.status.mysql.state}' logger.go:42: 12:13:26 | demand-backup/4-move-primary-before-backup | + [[ initializing == \r\e\a\d\y ]] logger.go:42: 12:13:26 | demand-backup/4-move-primary-before-backup | + echo 'waiting for cluster readyness (async)' logger.go:42: 12:13:26 | demand-backup/4-move-primary-before-backup | waiting for cluster readyness (async) logger.go:42: 12:13:26 | demand-backup/4-move-primary-before-backup | + sleep 15 logger.go:42: 12:13:41 | demand-backup/4-move-primary-before-backup | ++ kubectl get ps demand-backup -n kuttl-test-cosmic-ghoul -o 'jsonpath={.status.mysql.state}' logger.go:42: 12:13:42 | demand-backup/4-move-primary-before-backup | + [[ initializing == \r\e\a\d\y ]] logger.go:42: 12:13:42 | demand-backup/4-move-primary-before-backup | + echo 'waiting for cluster readyness (async)' logger.go:42: 12:13:42 | demand-backup/4-move-primary-before-backup | waiting for cluster readyness (async) logger.go:42: 12:13:42 | demand-backup/4-move-primary-before-backup | + sleep 15 logger.go:42: 12:13:57 | demand-backup/4-move-primary-before-backup | ++ kubectl get ps demand-backup -n kuttl-test-cosmic-ghoul -o 'jsonpath={.status.mysql.state}' logger.go:42: 12:13:58 | demand-backup/4-move-primary-before-backup | + [[ initializing == \r\e\a\d\y ]] logger.go:42: 12:13:58 | demand-backup/4-move-primary-before-backup | + echo 'waiting for cluster readyness (async)' logger.go:42: 12:13:58 | demand-backup/4-move-primary-before-backup | waiting for cluster readyness (async) logger.go:42: 12:13:58 | demand-backup/4-move-primary-before-backup | + sleep 15 logger.go:42: 12:14:13 | demand-backup/4-move-primary-before-backup | ++ kubectl get ps demand-backup -n kuttl-test-cosmic-ghoul -o 'jsonpath={.status.mysql.state}' logger.go:42: 12:14:14 | demand-backup/4-move-primary-before-backup | + [[ initializing == \r\e\a\d\y ]] logger.go:42: 12:14:14 | demand-backup/4-move-primary-before-backup | + echo 'waiting for cluster readyness (async)' logger.go:42: 12:14:14 | demand-backup/4-move-primary-before-backup | waiting for cluster readyness (async) logger.go:42: 12:14:14 | demand-backup/4-move-primary-before-backup | + sleep 15 logger.go:42: 12:14:29 | demand-backup/4-move-primary-before-backup | ++ kubectl get ps demand-backup -n kuttl-test-cosmic-ghoul -o 'jsonpath={.status.mysql.state}' logger.go:42: 12:14:30 | demand-backup/4-move-primary-before-backup | + [[ initializing == \r\e\a\d\y ]] logger.go:42: 12:14:30 | demand-backup/4-move-primary-before-backup | + echo 'waiting for cluster readyness (async)' logger.go:42: 12:14:30 | demand-backup/4-move-primary-before-backup | waiting for cluster readyness (async) logger.go:42: 12:14:30 | demand-backup/4-move-primary-before-backup | + sleep 15 logger.go:42: 12:14:45 | demand-backup/4-move-primary-before-backup | ++ kubectl get ps demand-backup -n kuttl-test-cosmic-ghoul -o 'jsonpath={.status.mysql.state}' logger.go:42: 12:14:46 | demand-backup/4-move-primary-before-backup | + [[ ready == \r\e\a\d\y ]] logger.go:42: 12:14:46 | demand-backup/4-move-primary-before-backup | ++ kubectl get ps demand-backup -n kuttl-test-cosmic-ghoul -o 'jsonpath={.status.mysql.ready}' logger.go:42: 12:14:46 | demand-backup/4-move-primary-before-backup | + [[ 3 == \3 ]] logger.go:42: 12:14:46 | demand-backup/4-move-primary-before-backup | ++ kubectl get ps demand-backup -n kuttl-test-cosmic-ghoul -o 'jsonpath={.status.orchestrator.ready}' logger.go:42: 12:14:47 | demand-backup/4-move-primary-before-backup | + [[ 3 == \3 ]] logger.go:42: 12:14:47 | demand-backup/4-move-primary-before-backup | ++ kubectl get ps demand-backup -n kuttl-test-cosmic-ghoul -o 'jsonpath={.status.orchestrator.state}' logger.go:42: 12:14:48 | demand-backup/4-move-primary-before-backup | + [[ ready == \r\e\a\d\y ]] logger.go:42: 12:14:48 | demand-backup/4-move-primary-before-backup | ++ kubectl get ps demand-backup -n kuttl-test-cosmic-ghoul -o 'jsonpath={.status.state}' logger.go:42: 12:14:49 | demand-backup/4-move-primary-before-backup | + [[ ready == \r\e\a\d\y ]] logger.go:42: 12:14:49 | demand-backup/4-move-primary-before-backup | ++ get_primary_from_label logger.go:42: 12:14:49 | demand-backup/4-move-primary-before-backup | ++ kubectl -n kuttl-test-cosmic-ghoul get pods -l mysql.percona.com/primary=true '-ojsonpath={.items[0].metadata.name}' logger.go:42: 12:14:49 | demand-backup/4-move-primary-before-backup | + new_primary_pod_from_label=demand-backup-mysql-2 logger.go:42: 12:14:49 | demand-backup/4-move-primary-before-backup | + '[' demand-backup-mysql-0 == demand-backup-mysql-2 ']' logger.go:42: 12:14:50 | demand-backup/4-move-primary-before-backup | test step completed 4-move-primary-before-backup logger.go:42: 12:14:50 | demand-backup/5-create-backup-minio | starting test step 5-create-backup-minio logger.go:42: 12:14:52 | demand-backup/5-create-backup-minio | PerconaServerMySQLBackup:kuttl-test-cosmic-ghoul/demand-backup-minio created logger.go:42: 12:15:09 | demand-backup/5-create-backup-minio | test step completed 5-create-backup-minio logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | starting test step 6-check-password-leak logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | running command: [sh -c set -o errexit set -o xtrace source ../../functions check_passwords_leak] logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | + source ../../functions logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ realpath ../../.. logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | ++++ mktemp -d logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ export TEMP_DIR=/tmp/tmp.rMhSfE1imL logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ TEMP_DIR=/tmp/tmp.rMhSfE1imL logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ GIT_BRANCH=PR-339 logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ VERSION=PR-339-3a8483b logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | ++++ which gdate logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | ++++ which date logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ date=/usr/bin/date logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ command -v oc logger.go:42: 12:15:09 | demand-backup/6-check-password-leak | +++ oc get projects logger.go:42: 12:15:14 | demand-backup/6-check-password-leak | error: the server doesn't have a resource type "projects" logger.go:42: 12:15:14 | demand-backup/6-check-password-leak | +++ kubectl get nodes logger.go:42: 12:15:14 | demand-backup/6-check-password-leak | +++ grep '^minikube' logger.go:42: 12:15:15 | demand-backup/6-check-password-leak | ++++ pwd logger.go:42: 12:15:15 | demand-backup/6-check-password-leak | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/demand-backup logger.go:42: 12:15:15 | demand-backup/6-check-password-leak | ++ test_name=demand-backup logger.go:42: 12:15:15 | demand-backup/6-check-password-leak | + check_passwords_leak logger.go:42: 12:15:15 | demand-backup/6-check-password-leak | ++ kubectl get secrets -o json logger.go:42: 12:15:15 | demand-backup/6-check-password-leak | ++ jq -r '.items[].data | to_entries | .[] | select(.key | (endswith(".crt") or endswith(".key") or endswith(".pub") or endswith(".pem") or endswith(".p12")) | not) | .value' logger.go:42: 12:15:16 | demand-backup/6-check-password-leak | + secrets= logger.go:42: 12:15:16 | demand-backup/6-check-password-leak | + passwords=' ' logger.go:42: 12:15:16 | demand-backup/6-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pods -o name logger.go:42: 12:15:16 | demand-backup/6-check-password-leak | ++ awk -F / '{print $2}' logger.go:42: 12:15:17 | demand-backup/6-check-password-leak | + pods='demand-backup-haproxy-0 logger.go:42: 12:15:17 | demand-backup/6-check-password-leak | demand-backup-haproxy-1 logger.go:42: 12:15:17 | demand-backup/6-check-password-leak | demand-backup-haproxy-2 logger.go:42: 12:15:17 | demand-backup/6-check-password-leak | demand-backup-mysql-0 logger.go:42: 12:15:17 | demand-backup/6-check-password-leak | demand-backup-mysql-1 logger.go:42: 12:15:17 | demand-backup/6-check-password-leak | demand-backup-mysql-2 logger.go:42: 12:15:17 | demand-backup/6-check-password-leak | demand-backup-orc-0 logger.go:42: 12:15:17 | demand-backup/6-check-password-leak | demand-backup-orc-1 logger.go:42: 12:15:17 | demand-backup/6-check-password-leak | demand-backup-orc-2 logger.go:42: 12:15:17 | demand-backup/6-check-password-leak | minio-service-6f47578c5b-2vr2g logger.go:42: 12:15:17 | demand-backup/6-check-password-leak | mysql-client logger.go:42: 12:15:17 | demand-backup/6-check-password-leak | percona-server-mysql-operator-6dcd78d857-h42q5 logger.go:42: 12:15:17 | demand-backup/6-check-password-leak | xb-demand-backup-minio-minio-xcl82' logger.go:42: 12:15:17 | demand-backup/6-check-password-leak | + collect_logs kuttl-test-cosmic-ghoul logger.go:42: 12:15:17 | demand-backup/6-check-password-leak | + NS=kuttl-test-cosmic-ghoul logger.go:42: 12:15:17 | demand-backup/6-check-password-leak | + for p in '$pods' logger.go:42: 12:15:17 | demand-backup/6-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod demand-backup-haproxy-0 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:15:18 | demand-backup/6-check-password-leak | + containers='haproxy mysql-monit' logger.go:42: 12:15:18 | demand-backup/6-check-password-leak | + for c in '$containers' logger.go:42: 12:15:18 | demand-backup/6-check-password-leak | + [[ haproxy =~ pmm ]] logger.go:42: 12:15:18 | demand-backup/6-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-haproxy-0 -c haproxy logger.go:42: 12:15:19 | demand-backup/6-check-password-leak | + echo logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-haproxy-0-haproxy.txt logger.go:42: 12:15:19 | demand-backup/6-check-password-leak | logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-haproxy-0-haproxy.txt logger.go:42: 12:15:19 | demand-backup/6-check-password-leak | + for c in '$containers' logger.go:42: 12:15:19 | demand-backup/6-check-password-leak | + [[ mysql-monit =~ pmm ]] logger.go:42: 12:15:19 | demand-backup/6-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-haproxy-0 -c mysql-monit logger.go:42: 12:15:20 | demand-backup/6-check-password-leak | + echo logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-haproxy-0-mysql-monit.txt logger.go:42: 12:15:20 | demand-backup/6-check-password-leak | logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-haproxy-0-mysql-monit.txt logger.go:42: 12:15:20 | demand-backup/6-check-password-leak | + echo logger.go:42: 12:15:20 | demand-backup/6-check-password-leak | logger.go:42: 12:15:20 | demand-backup/6-check-password-leak | + for p in '$pods' logger.go:42: 12:15:20 | demand-backup/6-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod demand-backup-haproxy-1 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:15:21 | demand-backup/6-check-password-leak | + containers='haproxy mysql-monit' logger.go:42: 12:15:21 | demand-backup/6-check-password-leak | + for c in '$containers' logger.go:42: 12:15:21 | demand-backup/6-check-password-leak | + [[ haproxy =~ pmm ]] logger.go:42: 12:15:21 | demand-backup/6-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-haproxy-1 -c haproxy logger.go:42: 12:15:22 | demand-backup/6-check-password-leak | + echo logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-haproxy-1-haproxy.txt logger.go:42: 12:15:22 | demand-backup/6-check-password-leak | logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-haproxy-1-haproxy.txt logger.go:42: 12:15:22 | demand-backup/6-check-password-leak | + for c in '$containers' logger.go:42: 12:15:22 | demand-backup/6-check-password-leak | + [[ mysql-monit =~ pmm ]] logger.go:42: 12:15:22 | demand-backup/6-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-haproxy-1 -c mysql-monit logger.go:42: 12:15:23 | demand-backup/6-check-password-leak | + echo logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-haproxy-1-mysql-monit.txt logger.go:42: 12:15:23 | demand-backup/6-check-password-leak | logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-haproxy-1-mysql-monit.txt logger.go:42: 12:15:23 | demand-backup/6-check-password-leak | + echo logger.go:42: 12:15:23 | demand-backup/6-check-password-leak | logger.go:42: 12:15:23 | demand-backup/6-check-password-leak | + for p in '$pods' logger.go:42: 12:15:23 | demand-backup/6-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod demand-backup-haproxy-2 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:15:24 | demand-backup/6-check-password-leak | + containers='haproxy mysql-monit' logger.go:42: 12:15:24 | demand-backup/6-check-password-leak | + for c in '$containers' logger.go:42: 12:15:24 | demand-backup/6-check-password-leak | + [[ haproxy =~ pmm ]] logger.go:42: 12:15:24 | demand-backup/6-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-haproxy-2 -c haproxy logger.go:42: 12:15:25 | demand-backup/6-check-password-leak | + echo logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-haproxy-2-haproxy.txt logger.go:42: 12:15:25 | demand-backup/6-check-password-leak | logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-haproxy-2-haproxy.txt logger.go:42: 12:15:25 | demand-backup/6-check-password-leak | + for c in '$containers' logger.go:42: 12:15:25 | demand-backup/6-check-password-leak | + [[ mysql-monit =~ pmm ]] logger.go:42: 12:15:25 | demand-backup/6-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-haproxy-2 -c mysql-monit logger.go:42: 12:15:26 | demand-backup/6-check-password-leak | + echo logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-haproxy-2-mysql-monit.txt logger.go:42: 12:15:26 | demand-backup/6-check-password-leak | logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-haproxy-2-mysql-monit.txt logger.go:42: 12:15:26 | demand-backup/6-check-password-leak | + echo logger.go:42: 12:15:26 | demand-backup/6-check-password-leak | logger.go:42: 12:15:26 | demand-backup/6-check-password-leak | + for p in '$pods' logger.go:42: 12:15:26 | demand-backup/6-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod demand-backup-mysql-0 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:15:27 | demand-backup/6-check-password-leak | + containers='mysql xtrabackup pt-heartbeat' logger.go:42: 12:15:27 | demand-backup/6-check-password-leak | + for c in '$containers' logger.go:42: 12:15:27 | demand-backup/6-check-password-leak | + [[ mysql =~ pmm ]] logger.go:42: 12:15:27 | demand-backup/6-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-mysql-0 -c mysql logger.go:42: 12:15:28 | demand-backup/6-check-password-leak | + echo logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-mysql-0-mysql.txt logger.go:42: 12:15:28 | demand-backup/6-check-password-leak | logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-mysql-0-mysql.txt logger.go:42: 12:15:28 | demand-backup/6-check-password-leak | + for c in '$containers' logger.go:42: 12:15:28 | demand-backup/6-check-password-leak | + [[ xtrabackup =~ pmm ]] logger.go:42: 12:15:28 | demand-backup/6-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-mysql-0 -c xtrabackup logger.go:42: 12:15:29 | demand-backup/6-check-password-leak | + echo logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-mysql-0-xtrabackup.txt logger.go:42: 12:15:29 | demand-backup/6-check-password-leak | logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-mysql-0-xtrabackup.txt logger.go:42: 12:15:29 | demand-backup/6-check-password-leak | + for c in '$containers' logger.go:42: 12:15:29 | demand-backup/6-check-password-leak | + [[ pt-heartbeat =~ pmm ]] logger.go:42: 12:15:29 | demand-backup/6-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-mysql-0 -c pt-heartbeat logger.go:42: 12:15:30 | demand-backup/6-check-password-leak | + echo logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-mysql-0-pt-heartbeat.txt logger.go:42: 12:15:30 | demand-backup/6-check-password-leak | logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-mysql-0-pt-heartbeat.txt logger.go:42: 12:15:30 | demand-backup/6-check-password-leak | + echo logger.go:42: 12:15:30 | demand-backup/6-check-password-leak | logger.go:42: 12:15:30 | demand-backup/6-check-password-leak | + for p in '$pods' logger.go:42: 12:15:30 | demand-backup/6-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod demand-backup-mysql-1 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:15:31 | demand-backup/6-check-password-leak | + containers='mysql xtrabackup pt-heartbeat' logger.go:42: 12:15:31 | demand-backup/6-check-password-leak | + for c in '$containers' logger.go:42: 12:15:31 | demand-backup/6-check-password-leak | + [[ mysql =~ pmm ]] logger.go:42: 12:15:31 | demand-backup/6-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-mysql-1 -c mysql logger.go:42: 12:15:32 | demand-backup/6-check-password-leak | + echo logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-mysql-1-mysql.txt logger.go:42: 12:15:32 | demand-backup/6-check-password-leak | logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-mysql-1-mysql.txt logger.go:42: 12:15:32 | demand-backup/6-check-password-leak | + for c in '$containers' logger.go:42: 12:15:32 | demand-backup/6-check-password-leak | + [[ xtrabackup =~ pmm ]] logger.go:42: 12:15:32 | demand-backup/6-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-mysql-1 -c xtrabackup logger.go:42: 12:15:33 | demand-backup/6-check-password-leak | + echo logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-mysql-1-xtrabackup.txt logger.go:42: 12:15:33 | demand-backup/6-check-password-leak | logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-mysql-1-xtrabackup.txt logger.go:42: 12:15:33 | demand-backup/6-check-password-leak | + for c in '$containers' logger.go:42: 12:15:33 | demand-backup/6-check-password-leak | + [[ pt-heartbeat =~ pmm ]] logger.go:42: 12:15:33 | demand-backup/6-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-mysql-1 -c pt-heartbeat logger.go:42: 12:15:34 | demand-backup/6-check-password-leak | + echo logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-mysql-1-pt-heartbeat.txt logger.go:42: 12:15:34 | demand-backup/6-check-password-leak | logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-mysql-1-pt-heartbeat.txt logger.go:42: 12:15:34 | demand-backup/6-check-password-leak | + echo logger.go:42: 12:15:34 | demand-backup/6-check-password-leak | logger.go:42: 12:15:34 | demand-backup/6-check-password-leak | + for p in '$pods' logger.go:42: 12:15:34 | demand-backup/6-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod demand-backup-mysql-2 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:15:35 | demand-backup/6-check-password-leak | + containers='mysql xtrabackup pt-heartbeat' logger.go:42: 12:15:35 | demand-backup/6-check-password-leak | + for c in '$containers' logger.go:42: 12:15:35 | demand-backup/6-check-password-leak | + [[ mysql =~ pmm ]] logger.go:42: 12:15:35 | demand-backup/6-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-mysql-2 -c mysql logger.go:42: 12:15:36 | demand-backup/6-check-password-leak | + echo logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-mysql-2-mysql.txt logger.go:42: 12:15:36 | demand-backup/6-check-password-leak | logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-mysql-2-mysql.txt logger.go:42: 12:15:36 | demand-backup/6-check-password-leak | + for c in '$containers' logger.go:42: 12:15:36 | demand-backup/6-check-password-leak | + [[ xtrabackup =~ pmm ]] logger.go:42: 12:15:36 | demand-backup/6-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-mysql-2 -c xtrabackup logger.go:42: 12:15:37 | demand-backup/6-check-password-leak | + echo logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-mysql-2-xtrabackup.txt logger.go:42: 12:15:37 | demand-backup/6-check-password-leak | logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-mysql-2-xtrabackup.txt logger.go:42: 12:15:37 | demand-backup/6-check-password-leak | + for c in '$containers' logger.go:42: 12:15:37 | demand-backup/6-check-password-leak | + [[ pt-heartbeat =~ pmm ]] logger.go:42: 12:15:37 | demand-backup/6-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-mysql-2 -c pt-heartbeat logger.go:42: 12:15:38 | demand-backup/6-check-password-leak | + echo logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-mysql-2-pt-heartbeat.txt logger.go:42: 12:15:38 | demand-backup/6-check-password-leak | logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-mysql-2-pt-heartbeat.txt logger.go:42: 12:15:38 | demand-backup/6-check-password-leak | + echo logger.go:42: 12:15:38 | demand-backup/6-check-password-leak | logger.go:42: 12:15:38 | demand-backup/6-check-password-leak | + for p in '$pods' logger.go:42: 12:15:38 | demand-backup/6-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod demand-backup-orc-0 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:15:38 | demand-backup/6-check-password-leak | + containers='orc mysql-monit' logger.go:42: 12:15:38 | demand-backup/6-check-password-leak | + for c in '$containers' logger.go:42: 12:15:38 | demand-backup/6-check-password-leak | + [[ orc =~ pmm ]] logger.go:42: 12:15:38 | demand-backup/6-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-orc-0 -c orc logger.go:42: 12:15:40 | demand-backup/6-check-password-leak | + echo logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-orc-0-orc.txt logger.go:42: 12:15:40 | demand-backup/6-check-password-leak | logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-orc-0-orc.txt logger.go:42: 12:15:40 | demand-backup/6-check-password-leak | + for c in '$containers' logger.go:42: 12:15:40 | demand-backup/6-check-password-leak | + [[ mysql-monit =~ pmm ]] logger.go:42: 12:15:40 | demand-backup/6-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-orc-0 -c mysql-monit logger.go:42: 12:15:41 | demand-backup/6-check-password-leak | + echo logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-orc-0-mysql-monit.txt logger.go:42: 12:15:41 | demand-backup/6-check-password-leak | logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-orc-0-mysql-monit.txt logger.go:42: 12:15:41 | demand-backup/6-check-password-leak | + echo logger.go:42: 12:15:41 | demand-backup/6-check-password-leak | logger.go:42: 12:15:41 | demand-backup/6-check-password-leak | + for p in '$pods' logger.go:42: 12:15:41 | demand-backup/6-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod demand-backup-orc-1 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:15:41 | demand-backup/6-check-password-leak | + containers='orc mysql-monit' logger.go:42: 12:15:41 | demand-backup/6-check-password-leak | + for c in '$containers' logger.go:42: 12:15:41 | demand-backup/6-check-password-leak | + [[ orc =~ pmm ]] logger.go:42: 12:15:41 | demand-backup/6-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-orc-1 -c orc logger.go:42: 12:15:43 | demand-backup/6-check-password-leak | + echo logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-orc-1-orc.txt logger.go:42: 12:15:43 | demand-backup/6-check-password-leak | logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-orc-1-orc.txt logger.go:42: 12:15:43 | demand-backup/6-check-password-leak | + for c in '$containers' logger.go:42: 12:15:43 | demand-backup/6-check-password-leak | + [[ mysql-monit =~ pmm ]] logger.go:42: 12:15:43 | demand-backup/6-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-orc-1 -c mysql-monit logger.go:42: 12:15:44 | demand-backup/6-check-password-leak | + echo logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-orc-1-mysql-monit.txt logger.go:42: 12:15:44 | demand-backup/6-check-password-leak | logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-orc-1-mysql-monit.txt logger.go:42: 12:15:44 | demand-backup/6-check-password-leak | + echo logger.go:42: 12:15:44 | demand-backup/6-check-password-leak | logger.go:42: 12:15:44 | demand-backup/6-check-password-leak | + for p in '$pods' logger.go:42: 12:15:44 | demand-backup/6-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod demand-backup-orc-2 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:15:45 | demand-backup/6-check-password-leak | + containers='orc mysql-monit' logger.go:42: 12:15:45 | demand-backup/6-check-password-leak | + for c in '$containers' logger.go:42: 12:15:45 | demand-backup/6-check-password-leak | + [[ orc =~ pmm ]] logger.go:42: 12:15:45 | demand-backup/6-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-orc-2 -c orc logger.go:42: 12:15:46 | demand-backup/6-check-password-leak | + echo logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-orc-2-orc.txt logger.go:42: 12:15:46 | demand-backup/6-check-password-leak | logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-orc-2-orc.txt logger.go:42: 12:15:46 | demand-backup/6-check-password-leak | + for c in '$containers' logger.go:42: 12:15:46 | demand-backup/6-check-password-leak | + [[ mysql-monit =~ pmm ]] logger.go:42: 12:15:46 | demand-backup/6-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-orc-2 -c mysql-monit logger.go:42: 12:15:47 | demand-backup/6-check-password-leak | + echo logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-orc-2-mysql-monit.txt logger.go:42: 12:15:47 | demand-backup/6-check-password-leak | logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-demand-backup-orc-2-mysql-monit.txt logger.go:42: 12:15:47 | demand-backup/6-check-password-leak | + echo logger.go:42: 12:15:47 | demand-backup/6-check-password-leak | logger.go:42: 12:15:47 | demand-backup/6-check-password-leak | + for p in '$pods' logger.go:42: 12:15:47 | demand-backup/6-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod minio-service-6f47578c5b-2vr2g -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:15:47 | demand-backup/6-check-password-leak | + containers=minio logger.go:42: 12:15:47 | demand-backup/6-check-password-leak | + for c in '$containers' logger.go:42: 12:15:47 | demand-backup/6-check-password-leak | + [[ minio =~ pmm ]] logger.go:42: 12:15:47 | demand-backup/6-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs minio-service-6f47578c5b-2vr2g -c minio logger.go:42: 12:15:48 | demand-backup/6-check-password-leak | + echo logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-minio-service-6f47578c5b-2vr2g-minio.txt logger.go:42: 12:15:48 | demand-backup/6-check-password-leak | logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-minio-service-6f47578c5b-2vr2g-minio.txt logger.go:42: 12:15:48 | demand-backup/6-check-password-leak | + echo logger.go:42: 12:15:48 | demand-backup/6-check-password-leak | logger.go:42: 12:15:48 | demand-backup/6-check-password-leak | + for p in '$pods' logger.go:42: 12:15:48 | demand-backup/6-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod mysql-client -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:15:49 | demand-backup/6-check-password-leak | + containers=mysql-client logger.go:42: 12:15:49 | demand-backup/6-check-password-leak | + for c in '$containers' logger.go:42: 12:15:49 | demand-backup/6-check-password-leak | + [[ mysql-client =~ pmm ]] logger.go:42: 12:15:49 | demand-backup/6-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs mysql-client -c mysql-client logger.go:42: 12:15:50 | demand-backup/6-check-password-leak | + echo logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-mysql-client-mysql-client.txt logger.go:42: 12:15:50 | demand-backup/6-check-password-leak | logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-mysql-client-mysql-client.txt logger.go:42: 12:15:50 | demand-backup/6-check-password-leak | + echo logger.go:42: 12:15:50 | demand-backup/6-check-password-leak | logger.go:42: 12:15:50 | demand-backup/6-check-password-leak | + for p in '$pods' logger.go:42: 12:15:50 | demand-backup/6-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod percona-server-mysql-operator-6dcd78d857-h42q5 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:15:51 | demand-backup/6-check-password-leak | + containers=manager logger.go:42: 12:15:51 | demand-backup/6-check-password-leak | + for c in '$containers' logger.go:42: 12:15:51 | demand-backup/6-check-password-leak | + [[ manager =~ pmm ]] logger.go:42: 12:15:51 | demand-backup/6-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs percona-server-mysql-operator-6dcd78d857-h42q5 -c manager logger.go:42: 12:15:52 | demand-backup/6-check-password-leak | + echo logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-percona-server-mysql-operator-6dcd78d857-h42q5-manager.txt logger.go:42: 12:15:52 | demand-backup/6-check-password-leak | logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-percona-server-mysql-operator-6dcd78d857-h42q5-manager.txt logger.go:42: 12:15:52 | demand-backup/6-check-password-leak | + echo logger.go:42: 12:15:52 | demand-backup/6-check-password-leak | logger.go:42: 12:15:52 | demand-backup/6-check-password-leak | + for p in '$pods' logger.go:42: 12:15:52 | demand-backup/6-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod xb-demand-backup-minio-minio-xcl82 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:15:52 | demand-backup/6-check-password-leak | + containers=xtrabackup logger.go:42: 12:15:52 | demand-backup/6-check-password-leak | + for c in '$containers' logger.go:42: 12:15:52 | demand-backup/6-check-password-leak | + [[ xtrabackup =~ pmm ]] logger.go:42: 12:15:52 | demand-backup/6-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs xb-demand-backup-minio-minio-xcl82 -c xtrabackup logger.go:42: 12:15:53 | demand-backup/6-check-password-leak | + echo logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-xb-demand-backup-minio-minio-xcl82-xtrabackup.txt logger.go:42: 12:15:53 | demand-backup/6-check-password-leak | logs saved in: /tmp/tmp.rMhSfE1imL/logs_output-xb-demand-backup-minio-minio-xcl82-xtrabackup.txt logger.go:42: 12:15:53 | demand-backup/6-check-password-leak | + echo logger.go:42: 12:15:53 | demand-backup/6-check-password-leak | logger.go:42: 12:15:53 | demand-backup/6-check-password-leak | + '[' -n ps-operator ']' logger.go:42: 12:15:53 | demand-backup/6-check-password-leak | ++ kubectl -n ps-operator get pods -o name logger.go:42: 12:15:53 | demand-backup/6-check-password-leak | ++ awk -F / '{print $2}' logger.go:42: 12:15:54 | demand-backup/6-check-password-leak | + pods= logger.go:42: 12:15:54 | demand-backup/6-check-password-leak | + collect_logs ps-operator logger.go:42: 12:15:54 | demand-backup/6-check-password-leak | + NS=ps-operator logger.go:42: 12:15:55 | demand-backup/6-check-password-leak | test step completed 6-check-password-leak logger.go:42: 12:15:55 | demand-backup/7-delete-data | starting test step 7-delete-data logger.go:42: 12:15:55 | demand-backup/7-delete-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "TRUNCATE TABLE myDB.myTable" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password" cluster_name=$(get_cluster_name) for i in 0 1 2; do data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 04-delete-data-minio-${i} --from-literal=data="${data}" done] logger.go:42: 12:15:55 | demand-backup/7-delete-data | + source ../../functions logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ realpath ../../.. logger.go:42: 12:15:55 | demand-backup/7-delete-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:15:55 | demand-backup/7-delete-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:15:55 | demand-backup/7-delete-data | ++++ mktemp -d logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ export TEMP_DIR=/tmp/tmp.eksTYI0Xii logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ TEMP_DIR=/tmp/tmp.eksTYI0Xii logger.go:42: 12:15:55 | demand-backup/7-delete-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ GIT_BRANCH=PR-339 logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ VERSION=PR-339-3a8483b logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:15:55 | demand-backup/7-delete-data | ++++ which gdate logger.go:42: 12:15:55 | demand-backup/7-delete-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:15:55 | demand-backup/7-delete-data | ++++ which date logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ date=/usr/bin/date logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ command -v oc logger.go:42: 12:15:55 | demand-backup/7-delete-data | +++ oc get projects logger.go:42: 12:16:01 | demand-backup/7-delete-data | error: the server doesn't have a resource type "projects" logger.go:42: 12:16:01 | demand-backup/7-delete-data | +++ kubectl get nodes logger.go:42: 12:16:01 | demand-backup/7-delete-data | +++ grep '^minikube' logger.go:42: 12:16:02 | demand-backup/7-delete-data | ++++ pwd logger.go:42: 12:16:02 | demand-backup/7-delete-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/demand-backup logger.go:42: 12:16:02 | demand-backup/7-delete-data | ++ test_name=demand-backup logger.go:42: 12:16:02 | demand-backup/7-delete-data | +++ get_cluster_name logger.go:42: 12:16:02 | demand-backup/7-delete-data | +++ kubectl -n kuttl-test-cosmic-ghoul get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 12:16:02 | demand-backup/7-delete-data | ++ get_haproxy_svc demand-backup logger.go:42: 12:16:02 | demand-backup/7-delete-data | ++ local cluster=demand-backup logger.go:42: 12:16:02 | demand-backup/7-delete-data | ++ echo demand-backup-haproxy logger.go:42: 12:16:02 | demand-backup/7-delete-data | + run_mysql 'TRUNCATE TABLE myDB.myTable' '-h demand-backup-haproxy -uroot -proot_password' logger.go:42: 12:16:02 | demand-backup/7-delete-data | + local 'command=TRUNCATE TABLE myDB.myTable' logger.go:42: 12:16:02 | demand-backup/7-delete-data | + local 'uri=-h demand-backup-haproxy -uroot -proot_password' logger.go:42: 12:16:02 | demand-backup/7-delete-data | + local pod= logger.go:42: 12:16:02 | demand-backup/7-delete-data | ++ get_client_pod logger.go:42: 12:16:02 | demand-backup/7-delete-data | ++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:16:03 | demand-backup/7-delete-data | + client_pod=mysql-client logger.go:42: 12:16:03 | demand-backup/7-delete-data | + wait_pod mysql-client logger.go:42: 12:16:03 | demand-backup/7-delete-data | + local pod=mysql-client logger.go:42: 12:16:03 | demand-backup/7-delete-data | + set +o xtrace logger.go:42: 12:16:04 | demand-backup/7-delete-data | mysql-clienttrue logger.go:42: 12:16:04 | demand-backup/7-delete-data | + kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "TRUNCATE TABLE myDB.myTable" | mysql -sN -h demand-backup-haproxy -uroot -proot_password' logger.go:42: 12:16:04 | demand-backup/7-delete-data | + sed -e 's/mysql: //' logger.go:42: 12:16:04 | demand-backup/7-delete-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:16:05 | demand-backup/7-delete-data | + : logger.go:42: 12:16:05 | demand-backup/7-delete-data | ++ get_cluster_name logger.go:42: 12:16:05 | demand-backup/7-delete-data | ++ kubectl -n kuttl-test-cosmic-ghoul get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 12:16:06 | demand-backup/7-delete-data | + cluster_name=demand-backup logger.go:42: 12:16:06 | demand-backup/7-delete-data | + for i in 0 1 2 logger.go:42: 12:16:06 | demand-backup/7-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:16:06 | demand-backup/7-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:16:06 | demand-backup/7-delete-data | ++ local 'uri=-h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:16:06 | demand-backup/7-delete-data | ++ local pod= logger.go:42: 12:16:06 | demand-backup/7-delete-data | +++ get_client_pod logger.go:42: 12:16:06 | demand-backup/7-delete-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:16:07 | demand-backup/7-delete-data | ++ client_pod=mysql-client logger.go:42: 12:16:07 | demand-backup/7-delete-data | ++ wait_pod mysql-client logger.go:42: 12:16:07 | demand-backup/7-delete-data | ++ local pod=mysql-client logger.go:42: 12:16:07 | demand-backup/7-delete-data | ++ set +o xtrace logger.go:42: 12:16:08 | demand-backup/7-delete-data | mysql-clienttrue logger.go:42: 12:16:08 | demand-backup/7-delete-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:16:08 | demand-backup/7-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 12:16:08 | demand-backup/7-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:16:09 | demand-backup/7-delete-data | ++ : logger.go:42: 12:16:09 | demand-backup/7-delete-data | + data= logger.go:42: 12:16:09 | demand-backup/7-delete-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 04-delete-data-minio-0 --from-literal=data= logger.go:42: 12:16:10 | demand-backup/7-delete-data | configmap/04-delete-data-minio-0 created logger.go:42: 12:16:10 | demand-backup/7-delete-data | + for i in 0 1 2 logger.go:42: 12:16:10 | demand-backup/7-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:16:10 | demand-backup/7-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:16:10 | demand-backup/7-delete-data | ++ local 'uri=-h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:16:10 | demand-backup/7-delete-data | ++ local pod= logger.go:42: 12:16:10 | demand-backup/7-delete-data | +++ get_client_pod logger.go:42: 12:16:10 | demand-backup/7-delete-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:16:11 | demand-backup/7-delete-data | ++ client_pod=mysql-client logger.go:42: 12:16:11 | demand-backup/7-delete-data | ++ wait_pod mysql-client logger.go:42: 12:16:11 | demand-backup/7-delete-data | ++ local pod=mysql-client logger.go:42: 12:16:11 | demand-backup/7-delete-data | ++ set +o xtrace logger.go:42: 12:16:11 | demand-backup/7-delete-data | mysql-clienttrue logger.go:42: 12:16:11 | demand-backup/7-delete-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:16:11 | demand-backup/7-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 12:16:11 | demand-backup/7-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:16:13 | demand-backup/7-delete-data | ++ : logger.go:42: 12:16:13 | demand-backup/7-delete-data | + data= logger.go:42: 12:16:13 | demand-backup/7-delete-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 04-delete-data-minio-1 --from-literal=data= logger.go:42: 12:16:14 | demand-backup/7-delete-data | configmap/04-delete-data-minio-1 created logger.go:42: 12:16:14 | demand-backup/7-delete-data | + for i in 0 1 2 logger.go:42: 12:16:14 | demand-backup/7-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:16:14 | demand-backup/7-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:16:14 | demand-backup/7-delete-data | ++ local 'uri=-h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:16:14 | demand-backup/7-delete-data | ++ local pod= logger.go:42: 12:16:14 | demand-backup/7-delete-data | +++ get_client_pod logger.go:42: 12:16:14 | demand-backup/7-delete-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:16:14 | demand-backup/7-delete-data | ++ client_pod=mysql-client logger.go:42: 12:16:14 | demand-backup/7-delete-data | ++ wait_pod mysql-client logger.go:42: 12:16:14 | demand-backup/7-delete-data | ++ local pod=mysql-client logger.go:42: 12:16:14 | demand-backup/7-delete-data | ++ set +o xtrace logger.go:42: 12:16:15 | demand-backup/7-delete-data | mysql-clienttrue logger.go:42: 12:16:15 | demand-backup/7-delete-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:16:15 | demand-backup/7-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 12:16:15 | demand-backup/7-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:16:17 | demand-backup/7-delete-data | ++ : logger.go:42: 12:16:17 | demand-backup/7-delete-data | + data= logger.go:42: 12:16:17 | demand-backup/7-delete-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 04-delete-data-minio-2 --from-literal=data= logger.go:42: 12:16:17 | demand-backup/7-delete-data | configmap/04-delete-data-minio-2 created logger.go:42: 12:16:19 | demand-backup/7-delete-data | test step completed 7-delete-data logger.go:42: 12:16:19 | demand-backup/8-restore-from-minio | starting test step 8-restore-from-minio logger.go:42: 12:16:20 | demand-backup/8-restore-from-minio | PerconaServerMySQLRestore:kuttl-test-cosmic-ghoul/demand-backup-restore-minio created logger.go:42: 12:21:23 | demand-backup/8-restore-from-minio | test step completed 8-restore-from-minio logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | starting test step 9-check-password-leak logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | running command: [sh -c set -o errexit set -o xtrace source ../../functions check_passwords_leak] logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | + source ../../functions logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ realpath ../../.. logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | ++++ mktemp -d logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ export TEMP_DIR=/tmp/tmp.hAItUbdZLD logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ TEMP_DIR=/tmp/tmp.hAItUbdZLD logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ GIT_BRANCH=PR-339 logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ VERSION=PR-339-3a8483b logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | ++++ which gdate logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | ++++ which date logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ date=/usr/bin/date logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ command -v oc logger.go:42: 12:21:23 | demand-backup/9-check-password-leak | +++ oc get projects logger.go:42: 12:21:29 | demand-backup/9-check-password-leak | error: the server doesn't have a resource type "projects" logger.go:42: 12:21:29 | demand-backup/9-check-password-leak | +++ grep '^minikube' logger.go:42: 12:21:29 | demand-backup/9-check-password-leak | +++ kubectl get nodes logger.go:42: 12:21:29 | demand-backup/9-check-password-leak | ++++ pwd logger.go:42: 12:21:29 | demand-backup/9-check-password-leak | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/demand-backup logger.go:42: 12:21:29 | demand-backup/9-check-password-leak | ++ test_name=demand-backup logger.go:42: 12:21:29 | demand-backup/9-check-password-leak | + check_passwords_leak logger.go:42: 12:21:29 | demand-backup/9-check-password-leak | ++ kubectl get secrets -o json logger.go:42: 12:21:29 | demand-backup/9-check-password-leak | ++ jq -r '.items[].data | to_entries | .[] | select(.key | (endswith(".crt") or endswith(".key") or endswith(".pub") or endswith(".pem") or endswith(".p12")) | not) | .value' logger.go:42: 12:21:30 | demand-backup/9-check-password-leak | + secrets= logger.go:42: 12:21:30 | demand-backup/9-check-password-leak | + passwords=' ' logger.go:42: 12:21:30 | demand-backup/9-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pods -o name logger.go:42: 12:21:30 | demand-backup/9-check-password-leak | ++ awk -F / '{print $2}' logger.go:42: 12:21:31 | demand-backup/9-check-password-leak | + pods='demand-backup-haproxy-0 logger.go:42: 12:21:31 | demand-backup/9-check-password-leak | demand-backup-haproxy-1 logger.go:42: 12:21:31 | demand-backup/9-check-password-leak | demand-backup-haproxy-2 logger.go:42: 12:21:31 | demand-backup/9-check-password-leak | demand-backup-mysql-0 logger.go:42: 12:21:31 | demand-backup/9-check-password-leak | demand-backup-mysql-1 logger.go:42: 12:21:31 | demand-backup/9-check-password-leak | demand-backup-mysql-2 logger.go:42: 12:21:31 | demand-backup/9-check-password-leak | demand-backup-orc-0 logger.go:42: 12:21:31 | demand-backup/9-check-password-leak | demand-backup-orc-1 logger.go:42: 12:21:31 | demand-backup/9-check-password-leak | demand-backup-orc-2 logger.go:42: 12:21:31 | demand-backup/9-check-password-leak | minio-service-6f47578c5b-2vr2g logger.go:42: 12:21:31 | demand-backup/9-check-password-leak | mysql-client logger.go:42: 12:21:31 | demand-backup/9-check-password-leak | percona-server-mysql-operator-6dcd78d857-h42q5 logger.go:42: 12:21:31 | demand-backup/9-check-password-leak | xb-demand-backup-minio-minio-xcl82 logger.go:42: 12:21:31 | demand-backup/9-check-password-leak | xb-restore-demand-backup-restore-minio-mhmx8' logger.go:42: 12:21:31 | demand-backup/9-check-password-leak | + collect_logs kuttl-test-cosmic-ghoul logger.go:42: 12:21:31 | demand-backup/9-check-password-leak | + NS=kuttl-test-cosmic-ghoul logger.go:42: 12:21:31 | demand-backup/9-check-password-leak | + for p in '$pods' logger.go:42: 12:21:31 | demand-backup/9-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod demand-backup-haproxy-0 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:21:32 | demand-backup/9-check-password-leak | + containers='haproxy mysql-monit' logger.go:42: 12:21:32 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:21:32 | demand-backup/9-check-password-leak | + [[ haproxy =~ pmm ]] logger.go:42: 12:21:32 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-haproxy-0 -c haproxy logger.go:42: 12:21:33 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-haproxy-0-haproxy.txt logger.go:42: 12:21:33 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-haproxy-0-haproxy.txt logger.go:42: 12:21:33 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:21:33 | demand-backup/9-check-password-leak | + [[ mysql-monit =~ pmm ]] logger.go:42: 12:21:33 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-haproxy-0 -c mysql-monit logger.go:42: 12:21:34 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-haproxy-0-mysql-monit.txt logger.go:42: 12:21:34 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-haproxy-0-mysql-monit.txt logger.go:42: 12:21:34 | demand-backup/9-check-password-leak | + echo logger.go:42: 12:21:34 | demand-backup/9-check-password-leak | logger.go:42: 12:21:34 | demand-backup/9-check-password-leak | + for p in '$pods' logger.go:42: 12:21:34 | demand-backup/9-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod demand-backup-haproxy-1 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:21:35 | demand-backup/9-check-password-leak | + containers='haproxy mysql-monit' logger.go:42: 12:21:35 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:21:35 | demand-backup/9-check-password-leak | + [[ haproxy =~ pmm ]] logger.go:42: 12:21:35 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-haproxy-1 -c haproxy logger.go:42: 12:21:36 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-haproxy-1-haproxy.txt logger.go:42: 12:21:36 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-haproxy-1-haproxy.txt logger.go:42: 12:21:36 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:21:36 | demand-backup/9-check-password-leak | + [[ mysql-monit =~ pmm ]] logger.go:42: 12:21:36 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-haproxy-1 -c mysql-monit logger.go:42: 12:21:37 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-haproxy-1-mysql-monit.txt logger.go:42: 12:21:37 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-haproxy-1-mysql-monit.txt logger.go:42: 12:21:37 | demand-backup/9-check-password-leak | + echo logger.go:42: 12:21:37 | demand-backup/9-check-password-leak | logger.go:42: 12:21:37 | demand-backup/9-check-password-leak | + for p in '$pods' logger.go:42: 12:21:37 | demand-backup/9-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod demand-backup-haproxy-2 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:21:38 | demand-backup/9-check-password-leak | + containers='haproxy mysql-monit' logger.go:42: 12:21:38 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:21:38 | demand-backup/9-check-password-leak | + [[ haproxy =~ pmm ]] logger.go:42: 12:21:38 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-haproxy-2 -c haproxy logger.go:42: 12:21:39 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-haproxy-2-haproxy.txt logger.go:42: 12:21:39 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-haproxy-2-haproxy.txt logger.go:42: 12:21:39 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:21:39 | demand-backup/9-check-password-leak | + [[ mysql-monit =~ pmm ]] logger.go:42: 12:21:39 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-haproxy-2 -c mysql-monit logger.go:42: 12:21:40 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-haproxy-2-mysql-monit.txt logger.go:42: 12:21:40 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-haproxy-2-mysql-monit.txt logger.go:42: 12:21:40 | demand-backup/9-check-password-leak | + echo logger.go:42: 12:21:40 | demand-backup/9-check-password-leak | logger.go:42: 12:21:40 | demand-backup/9-check-password-leak | + for p in '$pods' logger.go:42: 12:21:40 | demand-backup/9-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod demand-backup-mysql-0 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:21:41 | demand-backup/9-check-password-leak | + containers='mysql xtrabackup pt-heartbeat' logger.go:42: 12:21:41 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:21:41 | demand-backup/9-check-password-leak | + [[ mysql =~ pmm ]] logger.go:42: 12:21:41 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-mysql-0 -c mysql logger.go:42: 12:21:42 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-mysql-0-mysql.txt logger.go:42: 12:21:42 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-mysql-0-mysql.txt logger.go:42: 12:21:42 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:21:42 | demand-backup/9-check-password-leak | + [[ xtrabackup =~ pmm ]] logger.go:42: 12:21:42 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-mysql-0 -c xtrabackup logger.go:42: 12:21:42 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-mysql-0-xtrabackup.txt logger.go:42: 12:21:42 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-mysql-0-xtrabackup.txt logger.go:42: 12:21:42 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:21:42 | demand-backup/9-check-password-leak | + [[ pt-heartbeat =~ pmm ]] logger.go:42: 12:21:42 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-mysql-0 -c pt-heartbeat logger.go:42: 12:21:43 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-mysql-0-pt-heartbeat.txt logger.go:42: 12:21:43 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-mysql-0-pt-heartbeat.txt logger.go:42: 12:21:43 | demand-backup/9-check-password-leak | + echo logger.go:42: 12:21:43 | demand-backup/9-check-password-leak | logger.go:42: 12:21:43 | demand-backup/9-check-password-leak | + for p in '$pods' logger.go:42: 12:21:43 | demand-backup/9-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod demand-backup-mysql-1 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:21:44 | demand-backup/9-check-password-leak | + containers='mysql xtrabackup pt-heartbeat' logger.go:42: 12:21:44 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:21:44 | demand-backup/9-check-password-leak | + [[ mysql =~ pmm ]] logger.go:42: 12:21:44 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-mysql-1 -c mysql logger.go:42: 12:21:45 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-mysql-1-mysql.txt logger.go:42: 12:21:45 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-mysql-1-mysql.txt logger.go:42: 12:21:45 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:21:45 | demand-backup/9-check-password-leak | + [[ xtrabackup =~ pmm ]] logger.go:42: 12:21:45 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-mysql-1 -c xtrabackup logger.go:42: 12:21:46 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-mysql-1-xtrabackup.txt logger.go:42: 12:21:46 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-mysql-1-xtrabackup.txt logger.go:42: 12:21:46 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:21:46 | demand-backup/9-check-password-leak | + [[ pt-heartbeat =~ pmm ]] logger.go:42: 12:21:46 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-mysql-1 -c pt-heartbeat logger.go:42: 12:21:47 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-mysql-1-pt-heartbeat.txt logger.go:42: 12:21:47 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-mysql-1-pt-heartbeat.txt logger.go:42: 12:21:47 | demand-backup/9-check-password-leak | + echo logger.go:42: 12:21:47 | demand-backup/9-check-password-leak | logger.go:42: 12:21:47 | demand-backup/9-check-password-leak | + for p in '$pods' logger.go:42: 12:21:47 | demand-backup/9-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod demand-backup-mysql-2 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:21:48 | demand-backup/9-check-password-leak | + containers='mysql xtrabackup pt-heartbeat' logger.go:42: 12:21:48 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:21:48 | demand-backup/9-check-password-leak | + [[ mysql =~ pmm ]] logger.go:42: 12:21:48 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-mysql-2 -c mysql logger.go:42: 12:21:49 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-mysql-2-mysql.txt logger.go:42: 12:21:49 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-mysql-2-mysql.txt logger.go:42: 12:21:49 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:21:49 | demand-backup/9-check-password-leak | + [[ xtrabackup =~ pmm ]] logger.go:42: 12:21:49 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-mysql-2 -c xtrabackup logger.go:42: 12:21:50 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-mysql-2-xtrabackup.txt logger.go:42: 12:21:50 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-mysql-2-xtrabackup.txt logger.go:42: 12:21:50 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:21:50 | demand-backup/9-check-password-leak | + [[ pt-heartbeat =~ pmm ]] logger.go:42: 12:21:50 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-mysql-2 -c pt-heartbeat logger.go:42: 12:21:51 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-mysql-2-pt-heartbeat.txt logger.go:42: 12:21:51 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-mysql-2-pt-heartbeat.txt logger.go:42: 12:21:51 | demand-backup/9-check-password-leak | + echo logger.go:42: 12:21:51 | demand-backup/9-check-password-leak | logger.go:42: 12:21:51 | demand-backup/9-check-password-leak | + for p in '$pods' logger.go:42: 12:21:51 | demand-backup/9-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod demand-backup-orc-0 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:21:52 | demand-backup/9-check-password-leak | + containers='orc mysql-monit' logger.go:42: 12:21:52 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:21:52 | demand-backup/9-check-password-leak | + [[ orc =~ pmm ]] logger.go:42: 12:21:52 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-orc-0 -c orc logger.go:42: 12:21:53 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-orc-0-orc.txt logger.go:42: 12:21:53 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-orc-0-orc.txt logger.go:42: 12:21:53 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:21:53 | demand-backup/9-check-password-leak | + [[ mysql-monit =~ pmm ]] logger.go:42: 12:21:53 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-orc-0 -c mysql-monit logger.go:42: 12:21:54 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-orc-0-mysql-monit.txt logger.go:42: 12:21:54 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-orc-0-mysql-monit.txt logger.go:42: 12:21:54 | demand-backup/9-check-password-leak | + echo logger.go:42: 12:21:54 | demand-backup/9-check-password-leak | logger.go:42: 12:21:54 | demand-backup/9-check-password-leak | + for p in '$pods' logger.go:42: 12:21:54 | demand-backup/9-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod demand-backup-orc-1 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:21:55 | demand-backup/9-check-password-leak | + containers='orc mysql-monit' logger.go:42: 12:21:55 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:21:55 | demand-backup/9-check-password-leak | + [[ orc =~ pmm ]] logger.go:42: 12:21:55 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-orc-1 -c orc logger.go:42: 12:21:56 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-orc-1-orc.txt logger.go:42: 12:21:56 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-orc-1-orc.txt logger.go:42: 12:21:56 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:21:56 | demand-backup/9-check-password-leak | + [[ mysql-monit =~ pmm ]] logger.go:42: 12:21:56 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-orc-1 -c mysql-monit logger.go:42: 12:21:57 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-orc-1-mysql-monit.txt logger.go:42: 12:21:57 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-orc-1-mysql-monit.txt logger.go:42: 12:21:57 | demand-backup/9-check-password-leak | + echo logger.go:42: 12:21:57 | demand-backup/9-check-password-leak | logger.go:42: 12:21:57 | demand-backup/9-check-password-leak | + for p in '$pods' logger.go:42: 12:21:57 | demand-backup/9-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod demand-backup-orc-2 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:21:58 | demand-backup/9-check-password-leak | + containers='orc mysql-monit' logger.go:42: 12:21:58 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:21:58 | demand-backup/9-check-password-leak | + [[ orc =~ pmm ]] logger.go:42: 12:21:58 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-orc-2 -c orc logger.go:42: 12:21:59 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-orc-2-orc.txt logger.go:42: 12:21:59 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-orc-2-orc.txt logger.go:42: 12:21:59 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:21:59 | demand-backup/9-check-password-leak | + [[ mysql-monit =~ pmm ]] logger.go:42: 12:21:59 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs demand-backup-orc-2 -c mysql-monit logger.go:42: 12:22:00 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-orc-2-mysql-monit.txt logger.go:42: 12:22:00 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-demand-backup-orc-2-mysql-monit.txt logger.go:42: 12:22:00 | demand-backup/9-check-password-leak | + echo logger.go:42: 12:22:00 | demand-backup/9-check-password-leak | logger.go:42: 12:22:00 | demand-backup/9-check-password-leak | + for p in '$pods' logger.go:42: 12:22:00 | demand-backup/9-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod minio-service-6f47578c5b-2vr2g -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:22:00 | demand-backup/9-check-password-leak | + containers=minio logger.go:42: 12:22:00 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:22:00 | demand-backup/9-check-password-leak | + [[ minio =~ pmm ]] logger.go:42: 12:22:00 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs minio-service-6f47578c5b-2vr2g -c minio logger.go:42: 12:22:01 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-minio-service-6f47578c5b-2vr2g-minio.txt logger.go:42: 12:22:01 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-minio-service-6f47578c5b-2vr2g-minio.txt logger.go:42: 12:22:01 | demand-backup/9-check-password-leak | + echo logger.go:42: 12:22:01 | demand-backup/9-check-password-leak | logger.go:42: 12:22:01 | demand-backup/9-check-password-leak | + for p in '$pods' logger.go:42: 12:22:01 | demand-backup/9-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod mysql-client -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:22:02 | demand-backup/9-check-password-leak | + containers=mysql-client logger.go:42: 12:22:02 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:22:02 | demand-backup/9-check-password-leak | + [[ mysql-client =~ pmm ]] logger.go:42: 12:22:02 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs mysql-client -c mysql-client logger.go:42: 12:22:03 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-mysql-client-mysql-client.txt logger.go:42: 12:22:03 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-mysql-client-mysql-client.txt logger.go:42: 12:22:03 | demand-backup/9-check-password-leak | + echo logger.go:42: 12:22:03 | demand-backup/9-check-password-leak | logger.go:42: 12:22:03 | demand-backup/9-check-password-leak | + for p in '$pods' logger.go:42: 12:22:03 | demand-backup/9-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod percona-server-mysql-operator-6dcd78d857-h42q5 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:22:04 | demand-backup/9-check-password-leak | + containers=manager logger.go:42: 12:22:04 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:22:04 | demand-backup/9-check-password-leak | + [[ manager =~ pmm ]] logger.go:42: 12:22:04 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs percona-server-mysql-operator-6dcd78d857-h42q5 -c manager logger.go:42: 12:22:05 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-percona-server-mysql-operator-6dcd78d857-h42q5-manager.txt logger.go:42: 12:22:05 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-percona-server-mysql-operator-6dcd78d857-h42q5-manager.txt logger.go:42: 12:22:05 | demand-backup/9-check-password-leak | + echo logger.go:42: 12:22:05 | demand-backup/9-check-password-leak | logger.go:42: 12:22:05 | demand-backup/9-check-password-leak | + for p in '$pods' logger.go:42: 12:22:05 | demand-backup/9-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod xb-demand-backup-minio-minio-xcl82 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:22:05 | demand-backup/9-check-password-leak | + containers=xtrabackup logger.go:42: 12:22:05 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:22:05 | demand-backup/9-check-password-leak | + [[ xtrabackup =~ pmm ]] logger.go:42: 12:22:05 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs xb-demand-backup-minio-minio-xcl82 -c xtrabackup logger.go:42: 12:22:06 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-xb-demand-backup-minio-minio-xcl82-xtrabackup.txt logger.go:42: 12:22:06 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-xb-demand-backup-minio-minio-xcl82-xtrabackup.txt logger.go:42: 12:22:06 | demand-backup/9-check-password-leak | + echo logger.go:42: 12:22:06 | demand-backup/9-check-password-leak | logger.go:42: 12:22:06 | demand-backup/9-check-password-leak | + for p in '$pods' logger.go:42: 12:22:06 | demand-backup/9-check-password-leak | ++ kubectl -n kuttl-test-cosmic-ghoul get pod xb-restore-demand-backup-restore-minio-mhmx8 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 12:22:07 | demand-backup/9-check-password-leak | + containers=xtrabackup logger.go:42: 12:22:07 | demand-backup/9-check-password-leak | + for c in '$containers' logger.go:42: 12:22:07 | demand-backup/9-check-password-leak | + [[ xtrabackup =~ pmm ]] logger.go:42: 12:22:07 | demand-backup/9-check-password-leak | + kubectl -n kuttl-test-cosmic-ghoul logs xb-restore-demand-backup-restore-minio-mhmx8 -c xtrabackup logger.go:42: 12:22:08 | demand-backup/9-check-password-leak | + echo logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-xb-restore-demand-backup-restore-minio-mhmx8-xtrabackup.txt logger.go:42: 12:22:08 | demand-backup/9-check-password-leak | logs saved in: /tmp/tmp.hAItUbdZLD/logs_output-xb-restore-demand-backup-restore-minio-mhmx8-xtrabackup.txt logger.go:42: 12:22:08 | demand-backup/9-check-password-leak | + echo logger.go:42: 12:22:08 | demand-backup/9-check-password-leak | logger.go:42: 12:22:08 | demand-backup/9-check-password-leak | + '[' -n ps-operator ']' logger.go:42: 12:22:08 | demand-backup/9-check-password-leak | ++ kubectl -n ps-operator get pods -o name logger.go:42: 12:22:08 | demand-backup/9-check-password-leak | ++ awk -F / '{print $2}' logger.go:42: 12:22:09 | demand-backup/9-check-password-leak | + pods= logger.go:42: 12:22:09 | demand-backup/9-check-password-leak | + collect_logs ps-operator logger.go:42: 12:22:09 | demand-backup/9-check-password-leak | + NS=ps-operator logger.go:42: 12:22:10 | demand-backup/9-check-password-leak | test step completed 9-check-password-leak logger.go:42: 12:22:10 | demand-backup/10-read-data | starting test step 10-read-data logger.go:42: 12:22:10 | demand-backup/10-read-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions cluster_name=$(get_cluster_name) for i in 0 1 2; do data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 06-read-data-minio-${i} --from-literal=data="${data}" done] logger.go:42: 12:22:10 | demand-backup/10-read-data | + source ../../functions logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ realpath ../../.. logger.go:42: 12:22:10 | demand-backup/10-read-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:22:10 | demand-backup/10-read-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:22:10 | demand-backup/10-read-data | ++++ mktemp -d logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ export TEMP_DIR=/tmp/tmp.wG0xPNfYqA logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ TEMP_DIR=/tmp/tmp.wG0xPNfYqA logger.go:42: 12:22:10 | demand-backup/10-read-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ GIT_BRANCH=PR-339 logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ VERSION=PR-339-3a8483b logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:22:10 | demand-backup/10-read-data | ++++ which gdate logger.go:42: 12:22:10 | demand-backup/10-read-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:22:10 | demand-backup/10-read-data | ++++ which date logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ date=/usr/bin/date logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ command -v oc logger.go:42: 12:22:10 | demand-backup/10-read-data | +++ oc get projects logger.go:42: 12:22:15 | demand-backup/10-read-data | error: the server doesn't have a resource type "projects" logger.go:42: 12:22:15 | demand-backup/10-read-data | +++ kubectl get nodes logger.go:42: 12:22:15 | demand-backup/10-read-data | +++ grep '^minikube' logger.go:42: 12:22:16 | demand-backup/10-read-data | ++++ pwd logger.go:42: 12:22:16 | demand-backup/10-read-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/demand-backup logger.go:42: 12:22:16 | demand-backup/10-read-data | ++ test_name=demand-backup logger.go:42: 12:22:16 | demand-backup/10-read-data | ++ get_cluster_name logger.go:42: 12:22:16 | demand-backup/10-read-data | ++ kubectl -n kuttl-test-cosmic-ghoul get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 12:22:17 | demand-backup/10-read-data | + cluster_name=demand-backup logger.go:42: 12:22:17 | demand-backup/10-read-data | + for i in 0 1 2 logger.go:42: 12:22:17 | demand-backup/10-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:22:17 | demand-backup/10-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:22:17 | demand-backup/10-read-data | ++ local 'uri=-h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:22:17 | demand-backup/10-read-data | ++ local pod= logger.go:42: 12:22:17 | demand-backup/10-read-data | +++ get_client_pod logger.go:42: 12:22:17 | demand-backup/10-read-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:22:18 | demand-backup/10-read-data | ++ client_pod=mysql-client logger.go:42: 12:22:18 | demand-backup/10-read-data | ++ wait_pod mysql-client logger.go:42: 12:22:18 | demand-backup/10-read-data | ++ local pod=mysql-client logger.go:42: 12:22:18 | demand-backup/10-read-data | ++ set +o xtrace logger.go:42: 12:22:19 | demand-backup/10-read-data | mysql-clienttrue logger.go:42: 12:22:19 | demand-backup/10-read-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:22:19 | demand-backup/10-read-data | ++ sed -e 's/mysql: //' logger.go:42: 12:22:19 | demand-backup/10-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:22:20 | demand-backup/10-read-data | + data=100500 logger.go:42: 12:22:20 | demand-backup/10-read-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 06-read-data-minio-0 --from-literal=data=100500 logger.go:42: 12:22:21 | demand-backup/10-read-data | configmap/06-read-data-minio-0 created logger.go:42: 12:22:21 | demand-backup/10-read-data | + for i in 0 1 2 logger.go:42: 12:22:21 | demand-backup/10-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:22:21 | demand-backup/10-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:22:21 | demand-backup/10-read-data | ++ local 'uri=-h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:22:21 | demand-backup/10-read-data | ++ local pod= logger.go:42: 12:22:21 | demand-backup/10-read-data | +++ get_client_pod logger.go:42: 12:22:21 | demand-backup/10-read-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:22:22 | demand-backup/10-read-data | ++ client_pod=mysql-client logger.go:42: 12:22:22 | demand-backup/10-read-data | ++ wait_pod mysql-client logger.go:42: 12:22:22 | demand-backup/10-read-data | ++ local pod=mysql-client logger.go:42: 12:22:22 | demand-backup/10-read-data | ++ set +o xtrace logger.go:42: 12:22:22 | demand-backup/10-read-data | mysql-clienttrue logger.go:42: 12:22:22 | demand-backup/10-read-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:22:22 | demand-backup/10-read-data | ++ sed -e 's/mysql: //' logger.go:42: 12:22:22 | demand-backup/10-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:22:24 | demand-backup/10-read-data | + data=100500 logger.go:42: 12:22:24 | demand-backup/10-read-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 06-read-data-minio-1 --from-literal=data=100500 logger.go:42: 12:22:25 | demand-backup/10-read-data | configmap/06-read-data-minio-1 created logger.go:42: 12:22:25 | demand-backup/10-read-data | + for i in 0 1 2 logger.go:42: 12:22:25 | demand-backup/10-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:22:25 | demand-backup/10-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:22:25 | demand-backup/10-read-data | ++ local 'uri=-h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:22:25 | demand-backup/10-read-data | ++ local pod= logger.go:42: 12:22:25 | demand-backup/10-read-data | +++ get_client_pod logger.go:42: 12:22:25 | demand-backup/10-read-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:22:25 | demand-backup/10-read-data | ++ client_pod=mysql-client logger.go:42: 12:22:25 | demand-backup/10-read-data | ++ wait_pod mysql-client logger.go:42: 12:22:25 | demand-backup/10-read-data | ++ local pod=mysql-client logger.go:42: 12:22:25 | demand-backup/10-read-data | ++ set +o xtrace logger.go:42: 12:22:26 | demand-backup/10-read-data | mysql-clienttrue logger.go:42: 12:22:26 | demand-backup/10-read-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:22:26 | demand-backup/10-read-data | ++ sed -e 's/mysql: //' logger.go:42: 12:22:26 | demand-backup/10-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:22:28 | demand-backup/10-read-data | + data=100500 logger.go:42: 12:22:28 | demand-backup/10-read-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 06-read-data-minio-2 --from-literal=data=100500 logger.go:42: 12:22:28 | demand-backup/10-read-data | configmap/06-read-data-minio-2 created logger.go:42: 12:22:30 | demand-backup/10-read-data | test step completed 10-read-data logger.go:42: 12:22:30 | demand-backup/11-delete-data | starting test step 11-delete-data logger.go:42: 12:22:30 | demand-backup/11-delete-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "TRUNCATE TABLE myDB.myTable" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password" cluster_name=$(get_cluster_name) for i in 0 1 2 do data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 04-delete-data-minio-backup-source-${i} --from-literal=data="${data}" done] logger.go:42: 12:22:30 | demand-backup/11-delete-data | + source ../../functions logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ realpath ../../.. logger.go:42: 12:22:30 | demand-backup/11-delete-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:22:30 | demand-backup/11-delete-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:22:30 | demand-backup/11-delete-data | ++++ mktemp -d logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ export TEMP_DIR=/tmp/tmp.8VXZxqnSLg logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ TEMP_DIR=/tmp/tmp.8VXZxqnSLg logger.go:42: 12:22:30 | demand-backup/11-delete-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ GIT_BRANCH=PR-339 logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ VERSION=PR-339-3a8483b logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:22:30 | demand-backup/11-delete-data | ++++ which gdate logger.go:42: 12:22:30 | demand-backup/11-delete-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:22:30 | demand-backup/11-delete-data | ++++ which date logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ date=/usr/bin/date logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ command -v oc logger.go:42: 12:22:30 | demand-backup/11-delete-data | +++ oc get projects logger.go:42: 12:22:35 | demand-backup/11-delete-data | error: the server doesn't have a resource type "projects" logger.go:42: 12:22:35 | demand-backup/11-delete-data | +++ grep '^minikube' logger.go:42: 12:22:35 | demand-backup/11-delete-data | +++ kubectl get nodes logger.go:42: 12:22:36 | demand-backup/11-delete-data | ++++ pwd logger.go:42: 12:22:36 | demand-backup/11-delete-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/demand-backup logger.go:42: 12:22:36 | demand-backup/11-delete-data | ++ test_name=demand-backup logger.go:42: 12:22:36 | demand-backup/11-delete-data | +++ get_cluster_name logger.go:42: 12:22:36 | demand-backup/11-delete-data | +++ kubectl -n kuttl-test-cosmic-ghoul get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 12:22:37 | demand-backup/11-delete-data | ++ get_haproxy_svc demand-backup logger.go:42: 12:22:37 | demand-backup/11-delete-data | ++ local cluster=demand-backup logger.go:42: 12:22:37 | demand-backup/11-delete-data | ++ echo demand-backup-haproxy logger.go:42: 12:22:37 | demand-backup/11-delete-data | + run_mysql 'TRUNCATE TABLE myDB.myTable' '-h demand-backup-haproxy -uroot -proot_password' logger.go:42: 12:22:37 | demand-backup/11-delete-data | + local 'command=TRUNCATE TABLE myDB.myTable' logger.go:42: 12:22:37 | demand-backup/11-delete-data | + local 'uri=-h demand-backup-haproxy -uroot -proot_password' logger.go:42: 12:22:37 | demand-backup/11-delete-data | + local pod= logger.go:42: 12:22:37 | demand-backup/11-delete-data | ++ get_client_pod logger.go:42: 12:22:37 | demand-backup/11-delete-data | ++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:22:38 | demand-backup/11-delete-data | + client_pod=mysql-client logger.go:42: 12:22:38 | demand-backup/11-delete-data | + wait_pod mysql-client logger.go:42: 12:22:38 | demand-backup/11-delete-data | + local pod=mysql-client logger.go:42: 12:22:38 | demand-backup/11-delete-data | + set +o xtrace logger.go:42: 12:22:39 | demand-backup/11-delete-data | mysql-clienttrue logger.go:42: 12:22:39 | demand-backup/11-delete-data | + kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "TRUNCATE TABLE myDB.myTable" | mysql -sN -h demand-backup-haproxy -uroot -proot_password' logger.go:42: 12:22:39 | demand-backup/11-delete-data | + sed -e 's/mysql: //' logger.go:42: 12:22:39 | demand-backup/11-delete-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:22:40 | demand-backup/11-delete-data | + : logger.go:42: 12:22:40 | demand-backup/11-delete-data | ++ get_cluster_name logger.go:42: 12:22:40 | demand-backup/11-delete-data | ++ kubectl -n kuttl-test-cosmic-ghoul get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 12:22:41 | demand-backup/11-delete-data | + cluster_name=demand-backup logger.go:42: 12:22:41 | demand-backup/11-delete-data | + for i in 0 1 2 logger.go:42: 12:22:41 | demand-backup/11-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:22:41 | demand-backup/11-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:22:41 | demand-backup/11-delete-data | ++ local 'uri=-h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:22:41 | demand-backup/11-delete-data | ++ local pod= logger.go:42: 12:22:41 | demand-backup/11-delete-data | +++ get_client_pod logger.go:42: 12:22:41 | demand-backup/11-delete-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:22:42 | demand-backup/11-delete-data | ++ client_pod=mysql-client logger.go:42: 12:22:42 | demand-backup/11-delete-data | ++ wait_pod mysql-client logger.go:42: 12:22:42 | demand-backup/11-delete-data | ++ local pod=mysql-client logger.go:42: 12:22:42 | demand-backup/11-delete-data | ++ set +o xtrace logger.go:42: 12:22:43 | demand-backup/11-delete-data | mysql-clienttrue logger.go:42: 12:22:43 | demand-backup/11-delete-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:22:43 | demand-backup/11-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 12:22:43 | demand-backup/11-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:22:44 | demand-backup/11-delete-data | ++ : logger.go:42: 12:22:44 | demand-backup/11-delete-data | + data= logger.go:42: 12:22:44 | demand-backup/11-delete-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 04-delete-data-minio-backup-source-0 --from-literal=data= logger.go:42: 12:22:45 | demand-backup/11-delete-data | configmap/04-delete-data-minio-backup-source-0 created logger.go:42: 12:22:45 | demand-backup/11-delete-data | + for i in 0 1 2 logger.go:42: 12:22:45 | demand-backup/11-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:22:45 | demand-backup/11-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:22:45 | demand-backup/11-delete-data | ++ local 'uri=-h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:22:45 | demand-backup/11-delete-data | ++ local pod= logger.go:42: 12:22:45 | demand-backup/11-delete-data | +++ get_client_pod logger.go:42: 12:22:45 | demand-backup/11-delete-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:22:45 | demand-backup/11-delete-data | ++ client_pod=mysql-client logger.go:42: 12:22:45 | demand-backup/11-delete-data | ++ wait_pod mysql-client logger.go:42: 12:22:45 | demand-backup/11-delete-data | ++ local pod=mysql-client logger.go:42: 12:22:45 | demand-backup/11-delete-data | ++ set +o xtrace logger.go:42: 12:22:46 | demand-backup/11-delete-data | mysql-clienttrue logger.go:42: 12:22:46 | demand-backup/11-delete-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:22:46 | demand-backup/11-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 12:22:46 | demand-backup/11-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:22:48 | demand-backup/11-delete-data | ++ : logger.go:42: 12:22:48 | demand-backup/11-delete-data | + data= logger.go:42: 12:22:48 | demand-backup/11-delete-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 04-delete-data-minio-backup-source-1 --from-literal=data= logger.go:42: 12:22:48 | demand-backup/11-delete-data | configmap/04-delete-data-minio-backup-source-1 created logger.go:42: 12:22:48 | demand-backup/11-delete-data | + for i in 0 1 2 logger.go:42: 12:22:48 | demand-backup/11-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:22:48 | demand-backup/11-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:22:48 | demand-backup/11-delete-data | ++ local 'uri=-h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:22:48 | demand-backup/11-delete-data | ++ local pod= logger.go:42: 12:22:48 | demand-backup/11-delete-data | +++ get_client_pod logger.go:42: 12:22:48 | demand-backup/11-delete-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:22:49 | demand-backup/11-delete-data | ++ client_pod=mysql-client logger.go:42: 12:22:49 | demand-backup/11-delete-data | ++ wait_pod mysql-client logger.go:42: 12:22:49 | demand-backup/11-delete-data | ++ local pod=mysql-client logger.go:42: 12:22:49 | demand-backup/11-delete-data | ++ set +o xtrace logger.go:42: 12:22:50 | demand-backup/11-delete-data | mysql-clienttrue logger.go:42: 12:22:50 | demand-backup/11-delete-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:22:50 | demand-backup/11-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 12:22:50 | demand-backup/11-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:22:51 | demand-backup/11-delete-data | ++ : logger.go:42: 12:22:51 | demand-backup/11-delete-data | + data= logger.go:42: 12:22:51 | demand-backup/11-delete-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 04-delete-data-minio-backup-source-2 --from-literal=data= logger.go:42: 12:22:52 | demand-backup/11-delete-data | configmap/04-delete-data-minio-backup-source-2 created logger.go:42: 12:22:54 | demand-backup/11-delete-data | test step completed 11-delete-data logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | starting test step 12-restore-from-minio-backup-source logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | running command: [sh -c set -o errexit set -o xtrace source ../../functions storage_name="minio" backup_name="demand-backup-minio" restore_name="demand-backup-restore-minio-backup-source" cluster_name="${test_name}${name_suffix:+-$name_suffix}" destination=$(kubectl -n "${NAMESPACE}" get ps-backup "${backup_name}" -o jsonpath='{.status.destination}') cat "${DEPLOY_DIR}/restore.yaml" \ | yq eval "$(printf '.metadata.name="%s"' "${restore_name}")" - \ | yq eval "$(printf '.spec.clusterName="%s"' "${cluster_name}")" - \ | yq eval "del(.spec.backupName)" - \ | yq eval "$(printf '.spec.backupSource.destination="%s"' "${destination}")" - \ | yq eval '.spec.backupSource.storage.type="s3"' - \ | yq eval '.spec.backupSource.storage.s3.bucket="operator-testing"' - \ | yq eval '.spec.backupSource.storage.s3.credentialsSecret="minio-secret"' - \ | yq eval '.spec.backupSource.storage.s3.endpointUrl="http://minio-service:9000"' - \ | yq eval '.spec.backupSource.storage.s3.region="us-east-1"' - \ | kubectl apply -n "${NAMESPACE}" -f -] logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | + source ../../functions logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ realpath ../../.. logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | ++++ mktemp -d logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ export TEMP_DIR=/tmp/tmp.Cwt0Dlvhyg logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ TEMP_DIR=/tmp/tmp.Cwt0Dlvhyg logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ GIT_BRANCH=PR-339 logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ VERSION=PR-339-3a8483b logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | ++++ which gdate logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | ++++ which date logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ date=/usr/bin/date logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ command -v oc logger.go:42: 12:22:54 | demand-backup/12-restore-from-minio-backup-source | +++ oc get projects logger.go:42: 12:22:59 | demand-backup/12-restore-from-minio-backup-source | error: the server doesn't have a resource type "projects" logger.go:42: 12:22:59 | demand-backup/12-restore-from-minio-backup-source | +++ kubectl get nodes logger.go:42: 12:22:59 | demand-backup/12-restore-from-minio-backup-source | +++ grep '^minikube' logger.go:42: 12:23:00 | demand-backup/12-restore-from-minio-backup-source | ++++ pwd logger.go:42: 12:23:00 | demand-backup/12-restore-from-minio-backup-source | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/demand-backup logger.go:42: 12:23:00 | demand-backup/12-restore-from-minio-backup-source | ++ test_name=demand-backup logger.go:42: 12:23:00 | demand-backup/12-restore-from-minio-backup-source | + storage_name=minio logger.go:42: 12:23:00 | demand-backup/12-restore-from-minio-backup-source | + backup_name=demand-backup-minio logger.go:42: 12:23:00 | demand-backup/12-restore-from-minio-backup-source | + restore_name=demand-backup-restore-minio-backup-source logger.go:42: 12:23:00 | demand-backup/12-restore-from-minio-backup-source | + cluster_name=demand-backup logger.go:42: 12:23:00 | demand-backup/12-restore-from-minio-backup-source | ++ kubectl -n kuttl-test-cosmic-ghoul get ps-backup demand-backup-minio -o 'jsonpath={.status.destination}' logger.go:42: 12:23:01 | demand-backup/12-restore-from-minio-backup-source | + destination=s3://operator-testing/demand-backup-2023-08-09-12:14:52-full logger.go:42: 12:23:01 | demand-backup/12-restore-from-minio-backup-source | + yq eval 'del(.spec.backupName)' - logger.go:42: 12:23:01 | demand-backup/12-restore-from-minio-backup-source | ++ printf '.metadata.name="%s"' demand-backup-restore-minio-backup-source logger.go:42: 12:23:01 | demand-backup/12-restore-from-minio-backup-source | + yq eval '.metadata.name="demand-backup-restore-minio-backup-source"' - logger.go:42: 12:23:01 | demand-backup/12-restore-from-minio-backup-source | + cat /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/restore.yaml logger.go:42: 12:23:01 | demand-backup/12-restore-from-minio-backup-source | + yq eval '.spec.backupSource.storage.s3.bucket="operator-testing"' - logger.go:42: 12:23:01 | demand-backup/12-restore-from-minio-backup-source | + kubectl apply -n kuttl-test-cosmic-ghoul -f - logger.go:42: 12:23:01 | demand-backup/12-restore-from-minio-backup-source | ++ printf '.spec.clusterName="%s"' demand-backup logger.go:42: 12:23:01 | demand-backup/12-restore-from-minio-backup-source | + yq eval '.spec.backupSource.storage.type="s3"' - logger.go:42: 12:23:01 | demand-backup/12-restore-from-minio-backup-source | ++ printf '.spec.backupSource.destination="%s"' s3://operator-testing/demand-backup-2023-08-09-12:14:52-full logger.go:42: 12:23:01 | demand-backup/12-restore-from-minio-backup-source | + yq eval '.spec.backupSource.storage.s3.credentialsSecret="minio-secret"' - logger.go:42: 12:23:01 | demand-backup/12-restore-from-minio-backup-source | + yq eval '.spec.backupSource.storage.s3.region="us-east-1"' - logger.go:42: 12:23:01 | demand-backup/12-restore-from-minio-backup-source | + yq eval '.spec.backupSource.destination="s3://operator-testing/demand-backup-2023-08-09-12:14:52-full"' - logger.go:42: 12:23:01 | demand-backup/12-restore-from-minio-backup-source | + yq eval '.spec.clusterName="demand-backup"' - logger.go:42: 12:23:01 | demand-backup/12-restore-from-minio-backup-source | + yq eval '.spec.backupSource.storage.s3.endpointUrl="http://minio-service:9000"' - logger.go:42: 12:23:02 | demand-backup/12-restore-from-minio-backup-source | perconaservermysqlrestore.ps.percona.com/demand-backup-restore-minio-backup-source created logger.go:42: 12:27:58 | demand-backup/12-restore-from-minio-backup-source | test step completed 12-restore-from-minio-backup-source logger.go:42: 12:27:58 | demand-backup/13-read-data | starting test step 13-read-data logger.go:42: 12:27:58 | demand-backup/13-read-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions cluster_name=$(get_cluster_name) for i in 0 1 2 do data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 09-read-data-minio-backup-source-${i} --from-literal=data="${data}" done] logger.go:42: 12:27:58 | demand-backup/13-read-data | + source ../../functions logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ realpath ../../.. logger.go:42: 12:27:58 | demand-backup/13-read-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:27:58 | demand-backup/13-read-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:27:58 | demand-backup/13-read-data | ++++ mktemp -d logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ export TEMP_DIR=/tmp/tmp.8xB27iBljr logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ TEMP_DIR=/tmp/tmp.8xB27iBljr logger.go:42: 12:27:58 | demand-backup/13-read-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ GIT_BRANCH=PR-339 logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ VERSION=PR-339-3a8483b logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:27:58 | demand-backup/13-read-data | ++++ which gdate logger.go:42: 12:27:58 | demand-backup/13-read-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:27:58 | demand-backup/13-read-data | ++++ which date logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ date=/usr/bin/date logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ command -v oc logger.go:42: 12:27:58 | demand-backup/13-read-data | +++ oc get projects logger.go:42: 12:28:04 | demand-backup/13-read-data | error: the server doesn't have a resource type "projects" logger.go:42: 12:28:04 | demand-backup/13-read-data | +++ kubectl get nodes logger.go:42: 12:28:04 | demand-backup/13-read-data | +++ grep '^minikube' logger.go:42: 12:28:05 | demand-backup/13-read-data | ++++ pwd logger.go:42: 12:28:05 | demand-backup/13-read-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/demand-backup logger.go:42: 12:28:05 | demand-backup/13-read-data | ++ test_name=demand-backup logger.go:42: 12:28:05 | demand-backup/13-read-data | ++ get_cluster_name logger.go:42: 12:28:05 | demand-backup/13-read-data | ++ kubectl -n kuttl-test-cosmic-ghoul get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 12:28:06 | demand-backup/13-read-data | + cluster_name=demand-backup logger.go:42: 12:28:06 | demand-backup/13-read-data | + for i in 0 1 2 logger.go:42: 12:28:06 | demand-backup/13-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:28:06 | demand-backup/13-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:28:06 | demand-backup/13-read-data | ++ local 'uri=-h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:28:06 | demand-backup/13-read-data | ++ local pod= logger.go:42: 12:28:06 | demand-backup/13-read-data | +++ get_client_pod logger.go:42: 12:28:06 | demand-backup/13-read-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:28:06 | demand-backup/13-read-data | ++ client_pod=mysql-client logger.go:42: 12:28:06 | demand-backup/13-read-data | ++ wait_pod mysql-client logger.go:42: 12:28:06 | demand-backup/13-read-data | ++ local pod=mysql-client logger.go:42: 12:28:06 | demand-backup/13-read-data | ++ set +o xtrace logger.go:42: 12:28:07 | demand-backup/13-read-data | mysql-clienttrue logger.go:42: 12:28:07 | demand-backup/13-read-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:28:07 | demand-backup/13-read-data | ++ sed -e 's/mysql: //' logger.go:42: 12:28:07 | demand-backup/13-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:28:09 | demand-backup/13-read-data | + data=100500 logger.go:42: 12:28:09 | demand-backup/13-read-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 09-read-data-minio-backup-source-0 --from-literal=data=100500 logger.go:42: 12:28:09 | demand-backup/13-read-data | configmap/09-read-data-minio-backup-source-0 created logger.go:42: 12:28:09 | demand-backup/13-read-data | + for i in 0 1 2 logger.go:42: 12:28:09 | demand-backup/13-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:28:09 | demand-backup/13-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:28:09 | demand-backup/13-read-data | ++ local 'uri=-h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:28:09 | demand-backup/13-read-data | ++ local pod= logger.go:42: 12:28:09 | demand-backup/13-read-data | +++ get_client_pod logger.go:42: 12:28:09 | demand-backup/13-read-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:28:10 | demand-backup/13-read-data | ++ client_pod=mysql-client logger.go:42: 12:28:10 | demand-backup/13-read-data | ++ wait_pod mysql-client logger.go:42: 12:28:10 | demand-backup/13-read-data | ++ local pod=mysql-client logger.go:42: 12:28:10 | demand-backup/13-read-data | ++ set +o xtrace logger.go:42: 12:28:11 | demand-backup/13-read-data | mysql-clienttrue logger.go:42: 12:28:11 | demand-backup/13-read-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:28:11 | demand-backup/13-read-data | ++ sed -e 's/mysql: //' logger.go:42: 12:28:11 | demand-backup/13-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:28:12 | demand-backup/13-read-data | + data=100500 logger.go:42: 12:28:12 | demand-backup/13-read-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 09-read-data-minio-backup-source-1 --from-literal=data=100500 logger.go:42: 12:28:13 | demand-backup/13-read-data | configmap/09-read-data-minio-backup-source-1 created logger.go:42: 12:28:13 | demand-backup/13-read-data | + for i in 0 1 2 logger.go:42: 12:28:13 | demand-backup/13-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:28:13 | demand-backup/13-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:28:13 | demand-backup/13-read-data | ++ local 'uri=-h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:28:13 | demand-backup/13-read-data | ++ local pod= logger.go:42: 12:28:13 | demand-backup/13-read-data | +++ get_client_pod logger.go:42: 12:28:13 | demand-backup/13-read-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:28:14 | demand-backup/13-read-data | ++ client_pod=mysql-client logger.go:42: 12:28:14 | demand-backup/13-read-data | ++ wait_pod mysql-client logger.go:42: 12:28:14 | demand-backup/13-read-data | ++ local pod=mysql-client logger.go:42: 12:28:14 | demand-backup/13-read-data | ++ set +o xtrace logger.go:42: 12:28:14 | demand-backup/13-read-data | mysql-clienttrue logger.go:42: 12:28:14 | demand-backup/13-read-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:28:14 | demand-backup/13-read-data | ++ sed -e 's/mysql: //' logger.go:42: 12:28:14 | demand-backup/13-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:28:16 | demand-backup/13-read-data | + data=100500 logger.go:42: 12:28:16 | demand-backup/13-read-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 09-read-data-minio-backup-source-2 --from-literal=data=100500 logger.go:42: 12:28:17 | demand-backup/13-read-data | configmap/09-read-data-minio-backup-source-2 created logger.go:42: 12:28:18 | demand-backup/13-read-data | test step completed 13-read-data logger.go:42: 12:28:18 | demand-backup/14-create-backup-s3 | starting test step 14-create-backup-s3 logger.go:42: 12:28:20 | demand-backup/14-create-backup-s3 | PerconaServerMySQLBackup:kuttl-test-cosmic-ghoul/demand-backup-s3 created logger.go:42: 12:28:35 | demand-backup/14-create-backup-s3 | test step completed 14-create-backup-s3 logger.go:42: 12:28:35 | demand-backup/15-delete-data | starting test step 15-delete-data logger.go:42: 12:28:35 | demand-backup/15-delete-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "TRUNCATE TABLE myDB.myTable" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password" cluster_name=$(get_cluster_name) for i in 0 1 2; do data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 08-delete-data-s3-${i} --from-literal=data="${data}" done] logger.go:42: 12:28:35 | demand-backup/15-delete-data | + source ../../functions logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ realpath ../../.. logger.go:42: 12:28:35 | demand-backup/15-delete-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:28:35 | demand-backup/15-delete-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:28:35 | demand-backup/15-delete-data | ++++ mktemp -d logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ export TEMP_DIR=/tmp/tmp.tmbvPnakG8 logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ TEMP_DIR=/tmp/tmp.tmbvPnakG8 logger.go:42: 12:28:35 | demand-backup/15-delete-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ GIT_BRANCH=PR-339 logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ VERSION=PR-339-3a8483b logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:28:35 | demand-backup/15-delete-data | ++++ which gdate logger.go:42: 12:28:35 | demand-backup/15-delete-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:28:35 | demand-backup/15-delete-data | ++++ which date logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ date=/usr/bin/date logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ command -v oc logger.go:42: 12:28:35 | demand-backup/15-delete-data | +++ oc get projects logger.go:42: 12:28:41 | demand-backup/15-delete-data | error: the server doesn't have a resource type "projects" logger.go:42: 12:28:41 | demand-backup/15-delete-data | +++ grep '^minikube' logger.go:42: 12:28:41 | demand-backup/15-delete-data | +++ kubectl get nodes logger.go:42: 12:28:42 | demand-backup/15-delete-data | ++++ pwd logger.go:42: 12:28:42 | demand-backup/15-delete-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/demand-backup logger.go:42: 12:28:42 | demand-backup/15-delete-data | ++ test_name=demand-backup logger.go:42: 12:28:42 | demand-backup/15-delete-data | +++ get_cluster_name logger.go:42: 12:28:42 | demand-backup/15-delete-data | +++ kubectl -n kuttl-test-cosmic-ghoul get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 12:28:43 | demand-backup/15-delete-data | ++ get_haproxy_svc demand-backup logger.go:42: 12:28:43 | demand-backup/15-delete-data | ++ local cluster=demand-backup logger.go:42: 12:28:43 | demand-backup/15-delete-data | ++ echo demand-backup-haproxy logger.go:42: 12:28:43 | demand-backup/15-delete-data | + run_mysql 'TRUNCATE TABLE myDB.myTable' '-h demand-backup-haproxy -uroot -proot_password' logger.go:42: 12:28:43 | demand-backup/15-delete-data | + local 'command=TRUNCATE TABLE myDB.myTable' logger.go:42: 12:28:43 | demand-backup/15-delete-data | + local 'uri=-h demand-backup-haproxy -uroot -proot_password' logger.go:42: 12:28:43 | demand-backup/15-delete-data | + local pod= logger.go:42: 12:28:43 | demand-backup/15-delete-data | ++ get_client_pod logger.go:42: 12:28:43 | demand-backup/15-delete-data | ++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:28:43 | demand-backup/15-delete-data | + client_pod=mysql-client logger.go:42: 12:28:43 | demand-backup/15-delete-data | + wait_pod mysql-client logger.go:42: 12:28:43 | demand-backup/15-delete-data | + local pod=mysql-client logger.go:42: 12:28:43 | demand-backup/15-delete-data | + set +o xtrace logger.go:42: 12:28:44 | demand-backup/15-delete-data | mysql-clienttrue logger.go:42: 12:28:44 | demand-backup/15-delete-data | + kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "TRUNCATE TABLE myDB.myTable" | mysql -sN -h demand-backup-haproxy -uroot -proot_password' logger.go:42: 12:28:44 | demand-backup/15-delete-data | + sed -e 's/mysql: //' logger.go:42: 12:28:44 | demand-backup/15-delete-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:28:46 | demand-backup/15-delete-data | + : logger.go:42: 12:28:46 | demand-backup/15-delete-data | ++ get_cluster_name logger.go:42: 12:28:46 | demand-backup/15-delete-data | ++ kubectl -n kuttl-test-cosmic-ghoul get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 12:28:46 | demand-backup/15-delete-data | + cluster_name=demand-backup logger.go:42: 12:28:46 | demand-backup/15-delete-data | + for i in 0 1 2 logger.go:42: 12:28:46 | demand-backup/15-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:28:46 | demand-backup/15-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:28:46 | demand-backup/15-delete-data | ++ local 'uri=-h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:28:46 | demand-backup/15-delete-data | ++ local pod= logger.go:42: 12:28:46 | demand-backup/15-delete-data | +++ get_client_pod logger.go:42: 12:28:46 | demand-backup/15-delete-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:28:47 | demand-backup/15-delete-data | ++ client_pod=mysql-client logger.go:42: 12:28:47 | demand-backup/15-delete-data | ++ wait_pod mysql-client logger.go:42: 12:28:47 | demand-backup/15-delete-data | ++ local pod=mysql-client logger.go:42: 12:28:47 | demand-backup/15-delete-data | ++ set +o xtrace logger.go:42: 12:28:48 | demand-backup/15-delete-data | mysql-clienttrue logger.go:42: 12:28:48 | demand-backup/15-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 12:28:48 | demand-backup/15-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:28:48 | demand-backup/15-delete-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:28:49 | demand-backup/15-delete-data | ++ : logger.go:42: 12:28:49 | demand-backup/15-delete-data | + data= logger.go:42: 12:28:49 | demand-backup/15-delete-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 08-delete-data-s3-0 --from-literal=data= logger.go:42: 12:28:50 | demand-backup/15-delete-data | configmap/08-delete-data-s3-0 created logger.go:42: 12:28:50 | demand-backup/15-delete-data | + for i in 0 1 2 logger.go:42: 12:28:50 | demand-backup/15-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:28:50 | demand-backup/15-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:28:50 | demand-backup/15-delete-data | ++ local 'uri=-h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:28:50 | demand-backup/15-delete-data | ++ local pod= logger.go:42: 12:28:50 | demand-backup/15-delete-data | +++ get_client_pod logger.go:42: 12:28:50 | demand-backup/15-delete-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:28:51 | demand-backup/15-delete-data | ++ client_pod=mysql-client logger.go:42: 12:28:51 | demand-backup/15-delete-data | ++ wait_pod mysql-client logger.go:42: 12:28:51 | demand-backup/15-delete-data | ++ local pod=mysql-client logger.go:42: 12:28:51 | demand-backup/15-delete-data | ++ set +o xtrace logger.go:42: 12:28:52 | demand-backup/15-delete-data | mysql-clienttrue logger.go:42: 12:28:52 | demand-backup/15-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 12:28:52 | demand-backup/15-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:28:52 | demand-backup/15-delete-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:28:53 | demand-backup/15-delete-data | ++ : logger.go:42: 12:28:53 | demand-backup/15-delete-data | + data= logger.go:42: 12:28:53 | demand-backup/15-delete-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 08-delete-data-s3-1 --from-literal=data= logger.go:42: 12:28:54 | demand-backup/15-delete-data | configmap/08-delete-data-s3-1 created logger.go:42: 12:28:54 | demand-backup/15-delete-data | + for i in 0 1 2 logger.go:42: 12:28:54 | demand-backup/15-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:28:54 | demand-backup/15-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:28:54 | demand-backup/15-delete-data | ++ local 'uri=-h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:28:54 | demand-backup/15-delete-data | ++ local pod= logger.go:42: 12:28:54 | demand-backup/15-delete-data | +++ get_client_pod logger.go:42: 12:28:54 | demand-backup/15-delete-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:28:55 | demand-backup/15-delete-data | ++ client_pod=mysql-client logger.go:42: 12:28:55 | demand-backup/15-delete-data | ++ wait_pod mysql-client logger.go:42: 12:28:55 | demand-backup/15-delete-data | ++ local pod=mysql-client logger.go:42: 12:28:55 | demand-backup/15-delete-data | ++ set +o xtrace logger.go:42: 12:28:55 | demand-backup/15-delete-data | mysql-clienttrue logger.go:42: 12:28:55 | demand-backup/15-delete-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:28:55 | demand-backup/15-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 12:28:55 | demand-backup/15-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:28:57 | demand-backup/15-delete-data | ++ : logger.go:42: 12:28:57 | demand-backup/15-delete-data | + data= logger.go:42: 12:28:57 | demand-backup/15-delete-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 08-delete-data-s3-2 --from-literal=data= logger.go:42: 12:28:57 | demand-backup/15-delete-data | configmap/08-delete-data-s3-2 created logger.go:42: 12:28:59 | demand-backup/15-delete-data | test step completed 15-delete-data logger.go:42: 12:28:59 | demand-backup/16-restore-from-s3 | starting test step 16-restore-from-s3 logger.go:42: 12:29:00 | demand-backup/16-restore-from-s3 | PerconaServerMySQLRestore:kuttl-test-cosmic-ghoul/demand-backup-restore-s3 created logger.go:42: 12:33:51 | demand-backup/16-restore-from-s3 | test step completed 16-restore-from-s3 logger.go:42: 12:33:51 | demand-backup/17-read-data | starting test step 17-read-data logger.go:42: 12:33:51 | demand-backup/17-read-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions cluster_name=$(get_cluster_name) for i in 0 1 2; do data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 06-read-data-s3-${i} --from-literal=data="${data}" done] logger.go:42: 12:33:51 | demand-backup/17-read-data | + source ../../functions logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ realpath ../../.. logger.go:42: 12:33:51 | demand-backup/17-read-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:33:51 | demand-backup/17-read-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:33:51 | demand-backup/17-read-data | ++++ mktemp -d logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ export TEMP_DIR=/tmp/tmp.BWf8dYpEVZ logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ TEMP_DIR=/tmp/tmp.BWf8dYpEVZ logger.go:42: 12:33:51 | demand-backup/17-read-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ GIT_BRANCH=PR-339 logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ VERSION=PR-339-3a8483b logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:33:51 | demand-backup/17-read-data | ++++ which gdate logger.go:42: 12:33:51 | demand-backup/17-read-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:33:51 | demand-backup/17-read-data | ++++ which date logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ date=/usr/bin/date logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ command -v oc logger.go:42: 12:33:51 | demand-backup/17-read-data | +++ oc get projects logger.go:42: 12:33:57 | demand-backup/17-read-data | error: the server doesn't have a resource type "projects" logger.go:42: 12:33:57 | demand-backup/17-read-data | +++ grep '^minikube' logger.go:42: 12:33:57 | demand-backup/17-read-data | +++ kubectl get nodes logger.go:42: 12:33:58 | demand-backup/17-read-data | ++++ pwd logger.go:42: 12:33:58 | demand-backup/17-read-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/demand-backup logger.go:42: 12:33:58 | demand-backup/17-read-data | ++ test_name=demand-backup logger.go:42: 12:33:58 | demand-backup/17-read-data | ++ get_cluster_name logger.go:42: 12:33:58 | demand-backup/17-read-data | ++ kubectl -n kuttl-test-cosmic-ghoul get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 12:33:59 | demand-backup/17-read-data | + cluster_name=demand-backup logger.go:42: 12:33:59 | demand-backup/17-read-data | + for i in 0 1 2 logger.go:42: 12:33:59 | demand-backup/17-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:33:59 | demand-backup/17-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:33:59 | demand-backup/17-read-data | ++ local 'uri=-h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:33:59 | demand-backup/17-read-data | ++ local pod= logger.go:42: 12:33:59 | demand-backup/17-read-data | +++ get_client_pod logger.go:42: 12:33:59 | demand-backup/17-read-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:33:59 | demand-backup/17-read-data | ++ client_pod=mysql-client logger.go:42: 12:33:59 | demand-backup/17-read-data | ++ wait_pod mysql-client logger.go:42: 12:33:59 | demand-backup/17-read-data | ++ local pod=mysql-client logger.go:42: 12:33:59 | demand-backup/17-read-data | ++ set +o xtrace logger.go:42: 12:34:00 | demand-backup/17-read-data | mysql-clienttrue logger.go:42: 12:34:00 | demand-backup/17-read-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:34:00 | demand-backup/17-read-data | ++ sed -e 's/mysql: //' logger.go:42: 12:34:00 | demand-backup/17-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:34:02 | demand-backup/17-read-data | + data=100500 logger.go:42: 12:34:02 | demand-backup/17-read-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 06-read-data-s3-0 --from-literal=data=100500 logger.go:42: 12:34:02 | demand-backup/17-read-data | configmap/06-read-data-s3-0 created logger.go:42: 12:34:02 | demand-backup/17-read-data | + for i in 0 1 2 logger.go:42: 12:34:02 | demand-backup/17-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:34:02 | demand-backup/17-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:34:02 | demand-backup/17-read-data | ++ local 'uri=-h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:34:02 | demand-backup/17-read-data | ++ local pod= logger.go:42: 12:34:02 | demand-backup/17-read-data | +++ get_client_pod logger.go:42: 12:34:02 | demand-backup/17-read-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:34:03 | demand-backup/17-read-data | ++ client_pod=mysql-client logger.go:42: 12:34:03 | demand-backup/17-read-data | ++ wait_pod mysql-client logger.go:42: 12:34:03 | demand-backup/17-read-data | ++ local pod=mysql-client logger.go:42: 12:34:03 | demand-backup/17-read-data | ++ set +o xtrace logger.go:42: 12:34:04 | demand-backup/17-read-data | mysql-clienttrue logger.go:42: 12:34:04 | demand-backup/17-read-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:34:04 | demand-backup/17-read-data | ++ sed -e 's/mysql: //' logger.go:42: 12:34:04 | demand-backup/17-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:34:05 | demand-backup/17-read-data | + data=100500 logger.go:42: 12:34:05 | demand-backup/17-read-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 06-read-data-s3-1 --from-literal=data=100500 logger.go:42: 12:34:06 | demand-backup/17-read-data | configmap/06-read-data-s3-1 created logger.go:42: 12:34:06 | demand-backup/17-read-data | + for i in 0 1 2 logger.go:42: 12:34:06 | demand-backup/17-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:34:06 | demand-backup/17-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:34:06 | demand-backup/17-read-data | ++ local 'uri=-h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:34:06 | demand-backup/17-read-data | ++ local pod= logger.go:42: 12:34:06 | demand-backup/17-read-data | +++ get_client_pod logger.go:42: 12:34:06 | demand-backup/17-read-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:34:07 | demand-backup/17-read-data | ++ client_pod=mysql-client logger.go:42: 12:34:07 | demand-backup/17-read-data | ++ wait_pod mysql-client logger.go:42: 12:34:07 | demand-backup/17-read-data | ++ local pod=mysql-client logger.go:42: 12:34:07 | demand-backup/17-read-data | ++ set +o xtrace logger.go:42: 12:34:08 | demand-backup/17-read-data | mysql-clienttrue logger.go:42: 12:34:08 | demand-backup/17-read-data | ++ sed -e 's/mysql: //' logger.go:42: 12:34:08 | demand-backup/17-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:34:08 | demand-backup/17-read-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:34:09 | demand-backup/17-read-data | + data=100500 logger.go:42: 12:34:09 | demand-backup/17-read-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 06-read-data-s3-2 --from-literal=data=100500 logger.go:42: 12:34:10 | demand-backup/17-read-data | configmap/06-read-data-s3-2 created logger.go:42: 12:34:12 | demand-backup/17-read-data | test step completed 17-read-data logger.go:42: 12:34:12 | demand-backup/18-create-backup-gcp | starting test step 18-create-backup-gcp logger.go:42: 12:34:13 | demand-backup/18-create-backup-gcp | PerconaServerMySQLBackup:kuttl-test-cosmic-ghoul/demand-backup-gcp created logger.go:42: 12:34:29 | demand-backup/18-create-backup-gcp | test step completed 18-create-backup-gcp logger.go:42: 12:34:29 | demand-backup/19-delete-data | starting test step 19-delete-data logger.go:42: 12:34:29 | demand-backup/19-delete-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "TRUNCATE TABLE myDB.myTable" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password" cluster_name=$(get_cluster_name) for i in 0 1 2; do data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 12-delete-data-gcp-${i} --from-literal=data="${data}" done] logger.go:42: 12:34:29 | demand-backup/19-delete-data | + source ../../functions logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ realpath ../../.. logger.go:42: 12:34:29 | demand-backup/19-delete-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:34:29 | demand-backup/19-delete-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:34:29 | demand-backup/19-delete-data | ++++ mktemp -d logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ export TEMP_DIR=/tmp/tmp.NcP05nxdEi logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ TEMP_DIR=/tmp/tmp.NcP05nxdEi logger.go:42: 12:34:29 | demand-backup/19-delete-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ GIT_BRANCH=PR-339 logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ VERSION=PR-339-3a8483b logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:34:29 | demand-backup/19-delete-data | ++++ which gdate logger.go:42: 12:34:29 | demand-backup/19-delete-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:34:29 | demand-backup/19-delete-data | ++++ which date logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ date=/usr/bin/date logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ command -v oc logger.go:42: 12:34:29 | demand-backup/19-delete-data | +++ oc get projects logger.go:42: 12:34:34 | demand-backup/19-delete-data | error: the server doesn't have a resource type "projects" logger.go:42: 12:34:34 | demand-backup/19-delete-data | +++ kubectl get nodes logger.go:42: 12:34:34 | demand-backup/19-delete-data | +++ grep '^minikube' logger.go:42: 12:34:35 | demand-backup/19-delete-data | ++++ pwd logger.go:42: 12:34:35 | demand-backup/19-delete-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/demand-backup logger.go:42: 12:34:35 | demand-backup/19-delete-data | ++ test_name=demand-backup logger.go:42: 12:34:35 | demand-backup/19-delete-data | +++ get_cluster_name logger.go:42: 12:34:35 | demand-backup/19-delete-data | +++ kubectl -n kuttl-test-cosmic-ghoul get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 12:34:36 | demand-backup/19-delete-data | ++ get_haproxy_svc demand-backup logger.go:42: 12:34:36 | demand-backup/19-delete-data | ++ local cluster=demand-backup logger.go:42: 12:34:36 | demand-backup/19-delete-data | ++ echo demand-backup-haproxy logger.go:42: 12:34:36 | demand-backup/19-delete-data | + run_mysql 'TRUNCATE TABLE myDB.myTable' '-h demand-backup-haproxy -uroot -proot_password' logger.go:42: 12:34:36 | demand-backup/19-delete-data | + local 'command=TRUNCATE TABLE myDB.myTable' logger.go:42: 12:34:36 | demand-backup/19-delete-data | + local 'uri=-h demand-backup-haproxy -uroot -proot_password' logger.go:42: 12:34:36 | demand-backup/19-delete-data | + local pod= logger.go:42: 12:34:36 | demand-backup/19-delete-data | ++ get_client_pod logger.go:42: 12:34:36 | demand-backup/19-delete-data | ++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:34:37 | demand-backup/19-delete-data | + client_pod=mysql-client logger.go:42: 12:34:37 | demand-backup/19-delete-data | + wait_pod mysql-client logger.go:42: 12:34:37 | demand-backup/19-delete-data | + local pod=mysql-client logger.go:42: 12:34:37 | demand-backup/19-delete-data | + set +o xtrace logger.go:42: 12:34:37 | demand-backup/19-delete-data | mysql-clienttrue logger.go:42: 12:34:37 | demand-backup/19-delete-data | + kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "TRUNCATE TABLE myDB.myTable" | mysql -sN -h demand-backup-haproxy -uroot -proot_password' logger.go:42: 12:34:37 | demand-backup/19-delete-data | + sed -e 's/mysql: //' logger.go:42: 12:34:37 | demand-backup/19-delete-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:34:39 | demand-backup/19-delete-data | + : logger.go:42: 12:34:39 | demand-backup/19-delete-data | ++ get_cluster_name logger.go:42: 12:34:39 | demand-backup/19-delete-data | ++ kubectl -n kuttl-test-cosmic-ghoul get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 12:34:40 | demand-backup/19-delete-data | + cluster_name=demand-backup logger.go:42: 12:34:40 | demand-backup/19-delete-data | + for i in 0 1 2 logger.go:42: 12:34:40 | demand-backup/19-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:34:40 | demand-backup/19-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:34:40 | demand-backup/19-delete-data | ++ local 'uri=-h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:34:40 | demand-backup/19-delete-data | ++ local pod= logger.go:42: 12:34:40 | demand-backup/19-delete-data | +++ get_client_pod logger.go:42: 12:34:40 | demand-backup/19-delete-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:34:40 | demand-backup/19-delete-data | ++ client_pod=mysql-client logger.go:42: 12:34:40 | demand-backup/19-delete-data | ++ wait_pod mysql-client logger.go:42: 12:34:40 | demand-backup/19-delete-data | ++ local pod=mysql-client logger.go:42: 12:34:40 | demand-backup/19-delete-data | ++ set +o xtrace logger.go:42: 12:34:41 | demand-backup/19-delete-data | mysql-clienttrue logger.go:42: 12:34:41 | demand-backup/19-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 12:34:41 | demand-backup/19-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:34:41 | demand-backup/19-delete-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:34:43 | demand-backup/19-delete-data | ++ : logger.go:42: 12:34:43 | demand-backup/19-delete-data | + data= logger.go:42: 12:34:43 | demand-backup/19-delete-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 12-delete-data-gcp-0 --from-literal=data= logger.go:42: 12:34:43 | demand-backup/19-delete-data | configmap/12-delete-data-gcp-0 created logger.go:42: 12:34:43 | demand-backup/19-delete-data | + for i in 0 1 2 logger.go:42: 12:34:43 | demand-backup/19-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:34:43 | demand-backup/19-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:34:43 | demand-backup/19-delete-data | ++ local 'uri=-h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:34:43 | demand-backup/19-delete-data | ++ local pod= logger.go:42: 12:34:43 | demand-backup/19-delete-data | +++ get_client_pod logger.go:42: 12:34:43 | demand-backup/19-delete-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:34:44 | demand-backup/19-delete-data | ++ client_pod=mysql-client logger.go:42: 12:34:44 | demand-backup/19-delete-data | ++ wait_pod mysql-client logger.go:42: 12:34:44 | demand-backup/19-delete-data | ++ local pod=mysql-client logger.go:42: 12:34:44 | demand-backup/19-delete-data | ++ set +o xtrace logger.go:42: 12:34:45 | demand-backup/19-delete-data | mysql-clienttrue logger.go:42: 12:34:45 | demand-backup/19-delete-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:34:45 | demand-backup/19-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 12:34:45 | demand-backup/19-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:34:46 | demand-backup/19-delete-data | ++ : logger.go:42: 12:34:46 | demand-backup/19-delete-data | + data= logger.go:42: 12:34:46 | demand-backup/19-delete-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 12-delete-data-gcp-1 --from-literal=data= logger.go:42: 12:34:47 | demand-backup/19-delete-data | configmap/12-delete-data-gcp-1 created logger.go:42: 12:34:47 | demand-backup/19-delete-data | + for i in 0 1 2 logger.go:42: 12:34:47 | demand-backup/19-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:34:47 | demand-backup/19-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:34:47 | demand-backup/19-delete-data | ++ local 'uri=-h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:34:47 | demand-backup/19-delete-data | ++ local pod= logger.go:42: 12:34:47 | demand-backup/19-delete-data | +++ get_client_pod logger.go:42: 12:34:47 | demand-backup/19-delete-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:34:48 | demand-backup/19-delete-data | ++ client_pod=mysql-client logger.go:42: 12:34:48 | demand-backup/19-delete-data | ++ wait_pod mysql-client logger.go:42: 12:34:48 | demand-backup/19-delete-data | ++ local pod=mysql-client logger.go:42: 12:34:48 | demand-backup/19-delete-data | ++ set +o xtrace logger.go:42: 12:34:49 | demand-backup/19-delete-data | mysql-clienttrue logger.go:42: 12:34:49 | demand-backup/19-delete-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:34:49 | demand-backup/19-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 12:34:49 | demand-backup/19-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:34:50 | demand-backup/19-delete-data | ++ : logger.go:42: 12:34:50 | demand-backup/19-delete-data | + data= logger.go:42: 12:34:50 | demand-backup/19-delete-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 12-delete-data-gcp-2 --from-literal=data= logger.go:42: 12:34:51 | demand-backup/19-delete-data | configmap/12-delete-data-gcp-2 created logger.go:42: 12:34:52 | demand-backup/19-delete-data | test step completed 19-delete-data logger.go:42: 12:34:52 | demand-backup/20-restore-from-gcp | starting test step 20-restore-from-gcp logger.go:42: 12:34:54 | demand-backup/20-restore-from-gcp | PerconaServerMySQLRestore:kuttl-test-cosmic-ghoul/demand-backup-restore-gcp created logger.go:42: 12:40:02 | demand-backup/20-restore-from-gcp | test step completed 20-restore-from-gcp logger.go:42: 12:40:02 | demand-backup/21-read-data | starting test step 21-read-data logger.go:42: 12:40:02 | demand-backup/21-read-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions cluster_name=$(get_cluster_name) for i in 0 1 2; do data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 06-read-data-gcp-${i} --from-literal=data="${data}" done] logger.go:42: 12:40:02 | demand-backup/21-read-data | + source ../../functions logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ realpath ../../.. logger.go:42: 12:40:02 | demand-backup/21-read-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:40:02 | demand-backup/21-read-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:40:02 | demand-backup/21-read-data | ++++ mktemp -d logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ export TEMP_DIR=/tmp/tmp.HQj691KMli logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ TEMP_DIR=/tmp/tmp.HQj691KMli logger.go:42: 12:40:02 | demand-backup/21-read-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ GIT_BRANCH=PR-339 logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ VERSION=PR-339-3a8483b logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:40:02 | demand-backup/21-read-data | ++++ which gdate logger.go:42: 12:40:02 | demand-backup/21-read-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:40:02 | demand-backup/21-read-data | ++++ which date logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ date=/usr/bin/date logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ command -v oc logger.go:42: 12:40:02 | demand-backup/21-read-data | +++ oc get projects logger.go:42: 12:40:08 | demand-backup/21-read-data | error: the server doesn't have a resource type "projects" logger.go:42: 12:40:08 | demand-backup/21-read-data | +++ grep '^minikube' logger.go:42: 12:40:08 | demand-backup/21-read-data | +++ kubectl get nodes logger.go:42: 12:40:08 | demand-backup/21-read-data | ++++ pwd logger.go:42: 12:40:08 | demand-backup/21-read-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/demand-backup logger.go:42: 12:40:08 | demand-backup/21-read-data | ++ test_name=demand-backup logger.go:42: 12:40:08 | demand-backup/21-read-data | ++ get_cluster_name logger.go:42: 12:40:08 | demand-backup/21-read-data | ++ kubectl -n kuttl-test-cosmic-ghoul get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 12:40:09 | demand-backup/21-read-data | + cluster_name=demand-backup logger.go:42: 12:40:09 | demand-backup/21-read-data | + for i in 0 1 2 logger.go:42: 12:40:09 | demand-backup/21-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:40:09 | demand-backup/21-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:40:09 | demand-backup/21-read-data | ++ local 'uri=-h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:40:09 | demand-backup/21-read-data | ++ local pod= logger.go:42: 12:40:09 | demand-backup/21-read-data | +++ get_client_pod logger.go:42: 12:40:09 | demand-backup/21-read-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:40:10 | demand-backup/21-read-data | ++ client_pod=mysql-client logger.go:42: 12:40:10 | demand-backup/21-read-data | ++ wait_pod mysql-client logger.go:42: 12:40:10 | demand-backup/21-read-data | ++ local pod=mysql-client logger.go:42: 12:40:10 | demand-backup/21-read-data | ++ set +o xtrace logger.go:42: 12:40:11 | demand-backup/21-read-data | mysql-clienttrue logger.go:42: 12:40:11 | demand-backup/21-read-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:40:11 | demand-backup/21-read-data | ++ sed -e 's/mysql: //' logger.go:42: 12:40:11 | demand-backup/21-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:40:12 | demand-backup/21-read-data | + data=100500 logger.go:42: 12:40:12 | demand-backup/21-read-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 06-read-data-gcp-0 --from-literal=data=100500 logger.go:42: 12:40:13 | demand-backup/21-read-data | configmap/06-read-data-gcp-0 created logger.go:42: 12:40:13 | demand-backup/21-read-data | + for i in 0 1 2 logger.go:42: 12:40:13 | demand-backup/21-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:40:13 | demand-backup/21-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:40:13 | demand-backup/21-read-data | ++ local 'uri=-h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:40:13 | demand-backup/21-read-data | ++ local pod= logger.go:42: 12:40:13 | demand-backup/21-read-data | +++ get_client_pod logger.go:42: 12:40:13 | demand-backup/21-read-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:40:14 | demand-backup/21-read-data | ++ client_pod=mysql-client logger.go:42: 12:40:14 | demand-backup/21-read-data | ++ wait_pod mysql-client logger.go:42: 12:40:14 | demand-backup/21-read-data | ++ local pod=mysql-client logger.go:42: 12:40:14 | demand-backup/21-read-data | ++ set +o xtrace logger.go:42: 12:40:14 | demand-backup/21-read-data | mysql-clienttrue logger.go:42: 12:40:14 | demand-backup/21-read-data | ++ sed -e 's/mysql: //' logger.go:42: 12:40:14 | demand-backup/21-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:40:14 | demand-backup/21-read-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:40:16 | demand-backup/21-read-data | + data=100500 logger.go:42: 12:40:16 | demand-backup/21-read-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 06-read-data-gcp-1 --from-literal=data=100500 logger.go:42: 12:40:17 | demand-backup/21-read-data | configmap/06-read-data-gcp-1 created logger.go:42: 12:40:17 | demand-backup/21-read-data | + for i in 0 1 2 logger.go:42: 12:40:17 | demand-backup/21-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:40:17 | demand-backup/21-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:40:17 | demand-backup/21-read-data | ++ local 'uri=-h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:40:17 | demand-backup/21-read-data | ++ local pod= logger.go:42: 12:40:17 | demand-backup/21-read-data | +++ get_client_pod logger.go:42: 12:40:17 | demand-backup/21-read-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:40:17 | demand-backup/21-read-data | ++ client_pod=mysql-client logger.go:42: 12:40:17 | demand-backup/21-read-data | ++ wait_pod mysql-client logger.go:42: 12:40:17 | demand-backup/21-read-data | ++ local pod=mysql-client logger.go:42: 12:40:17 | demand-backup/21-read-data | ++ set +o xtrace logger.go:42: 12:40:18 | demand-backup/21-read-data | mysql-clienttrue logger.go:42: 12:40:18 | demand-backup/21-read-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:40:18 | demand-backup/21-read-data | ++ sed -e 's/mysql: //' logger.go:42: 12:40:18 | demand-backup/21-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:40:20 | demand-backup/21-read-data | + data=100500 logger.go:42: 12:40:20 | demand-backup/21-read-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 06-read-data-gcp-2 --from-literal=data=100500 logger.go:42: 12:40:20 | demand-backup/21-read-data | configmap/06-read-data-gcp-2 created logger.go:42: 12:40:22 | demand-backup/21-read-data | test step completed 21-read-data logger.go:42: 12:40:22 | demand-backup/22-create-backup-azure | starting test step 22-create-backup-azure logger.go:42: 12:40:23 | demand-backup/22-create-backup-azure | PerconaServerMySQLBackup:kuttl-test-cosmic-ghoul/demand-backup-azure created logger.go:42: 12:40:39 | demand-backup/22-create-backup-azure | test step completed 22-create-backup-azure logger.go:42: 12:40:39 | demand-backup/23-delete-data | starting test step 23-delete-data logger.go:42: 12:40:39 | demand-backup/23-delete-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "TRUNCATE TABLE myDB.myTable" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password" cluster_name=$(get_cluster_name) for i in 0 1 2; do data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 16-delete-data-azure-${i} --from-literal=data="${data}" done] logger.go:42: 12:40:39 | demand-backup/23-delete-data | + source ../../functions logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ realpath ../../.. logger.go:42: 12:40:39 | demand-backup/23-delete-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:40:39 | demand-backup/23-delete-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:40:39 | demand-backup/23-delete-data | ++++ mktemp -d logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ export TEMP_DIR=/tmp/tmp.h9KXnFvnmH logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ TEMP_DIR=/tmp/tmp.h9KXnFvnmH logger.go:42: 12:40:39 | demand-backup/23-delete-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ GIT_BRANCH=PR-339 logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ VERSION=PR-339-3a8483b logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:40:39 | demand-backup/23-delete-data | ++++ which gdate logger.go:42: 12:40:39 | demand-backup/23-delete-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:40:39 | demand-backup/23-delete-data | ++++ which date logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ date=/usr/bin/date logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ command -v oc logger.go:42: 12:40:39 | demand-backup/23-delete-data | +++ oc get projects logger.go:42: 12:40:45 | demand-backup/23-delete-data | error: the server doesn't have a resource type "projects" logger.go:42: 12:40:45 | demand-backup/23-delete-data | +++ kubectl get nodes logger.go:42: 12:40:45 | demand-backup/23-delete-data | +++ grep '^minikube' logger.go:42: 12:40:46 | demand-backup/23-delete-data | ++++ pwd logger.go:42: 12:40:46 | demand-backup/23-delete-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/demand-backup logger.go:42: 12:40:46 | demand-backup/23-delete-data | ++ test_name=demand-backup logger.go:42: 12:40:46 | demand-backup/23-delete-data | +++ get_cluster_name logger.go:42: 12:40:46 | demand-backup/23-delete-data | +++ kubectl -n kuttl-test-cosmic-ghoul get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 12:40:46 | demand-backup/23-delete-data | ++ get_haproxy_svc demand-backup logger.go:42: 12:40:46 | demand-backup/23-delete-data | ++ local cluster=demand-backup logger.go:42: 12:40:46 | demand-backup/23-delete-data | ++ echo demand-backup-haproxy logger.go:42: 12:40:46 | demand-backup/23-delete-data | + run_mysql 'TRUNCATE TABLE myDB.myTable' '-h demand-backup-haproxy -uroot -proot_password' logger.go:42: 12:40:46 | demand-backup/23-delete-data | + local 'command=TRUNCATE TABLE myDB.myTable' logger.go:42: 12:40:46 | demand-backup/23-delete-data | + local 'uri=-h demand-backup-haproxy -uroot -proot_password' logger.go:42: 12:40:46 | demand-backup/23-delete-data | + local pod= logger.go:42: 12:40:46 | demand-backup/23-delete-data | ++ get_client_pod logger.go:42: 12:40:46 | demand-backup/23-delete-data | ++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:40:47 | demand-backup/23-delete-data | + client_pod=mysql-client logger.go:42: 12:40:47 | demand-backup/23-delete-data | + wait_pod mysql-client logger.go:42: 12:40:47 | demand-backup/23-delete-data | + local pod=mysql-client logger.go:42: 12:40:47 | demand-backup/23-delete-data | + set +o xtrace logger.go:42: 12:40:48 | demand-backup/23-delete-data | mysql-clienttrue logger.go:42: 12:40:48 | demand-backup/23-delete-data | + sed -e 's/mysql: //' logger.go:42: 12:40:48 | demand-backup/23-delete-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:40:48 | demand-backup/23-delete-data | + kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "TRUNCATE TABLE myDB.myTable" | mysql -sN -h demand-backup-haproxy -uroot -proot_password' logger.go:42: 12:40:49 | demand-backup/23-delete-data | + : logger.go:42: 12:40:49 | demand-backup/23-delete-data | ++ get_cluster_name logger.go:42: 12:40:49 | demand-backup/23-delete-data | ++ kubectl -n kuttl-test-cosmic-ghoul get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 12:40:50 | demand-backup/23-delete-data | + cluster_name=demand-backup logger.go:42: 12:40:50 | demand-backup/23-delete-data | + for i in 0 1 2 logger.go:42: 12:40:50 | demand-backup/23-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:40:50 | demand-backup/23-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:40:50 | demand-backup/23-delete-data | ++ local 'uri=-h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:40:50 | demand-backup/23-delete-data | ++ local pod= logger.go:42: 12:40:50 | demand-backup/23-delete-data | +++ get_client_pod logger.go:42: 12:40:50 | demand-backup/23-delete-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:40:51 | demand-backup/23-delete-data | ++ client_pod=mysql-client logger.go:42: 12:40:51 | demand-backup/23-delete-data | ++ wait_pod mysql-client logger.go:42: 12:40:51 | demand-backup/23-delete-data | ++ local pod=mysql-client logger.go:42: 12:40:51 | demand-backup/23-delete-data | ++ set +o xtrace logger.go:42: 12:40:52 | demand-backup/23-delete-data | mysql-clienttrue logger.go:42: 12:40:52 | demand-backup/23-delete-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:40:52 | demand-backup/23-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 12:40:52 | demand-backup/23-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:40:53 | demand-backup/23-delete-data | ++ : logger.go:42: 12:40:53 | demand-backup/23-delete-data | + data= logger.go:42: 12:40:53 | demand-backup/23-delete-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 16-delete-data-azure-0 --from-literal=data= logger.go:42: 12:40:54 | demand-backup/23-delete-data | configmap/16-delete-data-azure-0 created logger.go:42: 12:40:54 | demand-backup/23-delete-data | + for i in 0 1 2 logger.go:42: 12:40:54 | demand-backup/23-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:40:54 | demand-backup/23-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:40:54 | demand-backup/23-delete-data | ++ local 'uri=-h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:40:54 | demand-backup/23-delete-data | ++ local pod= logger.go:42: 12:40:54 | demand-backup/23-delete-data | +++ get_client_pod logger.go:42: 12:40:54 | demand-backup/23-delete-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:40:55 | demand-backup/23-delete-data | ++ client_pod=mysql-client logger.go:42: 12:40:55 | demand-backup/23-delete-data | ++ wait_pod mysql-client logger.go:42: 12:40:55 | demand-backup/23-delete-data | ++ local pod=mysql-client logger.go:42: 12:40:55 | demand-backup/23-delete-data | ++ set +o xtrace logger.go:42: 12:40:55 | demand-backup/23-delete-data | mysql-clienttrue logger.go:42: 12:40:55 | demand-backup/23-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 12:40:55 | demand-backup/23-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:40:55 | demand-backup/23-delete-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:40:57 | demand-backup/23-delete-data | ++ : logger.go:42: 12:40:57 | demand-backup/23-delete-data | + data= logger.go:42: 12:40:57 | demand-backup/23-delete-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 16-delete-data-azure-1 --from-literal=data= logger.go:42: 12:40:58 | demand-backup/23-delete-data | configmap/16-delete-data-azure-1 created logger.go:42: 12:40:58 | demand-backup/23-delete-data | + for i in 0 1 2 logger.go:42: 12:40:58 | demand-backup/23-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:40:58 | demand-backup/23-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:40:58 | demand-backup/23-delete-data | ++ local 'uri=-h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:40:58 | demand-backup/23-delete-data | ++ local pod= logger.go:42: 12:40:58 | demand-backup/23-delete-data | +++ get_client_pod logger.go:42: 12:40:58 | demand-backup/23-delete-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:40:58 | demand-backup/23-delete-data | ++ client_pod=mysql-client logger.go:42: 12:40:58 | demand-backup/23-delete-data | ++ wait_pod mysql-client logger.go:42: 12:40:58 | demand-backup/23-delete-data | ++ local pod=mysql-client logger.go:42: 12:40:58 | demand-backup/23-delete-data | ++ set +o xtrace logger.go:42: 12:40:59 | demand-backup/23-delete-data | mysql-clienttrue logger.go:42: 12:40:59 | demand-backup/23-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 12:40:59 | demand-backup/23-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:40:59 | demand-backup/23-delete-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:41:01 | demand-backup/23-delete-data | ++ : logger.go:42: 12:41:01 | demand-backup/23-delete-data | + data= logger.go:42: 12:41:01 | demand-backup/23-delete-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 16-delete-data-azure-2 --from-literal=data= logger.go:42: 12:41:01 | demand-backup/23-delete-data | configmap/16-delete-data-azure-2 created logger.go:42: 12:41:03 | demand-backup/23-delete-data | test step completed 23-delete-data logger.go:42: 12:41:03 | demand-backup/24-restore-from-azure | starting test step 24-restore-from-azure logger.go:42: 12:41:04 | demand-backup/24-restore-from-azure | PerconaServerMySQLRestore:kuttl-test-cosmic-ghoul/demand-backup-restore-azure created logger.go:42: 12:46:01 | demand-backup/24-restore-from-azure | test step completed 24-restore-from-azure logger.go:42: 12:46:01 | demand-backup/25-read-data | starting test step 25-read-data logger.go:42: 12:46:01 | demand-backup/25-read-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions cluster_name=$(get_cluster_name) for i in 0 1 2; do data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 06-read-data-azure-${i} --from-literal=data="${data}" done] logger.go:42: 12:46:01 | demand-backup/25-read-data | + source ../../functions logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ realpath ../../.. logger.go:42: 12:46:01 | demand-backup/25-read-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:46:01 | demand-backup/25-read-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:46:01 | demand-backup/25-read-data | ++++ mktemp -d logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ export TEMP_DIR=/tmp/tmp.apH1NIKMwr logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ TEMP_DIR=/tmp/tmp.apH1NIKMwr logger.go:42: 12:46:01 | demand-backup/25-read-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ GIT_BRANCH=PR-339 logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ VERSION=PR-339-3a8483b logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:46:01 | demand-backup/25-read-data | ++++ which gdate logger.go:42: 12:46:01 | demand-backup/25-read-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:46:01 | demand-backup/25-read-data | ++++ which date logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ date=/usr/bin/date logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ command -v oc logger.go:42: 12:46:01 | demand-backup/25-read-data | +++ oc get projects logger.go:42: 12:46:07 | demand-backup/25-read-data | error: the server doesn't have a resource type "projects" logger.go:42: 12:46:07 | demand-backup/25-read-data | +++ grep '^minikube' logger.go:42: 12:46:07 | demand-backup/25-read-data | +++ kubectl get nodes logger.go:42: 12:46:07 | demand-backup/25-read-data | ++++ pwd logger.go:42: 12:46:07 | demand-backup/25-read-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/demand-backup logger.go:42: 12:46:07 | demand-backup/25-read-data | ++ test_name=demand-backup logger.go:42: 12:46:07 | demand-backup/25-read-data | ++ get_cluster_name logger.go:42: 12:46:07 | demand-backup/25-read-data | ++ kubectl -n kuttl-test-cosmic-ghoul get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 12:46:08 | demand-backup/25-read-data | + cluster_name=demand-backup logger.go:42: 12:46:08 | demand-backup/25-read-data | + for i in 0 1 2 logger.go:42: 12:46:08 | demand-backup/25-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:46:08 | demand-backup/25-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:46:08 | demand-backup/25-read-data | ++ local 'uri=-h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:46:08 | demand-backup/25-read-data | ++ local pod= logger.go:42: 12:46:08 | demand-backup/25-read-data | +++ get_client_pod logger.go:42: 12:46:08 | demand-backup/25-read-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:46:09 | demand-backup/25-read-data | ++ client_pod=mysql-client logger.go:42: 12:46:09 | demand-backup/25-read-data | ++ wait_pod mysql-client logger.go:42: 12:46:09 | demand-backup/25-read-data | ++ local pod=mysql-client logger.go:42: 12:46:09 | demand-backup/25-read-data | ++ set +o xtrace logger.go:42: 12:46:10 | demand-backup/25-read-data | mysql-clienttrue logger.go:42: 12:46:10 | demand-backup/25-read-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-0.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:46:10 | demand-backup/25-read-data | ++ sed -e 's/mysql: //' logger.go:42: 12:46:10 | demand-backup/25-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:46:11 | demand-backup/25-read-data | + data=100500 logger.go:42: 12:46:11 | demand-backup/25-read-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 06-read-data-azure-0 --from-literal=data=100500 logger.go:42: 12:46:12 | demand-backup/25-read-data | configmap/06-read-data-azure-0 created logger.go:42: 12:46:12 | demand-backup/25-read-data | + for i in 0 1 2 logger.go:42: 12:46:12 | demand-backup/25-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:46:12 | demand-backup/25-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:46:12 | demand-backup/25-read-data | ++ local 'uri=-h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:46:12 | demand-backup/25-read-data | ++ local pod= logger.go:42: 12:46:12 | demand-backup/25-read-data | +++ get_client_pod logger.go:42: 12:46:12 | demand-backup/25-read-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:46:13 | demand-backup/25-read-data | ++ client_pod=mysql-client logger.go:42: 12:46:13 | demand-backup/25-read-data | ++ wait_pod mysql-client logger.go:42: 12:46:13 | demand-backup/25-read-data | ++ local pod=mysql-client logger.go:42: 12:46:13 | demand-backup/25-read-data | ++ set +o xtrace logger.go:42: 12:46:13 | demand-backup/25-read-data | mysql-clienttrue logger.go:42: 12:46:13 | demand-backup/25-read-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-1.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:46:13 | demand-backup/25-read-data | ++ sed -e 's/mysql: //' logger.go:42: 12:46:13 | demand-backup/25-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:46:15 | demand-backup/25-read-data | + data=100500 logger.go:42: 12:46:15 | demand-backup/25-read-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 06-read-data-azure-1 --from-literal=data=100500 logger.go:42: 12:46:16 | demand-backup/25-read-data | configmap/06-read-data-azure-1 created logger.go:42: 12:46:16 | demand-backup/25-read-data | + for i in 0 1 2 logger.go:42: 12:46:16 | demand-backup/25-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:46:16 | demand-backup/25-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 12:46:16 | demand-backup/25-read-data | ++ local 'uri=-h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:46:16 | demand-backup/25-read-data | ++ local pod= logger.go:42: 12:46:16 | demand-backup/25-read-data | +++ get_client_pod logger.go:42: 12:46:16 | demand-backup/25-read-data | +++ kubectl -n kuttl-test-cosmic-ghoul get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:46:16 | demand-backup/25-read-data | ++ client_pod=mysql-client logger.go:42: 12:46:16 | demand-backup/25-read-data | ++ wait_pod mysql-client logger.go:42: 12:46:16 | demand-backup/25-read-data | ++ local pod=mysql-client logger.go:42: 12:46:16 | demand-backup/25-read-data | ++ set +o xtrace logger.go:42: 12:46:17 | demand-backup/25-read-data | mysql-clienttrue logger.go:42: 12:46:17 | demand-backup/25-read-data | ++ kubectl -n kuttl-test-cosmic-ghoul exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h demand-backup-mysql-2.demand-backup-mysql -uroot -proot_password' logger.go:42: 12:46:17 | demand-backup/25-read-data | ++ sed -e 's/mysql: //' logger.go:42: 12:46:17 | demand-backup/25-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:46:19 | demand-backup/25-read-data | + data=100500 logger.go:42: 12:46:19 | demand-backup/25-read-data | + kubectl create configmap -n kuttl-test-cosmic-ghoul 06-read-data-azure-2 --from-literal=data=100500 logger.go:42: 12:46:19 | demand-backup/25-read-data | configmap/06-read-data-azure-2 created logger.go:42: 12:46:21 | demand-backup/25-read-data | test step completed 25-read-data logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | starting test step 26-delete-all-backups logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | running command: [sh -c set -o errexit set -o xtrace source ../../functions kubectl delete ps-backup --all -n "${NAMESPACE}" backup_name_minio="demand-backup-minio" accessKey="$(kubectl -n "${NAMESPACE}" get secret minio-secret -o jsonpath='{.data.AWS_ACCESS_KEY_ID}' | base64 -d)" secretKey="$(kubectl -n "${NAMESPACE}" get secret minio-secret -o jsonpath='{.data.AWS_SECRET_ACCESS_KEY}' | base64 -d)" backup_exists=$( kubectl run -n "${NAMESPACE}" -i --rm aws-cli --image=perconalab/awscli --restart=Never -- \ /usr/bin/env AWS_ACCESS_KEY_ID="${accessKey}" AWS_SECRET_ACCESS_KEY="${secretKey}" AWS_DEFAULT_REGION=us-east-1 \ /usr/bin/aws --endpoint-url http://minio-service:9000 s3 ls 'operator-testing/' | grep -c "${backup_name_minio}/" | cat exit "${PIPESTATUS[0]}" ) if [[ 1 -eq $backup_exists ]]; then echo "Backup was not removed from bucket -- minio" exit 1 fi] logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | + source ../../functions logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ realpath ../../.. logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | ++++ mktemp -d logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ export TEMP_DIR=/tmp/tmp.ZIMjYbR8lm logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ TEMP_DIR=/tmp/tmp.ZIMjYbR8lm logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ GIT_BRANCH=PR-339 logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ VERSION=PR-339-3a8483b logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | ++++ which gdate logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | ++++ which date logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ date=/usr/bin/date logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ command -v oc logger.go:42: 12:46:21 | demand-backup/26-delete-all-backups | +++ oc get projects logger.go:42: 12:46:27 | demand-backup/26-delete-all-backups | error: the server doesn't have a resource type "projects" logger.go:42: 12:46:27 | demand-backup/26-delete-all-backups | +++ kubectl get nodes logger.go:42: 12:46:27 | demand-backup/26-delete-all-backups | +++ grep '^minikube' logger.go:42: 12:46:27 | demand-backup/26-delete-all-backups | ++++ pwd logger.go:42: 12:46:27 | demand-backup/26-delete-all-backups | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/demand-backup logger.go:42: 12:46:27 | demand-backup/26-delete-all-backups | ++ test_name=demand-backup logger.go:42: 12:46:27 | demand-backup/26-delete-all-backups | + kubectl delete ps-backup --all -n kuttl-test-cosmic-ghoul logger.go:42: 12:46:28 | demand-backup/26-delete-all-backups | perconaservermysqlbackup.ps.percona.com "demand-backup-azure" deleted logger.go:42: 12:46:28 | demand-backup/26-delete-all-backups | perconaservermysqlbackup.ps.percona.com "demand-backup-gcp" deleted logger.go:42: 12:46:29 | demand-backup/26-delete-all-backups | perconaservermysqlbackup.ps.percona.com "demand-backup-minio" deleted logger.go:42: 12:46:29 | demand-backup/26-delete-all-backups | perconaservermysqlbackup.ps.percona.com "demand-backup-s3" deleted logger.go:42: 12:46:33 | demand-backup/26-delete-all-backups | + backup_name_minio=demand-backup-minio logger.go:42: 12:46:33 | demand-backup/26-delete-all-backups | ++ kubectl -n kuttl-test-cosmic-ghoul get secret minio-secret -o 'jsonpath={.data.AWS_ACCESS_KEY_ID}' logger.go:42: 12:46:33 | demand-backup/26-delete-all-backups | ++ base64 -d logger.go:42: 12:46:34 | demand-backup/26-delete-all-backups | + accessKey='some-access$\n"-key' logger.go:42: 12:46:34 | demand-backup/26-delete-all-backups | ++ kubectl -n kuttl-test-cosmic-ghoul get secret minio-secret -o 'jsonpath={.data.AWS_SECRET_ACCESS_KEY}' logger.go:42: 12:46:34 | demand-backup/26-delete-all-backups | ++ base64 -d logger.go:42: 12:46:34 | demand-backup/26-delete-all-backups | + secretKey='some-$\n"secret-key' logger.go:42: 12:46:34 | demand-backup/26-delete-all-backups | ++ grep -c demand-backup-minio/ logger.go:42: 12:46:34 | demand-backup/26-delete-all-backups | ++ kubectl run -n kuttl-test-cosmic-ghoul -i --rm aws-cli --image=perconalab/awscli --restart=Never -- /usr/bin/env 'AWS_ACCESS_KEY_ID=some-access$\n"-key' 'AWS_SECRET_ACCESS_KEY=some-$\n"secret-key' AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url http://minio-service:9000 s3 ls operator-testing/ logger.go:42: 12:46:34 | demand-backup/26-delete-all-backups | ++ cat logger.go:42: 12:46:36 | demand-backup/26-delete-all-backups | If you don't see a command prompt, try pressing enter. logger.go:42: 12:46:37 | demand-backup/26-delete-all-backups | Error attaching, falling back to logs: Internal error occurred: error attaching to container: container is in CONTAINER_EXITED state logger.go:42: 12:46:40 | demand-backup/26-delete-all-backups | ++ exit 0 logger.go:42: 12:46:40 | demand-backup/26-delete-all-backups | + backup_exists=0 logger.go:42: 12:46:40 | demand-backup/26-delete-all-backups | + [[ 1 -eq 0 ]] logger.go:42: 12:46:41 | demand-backup/26-delete-all-backups | test step completed 26-delete-all-backups logger.go:42: 12:46:41 | demand-backup/27-drop-finalizer | starting test step 27-drop-finalizer logger.go:42: 12:46:42 | demand-backup/27-drop-finalizer | PerconaServerMySQL:kuttl-test-cosmic-ghoul/demand-backup updated logger.go:42: 12:46:42 | demand-backup/27-drop-finalizer | test step completed 27-drop-finalizer logger.go:42: 12:46:42 | demand-backup | demand-backup events from ns kuttl-test-cosmic-ghoul: logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:11 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-h42q5 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/percona-server-mysql-operator-6dcd78d857-h42q5 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:11 +0000 UTC Normal ReplicaSet.apps percona-server-mysql-operator-6dcd78d857 SuccessfulCreate Created pod: percona-server-mysql-operator-6dcd78d857-h42q5 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:11 +0000 UTC Normal Deployment.apps percona-server-mysql-operator ScalingReplicaSet Scaled up replica set percona-server-mysql-operator-6dcd78d857 to 1 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:13 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-h42q5.spec.containers{manager} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:14 +0000 UTC Normal Pod mysql-client Scheduled Successfully assigned kuttl-test-cosmic-ghoul/mysql-client to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:15 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Pulling Pulling image "percona/percona-server:8.0.25" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:20 +0000 UTC Normal PersistentVolumeClaim minio-service WaitForFirstConsumer waiting for first consumer to be created before binding logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:21 +0000 UTC Normal ReplicaSet.apps minio-service-6f47578c5b SuccessfulCreate Created pod: minio-service-6f47578c5b-2vr2g logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:21 +0000 UTC Normal Deployment.apps minio-service ScalingReplicaSet Scaled up replica set minio-service-6f47578c5b to 1 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:21 +0000 UTC Normal PersistentVolumeClaim minio-service ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:21 +0000 UTC Normal PersistentVolumeClaim minio-service Provisioning External provisioner is provisioning volume for claim "kuttl-test-cosmic-ghoul/minio-service" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:25 +0000 UTC Normal PersistentVolumeClaim minio-service ProvisioningSucceeded Successfully provisioned volume pvc-35c38c82-f0f4-4abf-a557-695f7d345625 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:26 +0000 UTC Normal Pod minio-service-6f47578c5b-2vr2g Scheduled Successfully assigned kuttl-test-cosmic-ghoul/minio-service-6f47578c5b-2vr2g to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w381 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:27 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Pulled Successfully pulled image "percona/percona-server:8.0.25" in 12.044464306s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:27 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Created Created container mysql-client logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:27 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Started Started container mysql-client logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:28 +0000 UTC Normal Lease.coordination.k8s.io 08db2feb.percona.com LeaderElection percona-server-mysql-operator-6dcd78d857-h42q5_c17e673a-1d81-421c-a950-170edf85c57a became leader logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:28 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-h42q5.spec.containers{manager} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 14.824480967s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:28 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-h42q5.spec.containers{manager} Created Created container manager logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:28 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-h42q5.spec.containers{manager} Started Started container manager logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:30 +0000 UTC Normal Pod minio-service-6f47578c5b-2vr2g SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-35c38c82-f0f4-4abf-a557-695f7d345625" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:32 +0000 UTC Normal Pod minio-service-6f47578c5b-2vr2g.spec.containers{minio} Pulling Pulling image "minio/minio:RELEASE.2020-11-19T23-48-16Z" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:34 +0000 UTC Normal Pod minio-service-6f47578c5b-2vr2g.spec.containers{minio} Pulled Successfully pulled image "minio/minio:RELEASE.2020-11-19T23-48-16Z" in 1.940498321s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:34 +0000 UTC Normal Pod minio-service-6f47578c5b-2vr2g.spec.containers{minio} Created Created container minio logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:34 +0000 UTC Normal Pod minio-service-6f47578c5b-2vr2g.spec.containers{minio} Started Started container minio logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:37 +0000 UTC Normal Pod aws-cli Scheduled Successfully assigned kuttl-test-cosmic-ghoul/aws-cli to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:38 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Pulling Pulling image "perconalab/awscli" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:42 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Pulled Successfully pulled image "perconalab/awscli" in 3.81845517s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:42 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Created Created container aws-cli logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:06:42 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Started Started container aws-cli logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:04 +0000 UTC Normal PersistentVolumeClaim datadir-demand-backup-mysql-0 WaitForFirstConsumer waiting for first consumer to be created before binding logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:04 +0000 UTC Normal StatefulSet.apps demand-backup-mysql SuccessfulCreate create Claim datadir-demand-backup-mysql-0 Pod demand-backup-mysql-0 in StatefulSet demand-backup-mysql success logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:04 +0000 UTC Normal StatefulSet.apps demand-backup-mysql SuccessfulCreate create Pod demand-backup-mysql-0 in StatefulSet demand-backup-mysql successful logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:04 +0000 UTC Normal Pod demand-backup-orc-0 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-orc-0 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:04 +0000 UTC Normal StatefulSet.apps demand-backup-orc SuccessfulCreate create Pod demand-backup-orc-0 in StatefulSet demand-backup-orc successful logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:05 +0000 UTC Normal PersistentVolumeClaim datadir-demand-backup-mysql-0 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:05 +0000 UTC Normal PersistentVolumeClaim datadir-demand-backup-mysql-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-cosmic-ghoul/datadir-demand-backup-mysql-0" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:05 +0000 UTC Normal Pod demand-backup-orc-0.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:05 +0000 UTC Normal Pod demand-backup-orc-0.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 404.604616ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:06 +0000 UTC Normal Pod demand-backup-orc-0.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:06 +0000 UTC Normal Pod demand-backup-orc-0.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:08 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:09 +0000 UTC Normal PersistentVolumeClaim datadir-demand-backup-mysql-0 ProvisioningSucceeded Successfully provisioned volume pvc-d0d9916e-94c2-4f55-b2d4-15e7447aeffb logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:09 +0000 UTC Normal Pod demand-backup-mysql-0 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-mysql-0 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:10 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 2.293896624s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:10 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Created Created container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:10 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Started Started container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:10 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:11 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 390.314024ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:11 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:11 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:17 +0000 UTC Normal Pod demand-backup-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-d0d9916e-94c2-4f55-b2d4-15e7447aeffb" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:20 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:20 +0000 UTC Warning Pod demand-backup-orc-0.spec.containers{orc} Unhealthy Liveness probe failed: Get "http://10.167.210.10:3000/api/lb-check": dial tcp 10.167.210.10:3000: connect: connection refused logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:35 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 15.030059549s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:35 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:35 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:40 +0000 UTC Normal Pod demand-backup-orc-1 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-orc-1 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:40 +0000 UTC Normal StatefulSet.apps demand-backup-orc SuccessfulCreate create Pod demand-backup-orc-1 in StatefulSet demand-backup-orc successful logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:42 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:42 +0000 UTC Normal Pod demand-backup-orc-1.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:43 +0000 UTC Normal Pod demand-backup-orc-1.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 377.023551ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:43 +0000 UTC Normal Pod demand-backup-orc-1.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:43 +0000 UTC Normal Pod demand-backup-orc-1.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:07:44 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:04 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 21.422432386s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:04 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Created Created container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:04 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Started Started container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:04 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:07 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 22.531379988s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:07 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Created Created container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:07 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Started Started container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:07 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:18 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 13.765815033s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:18 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:18 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:18 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:18 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 11.031461204s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:18 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:18 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:28 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 10.670753107s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:29 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:29 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:35 +0000 UTC Normal PersistentVolumeClaim datadir-demand-backup-mysql-1 WaitForFirstConsumer waiting for first consumer to be created before binding logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:35 +0000 UTC Normal PersistentVolumeClaim datadir-demand-backup-mysql-1 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:35 +0000 UTC Normal PersistentVolumeClaim datadir-demand-backup-mysql-1 Provisioning External provisioner is provisioning volume for claim "kuttl-test-cosmic-ghoul/datadir-demand-backup-mysql-1" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:35 +0000 UTC Normal StatefulSet.apps demand-backup-mysql SuccessfulCreate create Claim datadir-demand-backup-mysql-1 Pod demand-backup-mysql-1 in StatefulSet demand-backup-mysql success logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:35 +0000 UTC Normal StatefulSet.apps demand-backup-mysql SuccessfulCreate create Pod demand-backup-mysql-1 in StatefulSet demand-backup-mysql successful logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:39 +0000 UTC Normal PersistentVolumeClaim datadir-demand-backup-mysql-1 ProvisioningSucceeded Successfully provisioned volume pvc-a0ec110f-caed-420b-acb1-a96733d2c38d logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:40 +0000 UTC Normal Pod demand-backup-mysql-1 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-mysql-1 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:41 +0000 UTC Normal Pod demand-backup-haproxy-0 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-haproxy-0 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:41 +0000 UTC Normal StatefulSet.apps demand-backup-haproxy SuccessfulCreate create Pod demand-backup-haproxy-0 in StatefulSet demand-backup-haproxy successful logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:41 +0000 UTC Normal Pod demand-backup-orc-2 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-orc-2 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w381 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:41 +0000 UTC Normal Pod demand-backup-orc-2.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:41 +0000 UTC Normal StatefulSet.apps demand-backup-orc SuccessfulCreate create Pod demand-backup-orc-2 in StatefulSet demand-backup-orc successful logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:42 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:42 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 383.233115ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:42 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:42 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:44 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:45 +0000 UTC Normal Pod demand-backup-mysql-1 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-a0ec110f-caed-420b-acb1-a96733d2c38d" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:47 +0000 UTC Normal Pod demand-backup-mysql-1.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:48 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 4.034827833s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:48 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Created Created container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:48 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Started Started container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:48 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:49 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 608.830883ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:49 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:49 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:49 +0000 UTC Normal Pod demand-backup-haproxy-1 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-haproxy-1 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:49 +0000 UTC Normal StatefulSet.apps demand-backup-haproxy SuccessfulCreate create Pod demand-backup-haproxy-1 in StatefulSet demand-backup-haproxy successful logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:49 +0000 UTC Normal Pod demand-backup-mysql-1.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 1.923481022s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:49 +0000 UTC Normal Pod demand-backup-mysql-1.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:49 +0000 UTC Normal Pod demand-backup-mysql-1.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:50 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:51 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 387.164459ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:51 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:51 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:51 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:53 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:56 +0000 UTC Normal Pod demand-backup-orc-2.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 14.577858745s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:56 +0000 UTC Normal Pod demand-backup-orc-2.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:56 +0000 UTC Normal Pod demand-backup-orc-2.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:57 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 4.0002469s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:57 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Created Created container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:57 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Started Started container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:57 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:57 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 378.641609ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:57 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:57 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:58 +0000 UTC Normal Pod demand-backup-haproxy-2 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-haproxy-2 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w381 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:08:58 +0000 UTC Normal StatefulSet.apps demand-backup-haproxy SuccessfulCreate create Pod demand-backup-haproxy-2 in StatefulSet demand-backup-haproxy successful logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:03 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:03 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 451.36565ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:03 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:03 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:03 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:05 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:06 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 3.298146301s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:06 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Created Created container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:06 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Started Started container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:06 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:10 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 5.194123957s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:10 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Created Created container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:10 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Started Started container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:10 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:11 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 734.914463ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:11 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:11 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:11 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 4.223890245s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:11 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:11 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:12 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 21.09020893s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:12 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Created Created container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:12 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Started Started container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:12 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:16 +0000 UTC Warning Pod demand-backup-orc-2.spec.containers{orc} Unhealthy Liveness probe failed: Get "http://10.167.209.6:3000/api/lb-check": dial tcp 10.167.209.6:3000: connect: connection refused logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:24 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 11.749221416s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:25 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:25 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:25 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:36 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 11.239595326s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:36 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:36 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:47 +0000 UTC Warning Pod demand-backup-mysql-1.spec.containers{mysql} Unhealthy Startup probe failed: logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:48 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 480.696047ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:48 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:09:51 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 392.010313ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:10:26 +0000 UTC Normal PersistentVolumeClaim datadir-demand-backup-mysql-2 WaitForFirstConsumer waiting for first consumer to be created before binding logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:10:26 +0000 UTC Normal PersistentVolumeClaim datadir-demand-backup-mysql-2 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:10:26 +0000 UTC Normal PersistentVolumeClaim datadir-demand-backup-mysql-2 Provisioning External provisioner is provisioning volume for claim "kuttl-test-cosmic-ghoul/datadir-demand-backup-mysql-2" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:10:26 +0000 UTC Normal StatefulSet.apps demand-backup-mysql SuccessfulCreate create Claim datadir-demand-backup-mysql-2 Pod demand-backup-mysql-2 in StatefulSet demand-backup-mysql success logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:10:26 +0000 UTC Normal StatefulSet.apps demand-backup-mysql SuccessfulCreate create Pod demand-backup-mysql-2 in StatefulSet demand-backup-mysql successful logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:10:30 +0000 UTC Normal PersistentVolumeClaim datadir-demand-backup-mysql-2 ProvisioningSucceeded Successfully provisioned volume pvc-5cf7e610-5753-4e89-bf05-e10eedb63bbe logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:10:30 +0000 UTC Normal Pod demand-backup-mysql-2 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-mysql-2 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w381 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:10:38 +0000 UTC Normal Pod demand-backup-mysql-2 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-5cf7e610-5753-4e89-bf05-e10eedb63bbe" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:10:39 +0000 UTC Normal Pod demand-backup-mysql-2.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:10:40 +0000 UTC Normal Pod demand-backup-mysql-2.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 395.146686ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:10:40 +0000 UTC Normal Pod demand-backup-mysql-2.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:10:40 +0000 UTC Normal Pod demand-backup-mysql-2.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:10:41 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:11:03 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 21.337304421s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:11:03 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Created Created container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:11:03 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Started Started container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:11:03 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:11:15 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 11.907202249s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:11:15 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:11:15 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:11:15 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:11:26 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 10.837112913s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:11:26 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:11:26 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:11:30 +0000 UTC Warning Pod demand-backup-mysql-2.spec.containers{mysql} Unhealthy Startup probe failed: logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:11:30 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:11:38 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 401.341296ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:12:42 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Killing Stopping container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:12:42 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Killing Stopping container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:12:42 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Killing Stopping container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:12:45 +0000 UTC Warning Pod demand-backup-mysql-0.spec.containers{mysql} Unhealthy Readiness probe failed: 2023/08/09 12:12:45 readiness check failed: connect to db: ping database: dial tcp 10.167.208.5:33062: connect: connection refused logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:12:50 +0000 UTC Warning Pod demand-backup-mysql-0.spec.containers{mysql} Unhealthy Readiness probe failed: 2023/08/09 12:12:50 readiness check failed: connect to db: ping database: dial tcp 10.167.208.5:33062: connect: connection refused logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:12:55 +0000 UTC Warning Pod demand-backup-mysql-0.spec.containers{mysql} Unhealthy Readiness probe errored: rpc error: code = Unknown desc = failed to exec in container: container is in CONTAINER_EXITED state logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:13:03 +0000 UTC Normal Pod demand-backup-mysql-0 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-mysql-0 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:13:07 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:13:08 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 371.397892ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:13:08 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:13:08 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:13:10 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:13:11 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 398.207594ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:13:11 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Created Created container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:13:11 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Started Started container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:13:11 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:13:11 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 395.068222ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:13:11 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:13:11 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:13:11 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:13:12 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 416.621925ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:13:12 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:13:12 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:13:28 +0000 UTC Warning Pod demand-backup-mysql-0.spec.containers{mysql} Unhealthy Startup probe failed: logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:13:28 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:13:32 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 404.169806ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:14:53 +0000 UTC Normal Pod xb-demand-backup-minio-minio-xcl82 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/xb-demand-backup-minio-minio-xcl82 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:14:53 +0000 UTC Normal Job.batch xb-demand-backup-minio-minio SuccessfulCreate Created pod: xb-demand-backup-minio-minio-xcl82 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:14:54 +0000 UTC Normal Pod xb-demand-backup-minio-minio-xcl82.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:14:55 +0000 UTC Normal Pod xb-demand-backup-minio-minio-xcl82.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 427.559126ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:14:55 +0000 UTC Normal Pod xb-demand-backup-minio-minio-xcl82.spec.initContainers{xtrabackup-init} Created Created container xtrabackup-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:14:55 +0000 UTC Normal Pod xb-demand-backup-minio-minio-xcl82.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:14:56 +0000 UTC Normal Pod xb-demand-backup-minio-minio-xcl82.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:14:57 +0000 UTC Normal Pod xb-demand-backup-minio-minio-xcl82.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 405.365135ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:14:57 +0000 UTC Normal Pod xb-demand-backup-minio-minio-xcl82.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:14:57 +0000 UTC Normal Pod xb-demand-backup-minio-minio-xcl82.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:15:04 +0000 UTC Normal Job.batch xb-demand-backup-minio-minio Completed Job completed logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:25 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Killing Stopping container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:25 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Killing Stopping container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:25 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Killing Stopping container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:25 +0000 UTC Normal StatefulSet.apps demand-backup-mysql SuccessfulDelete delete Pod demand-backup-mysql-2 in StatefulSet demand-backup-mysql successful logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:26 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Killing Stopping container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:26 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:26 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Killing Stopping container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:26 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:26 +0000 UTC Normal StatefulSet.apps demand-backup-haproxy SuccessfulDelete delete Pod demand-backup-haproxy-2 in StatefulSet demand-backup-haproxy successful logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:26 +0000 UTC Normal StatefulSet.apps demand-backup-haproxy SuccessfulDelete delete Pod demand-backup-haproxy-1 in StatefulSet demand-backup-haproxy successful logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:26 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Killing Stopping container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:26 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:26 +0000 UTC Normal StatefulSet.apps demand-backup-orc SuccessfulDelete delete Pod demand-backup-orc-2 in StatefulSet demand-backup-orc successful logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:27 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Killing Stopping container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:27 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:27 +0000 UTC Normal StatefulSet.apps demand-backup-haproxy SuccessfulDelete delete Pod demand-backup-haproxy-0 in StatefulSet demand-backup-haproxy successful logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:29 +0000 UTC Warning Pod demand-backup-mysql-2.spec.containers{mysql} Unhealthy Readiness probe failed: 2023/08/09 12:16:29 readiness check failed: connect to db: ping database: dial tcp 10.167.209.8:33062: connect: connection refused logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:46 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Killing Stopping container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:46 +0000 UTC Normal StatefulSet.apps demand-backup-mysql SuccessfulDelete delete Pod demand-backup-mysql-1 in StatefulSet demand-backup-mysql successful logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:50 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Killing Stopping container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:50 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Killing Stopping container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:50 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Killing Stopping container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:50 +0000 UTC Normal StatefulSet.apps demand-backup-mysql SuccessfulDelete delete Pod demand-backup-mysql-0 in StatefulSet demand-backup-mysql successful logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:57 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Killing Stopping container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:57 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:16:57 +0000 UTC Normal StatefulSet.apps demand-backup-orc SuccessfulDelete delete Pod demand-backup-orc-1 in StatefulSet demand-backup-orc successful logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:17:27 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Killing Stopping container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:17:27 +0000 UTC Normal StatefulSet.apps demand-backup-orc SuccessfulDelete delete Pod demand-backup-orc-0 in StatefulSet demand-backup-orc successful logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:17:28 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:17:35 +0000 UTC Warning Service demand-backup-haproxy ProcessServiceFailed error processing service "kuttl-test-cosmic-ghoul/demand-backup-haproxy": services "demand-backup-haproxy" not found logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:03 +0000 UTC Normal Pod xb-restore-demand-backup-restore-minio-mhmx8 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/xb-restore-demand-backup-restore-minio-mhmx8 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:03 +0000 UTC Normal Job.batch xb-restore-demand-backup-restore-minio SuccessfulCreate Created pod: xb-restore-demand-backup-restore-minio-mhmx8 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:15 +0000 UTC Normal Pod xb-restore-demand-backup-restore-minio-mhmx8 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-d0d9916e-94c2-4f55-b2d4-15e7447aeffb" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:16 +0000 UTC Normal Pod xb-restore-demand-backup-restore-minio-mhmx8.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:17 +0000 UTC Normal Pod xb-restore-demand-backup-restore-minio-mhmx8.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 449.037454ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:17 +0000 UTC Normal Pod xb-restore-demand-backup-restore-minio-mhmx8.spec.initContainers{xtrabackup-init} Created Created container xtrabackup-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:17 +0000 UTC Normal Pod xb-restore-demand-backup-restore-minio-mhmx8.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:19 +0000 UTC Normal Pod xb-restore-demand-backup-restore-minio-mhmx8.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:19 +0000 UTC Normal Pod xb-restore-demand-backup-restore-minio-mhmx8.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 397.832042ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:19 +0000 UTC Normal Pod xb-restore-demand-backup-restore-minio-mhmx8.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:19 +0000 UTC Normal Pod xb-restore-demand-backup-restore-minio-mhmx8.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:29 +0000 UTC Normal Job.batch xb-restore-demand-backup-restore-minio Completed Job completed logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:32 +0000 UTC Normal Pod demand-backup-mysql-0 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-mysql-0 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:33 +0000 UTC Normal Pod demand-backup-orc-0 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-orc-0 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:34 +0000 UTC Normal Pod demand-backup-orc-0.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:34 +0000 UTC Normal Pod demand-backup-orc-0.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 432.831156ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:34 +0000 UTC Normal Pod demand-backup-orc-0.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:34 +0000 UTC Normal Pod demand-backup-orc-0.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:36 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:37 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:37 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 416.122811ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:37 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Created Created container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:37 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Started Started container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:37 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:37 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 357.726907ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:37 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:37 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:38 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 413.385084ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:38 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:38 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:40 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:40 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 352.753325ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:40 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Created Created container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:40 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Started Started container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:40 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:41 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 383.714684ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:41 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:41 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:41 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:41 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 408.641887ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:41 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:18:41 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:08 +0000 UTC Normal Pod demand-backup-orc-1 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-orc-1 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:09 +0000 UTC Normal Pod demand-backup-orc-1.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:10 +0000 UTC Normal Pod demand-backup-orc-1.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 404.409242ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:10 +0000 UTC Normal Pod demand-backup-orc-1.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:10 +0000 UTC Normal Pod demand-backup-orc-1.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:12 +0000 UTC Normal Pod demand-backup-mysql-1 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-mysql-1 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:12 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:12 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 364.332609ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:12 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Created Created container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:13 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Started Started container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:13 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:13 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 360.083543ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:13 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:13 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:16 +0000 UTC Normal Pod demand-backup-mysql-1 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-a0ec110f-caed-420b-acb1-a96733d2c38d" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:17 +0000 UTC Normal Pod demand-backup-haproxy-0 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-haproxy-0 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:18 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:18 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 406.25756ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:18 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:18 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:18 +0000 UTC Normal Pod demand-backup-mysql-1.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:18 +0000 UTC Normal Pod demand-backup-mysql-1.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 380.400465ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:18 +0000 UTC Normal Pod demand-backup-mysql-1.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:18 +0000 UTC Normal Pod demand-backup-mysql-1.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:20 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:20 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:21 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 394.836332ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:21 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Created Created container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:21 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Started Started container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:21 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:21 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 371.47483ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:21 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:21 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:21 +0000 UTC Normal Pod demand-backup-haproxy-1 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-haproxy-1 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w381 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:21 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 360.593468ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:21 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Created Created container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:21 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Started Started container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:21 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:21 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 351.038671ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:21 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:21 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:21 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:22 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:22 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 403.315178ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:22 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:22 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 403.61558ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:22 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:22 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:23 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:25 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:25 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 408.623258ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:25 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Created Created container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:25 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Started Started container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:25 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:26 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 369.1611ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:26 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:26 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:26 +0000 UTC Normal Pod demand-backup-haproxy-2 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-haproxy-2 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:27 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:27 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 354.279923ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:27 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:27 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:29 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:30 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 352.259412ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:30 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Created Created container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:30 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Started Started container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:30 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:30 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 366.215526ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:30 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:30 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:39 +0000 UTC Warning Pod demand-backup-mysql-1.spec.containers{mysql} Unhealthy Startup probe failed: logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:39 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:42 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 416.269629ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:44 +0000 UTC Normal Pod demand-backup-orc-2 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-orc-2 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w381 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:45 +0000 UTC Normal Pod demand-backup-orc-2.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:45 +0000 UTC Normal Pod demand-backup-orc-2.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 372.745695ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:45 +0000 UTC Normal Pod demand-backup-orc-2.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:45 +0000 UTC Normal Pod demand-backup-orc-2.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:47 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:47 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 365.32268ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:47 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Created Created container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:47 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Started Started container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:47 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:48 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 371.166173ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:48 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:19:48 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:20:18 +0000 UTC Normal Pod demand-backup-mysql-2 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-mysql-2 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w381 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:20:23 +0000 UTC Normal Pod demand-backup-mysql-2 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-5cf7e610-5753-4e89-bf05-e10eedb63bbe" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:20:28 +0000 UTC Normal Pod demand-backup-mysql-2.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:20:28 +0000 UTC Normal Pod demand-backup-mysql-2.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 404.275936ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:20:28 +0000 UTC Normal Pod demand-backup-mysql-2.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:20:28 +0000 UTC Normal Pod demand-backup-mysql-2.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:20:30 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:20:30 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 357.628607ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:20:30 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Created Created container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:20:30 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Started Started container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:20:30 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:20:31 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 367.732415ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:20:31 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:20:31 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:20:31 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:20:31 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 367.972656ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:20:31 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:20:31 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:20:48 +0000 UTC Warning Pod demand-backup-mysql-2.spec.containers{mysql} Unhealthy Startup probe failed: logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:20:48 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:20:52 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 363.741591ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:23:08 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Killing Stopping container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:23:09 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Killing Stopping container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:23:09 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:23:09 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Killing Stopping container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:23:09 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Killing Stopping container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:23:09 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Killing Stopping container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:23:09 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:23:10 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Killing Stopping container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:23:10 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:23:11 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Killing Stopping container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:23:11 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:23:13 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Killing Stopping container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:23:13 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Killing Stopping container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:23:13 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Killing Stopping container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:23:17 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Killing Stopping container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:23:17 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Killing Stopping container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:23:17 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Killing Stopping container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:23:22 +0000 UTC Warning Pod demand-backup-mysql-0.spec.containers{mysql} Unhealthy Readiness probe errored: rpc error: code = Unknown desc = failed to exec in container: container is in CONTAINER_EXITED state logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:23:24 +0000 UTC Warning Service demand-backup-haproxy ProcessServiceFailed error processing service "kuttl-test-cosmic-ghoul/demand-backup-haproxy": services "demand-backup-haproxy" not found logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:23:40 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Killing Stopping container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:23:40 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:24:08 +0000 UTC Warning Service demand-backup-haproxy ProcessServiceFailed error processing service "kuttl-test-cosmic-ghoul/demand-backup-haproxy": services "demand-backup-haproxy" not found logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:24:11 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Killing Stopping container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:24:11 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:24:44 +0000 UTC Normal Pod xb-restore-demand-backup-restore-minio-backup-source-r6vz9 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/xb-restore-demand-backup-restore-minio-backup-source-r6vz9 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:24:44 +0000 UTC Normal Job.batch xb-restore-demand-backup-restore-minio-backup-source SuccessfulCreate Created pod: xb-restore-demand-backup-restore-minio-backup-source-r6vz9 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:24:51 +0000 UTC Normal Pod xb-restore-demand-backup-restore-minio-backup-source-r6vz9 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-d0d9916e-94c2-4f55-b2d4-15e7447aeffb" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:24:52 +0000 UTC Normal Pod xb-restore-demand-backup-restore-minio-backup-source-r6vz9.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:24:53 +0000 UTC Normal Pod xb-restore-demand-backup-restore-minio-backup-source-r6vz9.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 409.710011ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:24:53 +0000 UTC Normal Pod xb-restore-demand-backup-restore-minio-backup-source-r6vz9.spec.initContainers{xtrabackup-init} Created Created container xtrabackup-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:24:53 +0000 UTC Normal Pod xb-restore-demand-backup-restore-minio-backup-source-r6vz9.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:24:55 +0000 UTC Normal Pod xb-restore-demand-backup-restore-minio-backup-source-r6vz9.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:24:56 +0000 UTC Normal Pod xb-restore-demand-backup-restore-minio-backup-source-r6vz9.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 381.081574ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:24:56 +0000 UTC Normal Pod xb-restore-demand-backup-restore-minio-backup-source-r6vz9.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:24:56 +0000 UTC Normal Pod xb-restore-demand-backup-restore-minio-backup-source-r6vz9.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:05 +0000 UTC Normal Job.batch xb-restore-demand-backup-restore-minio-backup-source Completed Job completed logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:08 +0000 UTC Normal Pod demand-backup-mysql-0 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-mysql-0 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:09 +0000 UTC Normal Pod demand-backup-orc-0 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-orc-0 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:09 +0000 UTC Normal Pod demand-backup-orc-0.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:10 +0000 UTC Normal Pod demand-backup-orc-0.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 402.473213ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:10 +0000 UTC Normal Pod demand-backup-orc-0.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:10 +0000 UTC Normal Pod demand-backup-orc-0.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:11 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:12 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 397.674191ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:12 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Created Created container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:12 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Started Started container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:12 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:12 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 365.072326ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:12 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:12 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:13 +0000 UTC Normal Pod demand-backup-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-d0d9916e-94c2-4f55-b2d4-15e7447aeffb" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:16 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:16 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 414.105719ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:16 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:16 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:18 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:19 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 362.829108ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:19 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Created Created container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:19 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Started Started container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:19 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:19 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 372.847201ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:19 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:19 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:19 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:20 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 389.033993ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:20 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:20 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:44 +0000 UTC Normal Pod demand-backup-orc-1 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-orc-1 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:45 +0000 UTC Normal Pod demand-backup-orc-1.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:45 +0000 UTC Normal Pod demand-backup-orc-1.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 418.294536ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:45 +0000 UTC Normal Pod demand-backup-orc-1.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:45 +0000 UTC Normal Pod demand-backup-orc-1.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:47 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:48 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 416.527975ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:48 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Created Created container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:48 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Started Started container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:48 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:48 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 370.724698ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:48 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:48 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:51 +0000 UTC Normal Pod demand-backup-mysql-1 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-mysql-1 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:52 +0000 UTC Normal Pod demand-backup-haproxy-0 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-haproxy-0 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:52 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:53 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 355.053672ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:53 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:53 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:54 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:55 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 378.008676ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:55 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Created Created container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:55 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Started Started container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:55 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:55 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 373.82002ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:55 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:55 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:55 +0000 UTC Normal Pod demand-backup-haproxy-1 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-haproxy-1 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:55 +0000 UTC Normal Pod demand-backup-mysql-1 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-a0ec110f-caed-420b-acb1-a96733d2c38d" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:56 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:56 +0000 UTC Normal Pod demand-backup-mysql-1.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:57 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 394.181536ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:57 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:57 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:57 +0000 UTC Normal Pod demand-backup-mysql-1.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 609.150951ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:57 +0000 UTC Normal Pod demand-backup-mysql-1.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:57 +0000 UTC Normal Pod demand-backup-mysql-1.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:59 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:25:59 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:00 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 375.615131ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:00 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Created Created container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:00 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Started Started container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:00 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:00 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 730.734396ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:00 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Created Created container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:00 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Started Started container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:00 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:01 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 640.061066ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:01 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:01 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:01 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 682.497717ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:01 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:01 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:01 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:01 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 371.631368ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:01 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:02 +0000 UTC Normal Pod demand-backup-haproxy-2 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-haproxy-2 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w381 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:02 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:02 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:03 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 461.637652ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:03 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:03 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:05 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:05 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 446.472454ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:05 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Created Created container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:05 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Started Started container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:05 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:06 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 377.811493ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:06 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:06 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:17 +0000 UTC Warning Pod demand-backup-mysql-1.spec.containers{mysql} Unhealthy Startup probe failed: logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:17 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:19 +0000 UTC Normal Pod demand-backup-orc-2 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-orc-2 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w381 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:20 +0000 UTC Normal Pod demand-backup-orc-2.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:21 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 397.074966ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:21 +0000 UTC Normal Pod demand-backup-orc-2.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 370.436819ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:21 +0000 UTC Normal Pod demand-backup-orc-2.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:21 +0000 UTC Normal Pod demand-backup-orc-2.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:22 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 386.440872ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:23 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:23 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 380.55817ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:23 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Created Created container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:23 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Started Started container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:24 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:24 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 364.901462ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:24 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:24 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:51 +0000 UTC Normal Pod demand-backup-mysql-2 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-mysql-2 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w381 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:55 +0000 UTC Normal Pod demand-backup-mysql-2 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-5cf7e610-5753-4e89-bf05-e10eedb63bbe" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:57 +0000 UTC Normal Pod demand-backup-mysql-2.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:57 +0000 UTC Normal Pod demand-backup-mysql-2.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 365.221557ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:57 +0000 UTC Normal Pod demand-backup-mysql-2.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:57 +0000 UTC Normal Pod demand-backup-mysql-2.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:26:59 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:27:00 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 402.985075ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:27:00 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Created Created container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:27:00 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Started Started container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:27:00 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:27:00 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 379.412406ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:27:00 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:27:00 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:27:00 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:27:01 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 404.545049ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:27:01 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:27:01 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:27:18 +0000 UTC Warning Pod demand-backup-mysql-2.spec.containers{mysql} Unhealthy Startup probe failed: logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:27:18 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:27:21 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 440.539798ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:28:20 +0000 UTC Normal Pod xb-demand-backup-s3-aws-s3-8vstg Scheduled Successfully assigned kuttl-test-cosmic-ghoul/xb-demand-backup-s3-aws-s3-8vstg to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:28:20 +0000 UTC Normal Job.batch xb-demand-backup-s3-aws-s3 SuccessfulCreate Created pod: xb-demand-backup-s3-aws-s3-8vstg logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:28:21 +0000 UTC Normal Pod xb-demand-backup-s3-aws-s3-8vstg.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:28:21 +0000 UTC Normal Pod xb-demand-backup-s3-aws-s3-8vstg.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 405.213535ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:28:21 +0000 UTC Normal Pod xb-demand-backup-s3-aws-s3-8vstg.spec.initContainers{xtrabackup-init} Created Created container xtrabackup-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:28:21 +0000 UTC Normal Pod xb-demand-backup-s3-aws-s3-8vstg.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:28:24 +0000 UTC Normal Pod xb-demand-backup-s3-aws-s3-8vstg.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:28:24 +0000 UTC Normal Pod xb-demand-backup-s3-aws-s3-8vstg.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 354.692881ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:28:24 +0000 UTC Normal Pod xb-demand-backup-s3-aws-s3-8vstg.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:28:24 +0000 UTC Normal Pod xb-demand-backup-s3-aws-s3-8vstg.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:28:32 +0000 UTC Normal Job.batch xb-demand-backup-s3-aws-s3 Completed Job completed logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:29:02 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Killing Stopping container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:29:03 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Killing Stopping container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:29:03 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:29:03 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Killing Stopping container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:29:03 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Killing Stopping container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:29:03 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Killing Stopping container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:29:03 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:29:04 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Killing Stopping container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:29:04 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:29:05 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Killing Stopping container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:29:05 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:29:09 +0000 UTC Warning Service demand-backup-haproxy ProcessServiceFailed error processing service "kuttl-test-cosmic-ghoul/demand-backup-haproxy": services "demand-backup-haproxy" not found logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:29:11 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Killing Stopping container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:29:11 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Killing Stopping container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:29:11 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Killing Stopping container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:29:15 +0000 UTC Warning Pod demand-backup-mysql-0.spec.containers{mysql} Unhealthy Readiness probe errored: rpc error: code = Unknown desc = failed to exec in container: container is in CONTAINER_EXITED state logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:29:34 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Killing Stopping container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:29:34 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:29:43 +0000 UTC Warning Service demand-backup-haproxy ProcessServiceFailed error processing service "kuttl-test-cosmic-ghoul/demand-backup-haproxy": services "demand-backup-haproxy" not found logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:30:05 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Killing Stopping container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:30:05 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:30:36 +0000 UTC Normal Pod xb-restore-demand-backup-restore-s3-wq5bg Scheduled Successfully assigned kuttl-test-cosmic-ghoul/xb-restore-demand-backup-restore-s3-wq5bg to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:30:36 +0000 UTC Normal Job.batch xb-restore-demand-backup-restore-s3 SuccessfulCreate Created pod: xb-restore-demand-backup-restore-s3-wq5bg logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:30:42 +0000 UTC Normal Pod xb-restore-demand-backup-restore-s3-wq5bg SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-d0d9916e-94c2-4f55-b2d4-15e7447aeffb" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:30:43 +0000 UTC Normal Pod xb-restore-demand-backup-restore-s3-wq5bg.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:30:44 +0000 UTC Normal Pod xb-restore-demand-backup-restore-s3-wq5bg.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 429.226077ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:30:44 +0000 UTC Normal Pod xb-restore-demand-backup-restore-s3-wq5bg.spec.initContainers{xtrabackup-init} Created Created container xtrabackup-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:30:44 +0000 UTC Normal Pod xb-restore-demand-backup-restore-s3-wq5bg.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:30:45 +0000 UTC Normal Pod xb-restore-demand-backup-restore-s3-wq5bg.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:30:46 +0000 UTC Normal Pod xb-restore-demand-backup-restore-s3-wq5bg.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 387.640194ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:30:46 +0000 UTC Normal Pod xb-restore-demand-backup-restore-s3-wq5bg.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:30:46 +0000 UTC Normal Pod xb-restore-demand-backup-restore-s3-wq5bg.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:30:58 +0000 UTC Normal Job.batch xb-restore-demand-backup-restore-s3 Completed Job completed logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:02 +0000 UTC Normal Pod demand-backup-mysql-0 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-mysql-0 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:02 +0000 UTC Normal Pod demand-backup-orc-0 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-orc-0 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:03 +0000 UTC Warning Pod demand-backup-orc-0 FailedMount MountVolume.SetUp failed for volume "config" : failed to sync configmap cache: timed out waiting for the condition logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:03 +0000 UTC Warning Pod demand-backup-orc-0 FailedMount MountVolume.SetUp failed for volume "users" : failed to sync secret cache: timed out waiting for the condition logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:03 +0000 UTC Warning Pod demand-backup-orc-0 FailedMount MountVolume.SetUp failed for volume "tls" : failed to sync secret cache: timed out waiting for the condition logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:04 +0000 UTC Normal Pod demand-backup-orc-0.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:05 +0000 UTC Normal Pod demand-backup-orc-0.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 410.784688ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:05 +0000 UTC Normal Pod demand-backup-orc-0.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:05 +0000 UTC Normal Pod demand-backup-orc-0.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:06 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:07 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 412.374335ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:07 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Created Created container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:07 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Started Started container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:07 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:07 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 373.224888ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:07 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:07 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:11 +0000 UTC Normal Pod demand-backup-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-d0d9916e-94c2-4f55-b2d4-15e7447aeffb" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:13 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:13 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 395.211714ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:13 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:14 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:15 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:16 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 407.064014ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:16 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Created Created container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:16 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Started Started container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:16 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:16 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 352.549747ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:16 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:16 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:16 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:17 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 385.705855ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:17 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:17 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:39 +0000 UTC Normal Pod demand-backup-orc-1 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-orc-1 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:40 +0000 UTC Normal Pod demand-backup-orc-1.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:40 +0000 UTC Normal Pod demand-backup-orc-1.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 432.069179ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:40 +0000 UTC Normal Pod demand-backup-orc-1.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:40 +0000 UTC Normal Pod demand-backup-orc-1.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:42 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:43 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 369.995666ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:43 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Created Created container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:43 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Started Started container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:43 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:43 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 405.723043ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:43 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:43 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:48 +0000 UTC Normal Pod demand-backup-mysql-1 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-mysql-1 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:51 +0000 UTC Normal Pod demand-backup-haproxy-0 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-haproxy-0 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:52 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:52 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 350.966362ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:52 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:52 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:52 +0000 UTC Normal Pod demand-backup-mysql-1 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-a0ec110f-caed-420b-acb1-a96733d2c38d" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:54 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:54 +0000 UTC Normal Pod demand-backup-mysql-1.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:54 +0000 UTC Normal Pod demand-backup-mysql-1.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 397.358059ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:54 +0000 UTC Normal Pod demand-backup-mysql-1.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:54 +0000 UTC Normal Pod demand-backup-mysql-1.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:55 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 390.646739ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:55 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Created Created container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:55 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Started Started container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:55 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:55 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 376.525445ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:55 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:55 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:56 +0000 UTC Normal Pod demand-backup-haproxy-1 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-haproxy-1 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:56 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:57 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 379.155305ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:57 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:57 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:57 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:57 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 613.684426ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:57 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Created Created container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:58 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Started Started container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:58 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:58 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 738.011271ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:58 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:58 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:58 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:59 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:59 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 645.212172ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:59 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:31:59 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:00 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 941.85046ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:00 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Created Created container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:00 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Started Started container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:00 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:00 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 369.771532ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:00 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:00 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:01 +0000 UTC Normal Pod demand-backup-haproxy-2 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-haproxy-2 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w381 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:01 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:02 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 426.470271ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:02 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:02 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:04 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:04 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 358.594842ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:04 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Created Created container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:04 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Started Started container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:04 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:05 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 440.168203ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:05 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:05 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:14 +0000 UTC Warning Pod demand-backup-mysql-1.spec.containers{mysql} Unhealthy Startup probe failed: logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:14 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:14 +0000 UTC Normal Pod demand-backup-orc-2 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-orc-2 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w381 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:15 +0000 UTC Normal Pod demand-backup-orc-2.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:15 +0000 UTC Normal Pod demand-backup-orc-2.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 364.135276ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:15 +0000 UTC Normal Pod demand-backup-orc-2.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:15 +0000 UTC Normal Pod demand-backup-orc-2.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:17 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:17 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 373.461696ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:17 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Created Created container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:17 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Started Started container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:17 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:18 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 373.903409ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:18 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 852.880277ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:18 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:18 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:48 +0000 UTC Normal Pod demand-backup-mysql-2 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-mysql-2 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w381 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:53 +0000 UTC Normal Pod demand-backup-mysql-2 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-5cf7e610-5753-4e89-bf05-e10eedb63bbe" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:54 +0000 UTC Normal Pod demand-backup-mysql-2.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:55 +0000 UTC Normal Pod demand-backup-mysql-2.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 424.540996ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:55 +0000 UTC Normal Pod demand-backup-mysql-2.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:55 +0000 UTC Normal Pod demand-backup-mysql-2.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:57 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:57 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 355.717322ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:57 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Created Created container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:58 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Started Started container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:58 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:58 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 407.281811ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:58 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:58 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:58 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:58 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 371.64954ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:58 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:32:59 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:33:15 +0000 UTC Warning Pod demand-backup-mysql-2.spec.containers{mysql} Unhealthy Startup probe failed: logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:33:15 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:33:19 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 363.314951ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:34:13 +0000 UTC Normal Pod xb-demand-backup-gcp-gcp-cs-l66kh Scheduled Successfully assigned kuttl-test-cosmic-ghoul/xb-demand-backup-gcp-gcp-cs-l66kh to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:34:13 +0000 UTC Normal Job.batch xb-demand-backup-gcp-gcp-cs SuccessfulCreate Created pod: xb-demand-backup-gcp-gcp-cs-l66kh logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:34:14 +0000 UTC Normal Pod xb-demand-backup-gcp-gcp-cs-l66kh.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:34:14 +0000 UTC Normal Pod xb-demand-backup-gcp-gcp-cs-l66kh.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 372.229981ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:34:15 +0000 UTC Normal Pod xb-demand-backup-gcp-gcp-cs-l66kh.spec.initContainers{xtrabackup-init} Created Created container xtrabackup-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:34:15 +0000 UTC Normal Pod xb-demand-backup-gcp-gcp-cs-l66kh.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:34:17 +0000 UTC Normal Pod xb-demand-backup-gcp-gcp-cs-l66kh.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:34:17 +0000 UTC Normal Pod xb-demand-backup-gcp-gcp-cs-l66kh.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 410.617148ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:34:17 +0000 UTC Normal Pod xb-demand-backup-gcp-gcp-cs-l66kh.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:34:17 +0000 UTC Normal Pod xb-demand-backup-gcp-gcp-cs-l66kh.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:34:24 +0000 UTC Normal Job.batch xb-demand-backup-gcp-gcp-cs Completed Job completed logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:34:59 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Killing Stopping container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:34:59 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:34:59 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Killing Stopping container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:34:59 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Killing Stopping container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:34:59 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Killing Stopping container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:34:59 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Killing Stopping container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:34:59 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:35:00 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Killing Stopping container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:35:00 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:35:01 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Killing Stopping container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:35:01 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:35:03 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Killing Stopping container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:35:03 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Killing Stopping container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:35:03 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Killing Stopping container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:35:07 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Killing Stopping container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:35:07 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Killing Stopping container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:35:07 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Killing Stopping container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:35:08 +0000 UTC Warning Pod demand-backup-mysql-0.spec.containers{mysql} Unhealthy Readiness probe failed: 2023/08/09 12:35:08 readiness check failed: connect to db: ping database: dial tcp 10.167.210.24:33062: connect: connection refused logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:35:13 +0000 UTC Warning Pod demand-backup-mysql-0.spec.containers{mysql} Unhealthy Readiness probe errored: rpc error: code = Unknown desc = failed to exec in container: container is in CONTAINER_EXITED state logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:35:30 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Killing Stopping container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:35:30 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:36:00 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Killing Stopping container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:36:00 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:36:09 +0000 UTC Warning Service demand-backup-haproxy ProcessServiceFailed error processing service "kuttl-test-cosmic-ghoul/demand-backup-haproxy": services "demand-backup-haproxy" not found logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:36:23 +0000 UTC Warning Service demand-backup-haproxy ProcessServiceFailed error processing service "kuttl-test-cosmic-ghoul/demand-backup-haproxy": services "demand-backup-haproxy" not found logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:36:35 +0000 UTC Normal Pod xb-restore-demand-backup-restore-gcp-4ck6k Scheduled Successfully assigned kuttl-test-cosmic-ghoul/xb-restore-demand-backup-restore-gcp-4ck6k to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:36:35 +0000 UTC Normal Job.batch xb-restore-demand-backup-restore-gcp SuccessfulCreate Created pod: xb-restore-demand-backup-restore-gcp-4ck6k logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:36:39 +0000 UTC Normal Pod xb-restore-demand-backup-restore-gcp-4ck6k SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-d0d9916e-94c2-4f55-b2d4-15e7447aeffb" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:36:41 +0000 UTC Normal Pod xb-restore-demand-backup-restore-gcp-4ck6k.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:36:41 +0000 UTC Normal Pod xb-restore-demand-backup-restore-gcp-4ck6k.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 506.590725ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:36:41 +0000 UTC Normal Pod xb-restore-demand-backup-restore-gcp-4ck6k.spec.initContainers{xtrabackup-init} Created Created container xtrabackup-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:36:41 +0000 UTC Normal Pod xb-restore-demand-backup-restore-gcp-4ck6k.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:36:43 +0000 UTC Normal Pod xb-restore-demand-backup-restore-gcp-4ck6k.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:36:44 +0000 UTC Normal Pod xb-restore-demand-backup-restore-gcp-4ck6k.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 353.344187ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:36:44 +0000 UTC Normal Pod xb-restore-demand-backup-restore-gcp-4ck6k.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:36:44 +0000 UTC Normal Pod xb-restore-demand-backup-restore-gcp-4ck6k.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:36:56 +0000 UTC Normal Job.batch xb-restore-demand-backup-restore-gcp Completed Job completed logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:08 +0000 UTC Normal Pod demand-backup-mysql-0 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-mysql-0 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:08 +0000 UTC Normal Pod demand-backup-orc-0 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-orc-0 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:09 +0000 UTC Normal Pod demand-backup-orc-0.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:09 +0000 UTC Normal Pod demand-backup-orc-0.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 363.596936ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:09 +0000 UTC Normal Pod demand-backup-orc-0.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:09 +0000 UTC Normal Pod demand-backup-orc-0.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:11 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:12 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 601.426066ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:12 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Created Created container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:12 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Started Started container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:12 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:13 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 364.919939ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:13 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:13 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:15 +0000 UTC Normal Pod demand-backup-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-d0d9916e-94c2-4f55-b2d4-15e7447aeffb" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:17 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:17 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 365.377484ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:17 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:17 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:20 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:20 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 388.016903ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:20 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Created Created container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:20 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Started Started container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:20 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:20 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 368.348127ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:20 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:20 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:20 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:21 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 403.459069ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:21 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:21 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:44 +0000 UTC Normal Pod demand-backup-orc-1 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-orc-1 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:44 +0000 UTC Normal Pod demand-backup-orc-1.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:45 +0000 UTC Normal Pod demand-backup-orc-1.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 361.205298ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:45 +0000 UTC Normal Pod demand-backup-orc-1.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:45 +0000 UTC Normal Pod demand-backup-orc-1.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:47 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:47 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 395.278468ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:47 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Created Created container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:47 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Started Started container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:47 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:47 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 403.424296ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:48 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:48 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:52 +0000 UTC Normal Pod demand-backup-mysql-1 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-mysql-1 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:53 +0000 UTC Normal Pod demand-backup-haproxy-0 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-haproxy-0 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:53 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:54 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 388.479931ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:54 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:54 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:56 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:56 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 374.115397ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:56 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Created Created container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:56 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Started Started container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:56 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:56 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 374.480557ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:57 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:57 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:57 +0000 UTC Normal Pod demand-backup-haproxy-1 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-haproxy-1 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w381 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:57 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:58 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 432.262928ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:58 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:58 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:37:59 +0000 UTC Normal Pod demand-backup-mysql-1 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-a0ec110f-caed-420b-acb1-a96733d2c38d" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:00 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:00 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 369.605066ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:00 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Created Created container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:00 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Started Started container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:00 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:01 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 380.17393ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:01 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:01 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:01 +0000 UTC Normal Pod demand-backup-haproxy-2 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-haproxy-2 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:01 +0000 UTC Normal Pod demand-backup-mysql-1.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:01 +0000 UTC Normal Pod demand-backup-mysql-1.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 537.733756ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:01 +0000 UTC Normal Pod demand-backup-mysql-1.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:01 +0000 UTC Normal Pod demand-backup-mysql-1.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:02 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:03 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 1.046985793s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:03 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:03 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:04 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:04 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 844.450697ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:04 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Created Created container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:05 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:05 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 713.431183ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:05 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Created Created container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:05 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Started Started container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:05 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:05 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Started Started container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:05 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:05 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 370.993947ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:05 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:05 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:05 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:06 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 724.98498ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:06 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:06 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:06 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 687.143835ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:06 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:06 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:19 +0000 UTC Normal Pod demand-backup-orc-2 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-orc-2 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w381 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:21 +0000 UTC Warning Pod demand-backup-mysql-1.spec.containers{mysql} Unhealthy Startup probe failed: logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:21 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:21 +0000 UTC Normal Pod demand-backup-orc-2.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:21 +0000 UTC Normal Pod demand-backup-orc-2.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 364.176882ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:21 +0000 UTC Normal Pod demand-backup-orc-2.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:21 +0000 UTC Normal Pod demand-backup-orc-2.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:23 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:23 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 360.903172ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:23 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Created Created container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:23 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Started Started container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:23 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:24 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 371.525271ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:24 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:24 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:25 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 383.914487ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:38:55 +0000 UTC Normal Pod demand-backup-mysql-2 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-mysql-2 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w381 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:39:00 +0000 UTC Normal Pod demand-backup-mysql-2 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-5cf7e610-5753-4e89-bf05-e10eedb63bbe" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:39:02 +0000 UTC Normal Pod demand-backup-mysql-2.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:39:02 +0000 UTC Normal Pod demand-backup-mysql-2.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 652.648718ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:39:02 +0000 UTC Normal Pod demand-backup-mysql-2.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:39:02 +0000 UTC Normal Pod demand-backup-mysql-2.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:39:04 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:39:04 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 375.528504ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:39:04 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Created Created container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:39:05 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Started Started container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:39:05 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:39:05 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 366.787132ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:39:05 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:39:05 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:39:05 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:39:05 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 403.246907ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:39:05 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:39:06 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:39:22 +0000 UTC Warning Pod demand-backup-mysql-2.spec.containers{mysql} Unhealthy Startup probe failed: logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:39:22 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:39:26 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 365.75322ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:40:24 +0000 UTC Normal Pod xb-demand-backup-azure-azure-blob-wc9rn Scheduled Successfully assigned kuttl-test-cosmic-ghoul/xb-demand-backup-azure-azure-blob-wc9rn to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:40:24 +0000 UTC Normal Job.batch xb-demand-backup-azure-azure-blob SuccessfulCreate Created pod: xb-demand-backup-azure-azure-blob-wc9rn logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:40:25 +0000 UTC Normal Pod xb-demand-backup-azure-azure-blob-wc9rn.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:40:25 +0000 UTC Normal Pod xb-demand-backup-azure-azure-blob-wc9rn.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 525.388207ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:40:25 +0000 UTC Normal Pod xb-demand-backup-azure-azure-blob-wc9rn.spec.initContainers{xtrabackup-init} Created Created container xtrabackup-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:40:25 +0000 UTC Normal Pod xb-demand-backup-azure-azure-blob-wc9rn.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:40:27 +0000 UTC Normal Pod xb-demand-backup-azure-azure-blob-wc9rn.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:40:27 +0000 UTC Normal Pod xb-demand-backup-azure-azure-blob-wc9rn.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 373.979936ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:40:28 +0000 UTC Normal Pod xb-demand-backup-azure-azure-blob-wc9rn.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:40:28 +0000 UTC Normal Pod xb-demand-backup-azure-azure-blob-wc9rn.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:40:36 +0000 UTC Normal Job.batch xb-demand-backup-azure-azure-blob Completed Job completed logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:41:09 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Killing Stopping container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:41:09 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Killing Stopping container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:41:09 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Killing Stopping container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:41:09 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:41:10 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Killing Stopping container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:41:10 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:41:10 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Killing Stopping container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:41:10 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:41:10 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Killing Stopping container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:41:12 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Killing Stopping container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:41:12 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:41:14 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Killing Stopping container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:41:14 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Killing Stopping container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:41:14 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Killing Stopping container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:41:18 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Killing Stopping container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:41:18 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Killing Stopping container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:41:19 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Killing Stopping container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:41:21 +0000 UTC Warning Pod demand-backup-mysql-0.spec.containers{mysql} Unhealthy Readiness probe failed: 2023/08/09 12:41:21 readiness check failed: connect to db: ping database: dial tcp 10.167.210.29:33062: connect: connection refused logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:41:26 +0000 UTC Warning Pod demand-backup-mysql-0.spec.containers{mysql} Unhealthy Readiness probe errored: rpc error: code = Unknown desc = failed to exec in container: container is in CONTAINER_EXITED state logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:41:40 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Killing Stopping container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:41:40 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:42:11 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Killing Stopping container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:42:11 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:42:46 +0000 UTC Normal Pod xb-restore-demand-backup-restore-azure-xgkkk Scheduled Successfully assigned kuttl-test-cosmic-ghoul/xb-restore-demand-backup-restore-azure-xgkkk to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:42:46 +0000 UTC Normal Job.batch xb-restore-demand-backup-restore-azure SuccessfulCreate Created pod: xb-restore-demand-backup-restore-azure-xgkkk logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:42:50 +0000 UTC Normal Pod xb-restore-demand-backup-restore-azure-xgkkk SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-d0d9916e-94c2-4f55-b2d4-15e7447aeffb" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:42:51 +0000 UTC Normal Pod xb-restore-demand-backup-restore-azure-xgkkk.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:42:52 +0000 UTC Normal Pod xb-restore-demand-backup-restore-azure-xgkkk.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 417.218654ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:42:52 +0000 UTC Normal Pod xb-restore-demand-backup-restore-azure-xgkkk.spec.initContainers{xtrabackup-init} Created Created container xtrabackup-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:42:52 +0000 UTC Normal Pod xb-restore-demand-backup-restore-azure-xgkkk.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:42:54 +0000 UTC Normal Pod xb-restore-demand-backup-restore-azure-xgkkk.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:42:54 +0000 UTC Normal Pod xb-restore-demand-backup-restore-azure-xgkkk.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 393.537275ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:42:54 +0000 UTC Normal Pod xb-restore-demand-backup-restore-azure-xgkkk.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:42:54 +0000 UTC Normal Pod xb-restore-demand-backup-restore-azure-xgkkk.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:06 +0000 UTC Normal Job.batch xb-restore-demand-backup-restore-azure Completed Job completed logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:11 +0000 UTC Normal Pod demand-backup-mysql-0 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-mysql-0 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:11 +0000 UTC Normal Pod demand-backup-orc-0 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-orc-0 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:12 +0000 UTC Normal Pod demand-backup-orc-0.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:12 +0000 UTC Normal Pod demand-backup-orc-0.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 424.739354ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:12 +0000 UTC Normal Pod demand-backup-orc-0.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:12 +0000 UTC Normal Pod demand-backup-orc-0.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:14 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:15 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 481.513897ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:15 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Created Created container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:15 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{orc} Started Started container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:15 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:15 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 420.27266ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:16 +0000 UTC Normal Pod demand-backup-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-d0d9916e-94c2-4f55-b2d4-15e7447aeffb" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:16 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:16 +0000 UTC Normal Pod demand-backup-orc-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:19 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:19 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 394.758857ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:19 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:19 +0000 UTC Normal Pod demand-backup-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:22 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:22 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 401.923621ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:22 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Created Created container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:22 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{mysql} Started Started container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:22 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:23 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 390.629215ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:23 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:23 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:23 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:23 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 429.802099ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:23 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:23 +0000 UTC Normal Pod demand-backup-mysql-0.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:46 +0000 UTC Normal Pod demand-backup-orc-1 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-orc-1 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:47 +0000 UTC Normal Pod demand-backup-orc-1.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:47 +0000 UTC Normal Pod demand-backup-orc-1.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 401.080462ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:47 +0000 UTC Normal Pod demand-backup-orc-1.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:48 +0000 UTC Normal Pod demand-backup-orc-1.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:50 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:50 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 443.732309ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:50 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Created Created container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:50 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{orc} Started Started container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:50 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:51 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 379.785043ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:51 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:51 +0000 UTC Normal Pod demand-backup-orc-1.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:54 +0000 UTC Normal Pod demand-backup-mysql-1 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-mysql-1 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:58 +0000 UTC Normal Pod demand-backup-haproxy-0 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-haproxy-0 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:58 +0000 UTC Normal Pod demand-backup-mysql-1 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-a0ec110f-caed-420b-acb1-a96733d2c38d" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:59 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:59 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 376.53248ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:59 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:43:59 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:00 +0000 UTC Normal Pod demand-backup-mysql-1.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:00 +0000 UTC Normal Pod demand-backup-mysql-1.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 452.70353ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:00 +0000 UTC Normal Pod demand-backup-mysql-1.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:00 +0000 UTC Normal Pod demand-backup-mysql-1.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:02 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:02 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 592.190516ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:02 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:02 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 540.416571ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:02 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Created Created container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:02 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Started Started container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:02 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:03 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Created Created container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:03 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{haproxy} Started Started container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:03 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:03 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 416.507594ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:03 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:03 +0000 UTC Normal Pod demand-backup-haproxy-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:03 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 402.116367ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:03 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:03 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:03 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:03 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 375.054581ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:03 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:03 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:04 +0000 UTC Normal Pod demand-backup-haproxy-1 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-haproxy-1 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w381 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:05 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:05 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 394.556978ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:05 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:05 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:07 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:07 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 409.328999ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:07 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Created Created container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:07 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{haproxy} Started Started container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:07 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:08 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 388.127904ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:08 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:08 +0000 UTC Normal Pod demand-backup-haproxy-1.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:08 +0000 UTC Normal Pod demand-backup-haproxy-2 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-haproxy-2 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:09 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:09 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 400.043029ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:09 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:09 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:12 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:13 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 859.398415ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:13 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Created Created container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:13 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{haproxy} Started Started container haproxy logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:13 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:13 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 391.145204ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:13 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:13 +0000 UTC Normal Pod demand-backup-haproxy-2.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:20 +0000 UTC Warning Pod demand-backup-mysql-1.spec.containers{mysql} Unhealthy Startup probe failed: logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:20 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:22 +0000 UTC Normal Pod demand-backup-orc-2 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-orc-2 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w381 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:22 +0000 UTC Normal Pod demand-backup-orc-2.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:23 +0000 UTC Normal Pod demand-backup-orc-2.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 372.107164ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:23 +0000 UTC Normal Pod demand-backup-orc-2.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:23 +0000 UTC Normal Pod demand-backup-orc-2.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:24 +0000 UTC Normal Pod demand-backup-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 385.389317ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:25 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:25 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 380.940111ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:25 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Created Created container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:26 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{orc} Started Started container orc logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:26 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:26 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 422.192256ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:26 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:26 +0000 UTC Normal Pod demand-backup-orc-2.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:44:54 +0000 UTC Normal Pod demand-backup-mysql-2 Scheduled Successfully assigned kuttl-test-cosmic-ghoul/demand-backup-mysql-2 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w381 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:45:03 +0000 UTC Normal Pod demand-backup-mysql-2 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-5cf7e610-5753-4e89-bf05-e10eedb63bbe" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:45:04 +0000 UTC Normal Pod demand-backup-mysql-2.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:45:06 +0000 UTC Normal Pod demand-backup-mysql-2.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 1.239293234s logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:45:06 +0000 UTC Normal Pod demand-backup-mysql-2.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:45:06 +0000 UTC Normal Pod demand-backup-mysql-2.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:45:07 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:45:08 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 398.15705ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:45:08 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Created Created container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:45:08 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Started Started container mysql logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:45:08 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:45:08 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 393.409912ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:45:08 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:45:08 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:45:08 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:45:09 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 474.605713ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:45:09 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:45:09 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:45:25 +0000 UTC Warning Pod demand-backup-mysql-2.spec.containers{mysql} Unhealthy Startup probe failed: logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:45:25 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:45:29 +0000 UTC Normal Pod demand-backup-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 530.019935ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:46:35 +0000 UTC Normal Pod aws-cli Scheduled Successfully assigned kuttl-test-cosmic-ghoul/aws-cli to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:46:36 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Pulling Pulling image "perconalab/awscli" logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:46:36 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Pulled Successfully pulled image "perconalab/awscli" in 364.881574ms logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:46:36 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Created Created container aws-cli logger.go:42: 12:46:42 | demand-backup | 2023-08-09 12:46:36 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Started Started container aws-cli logger.go:42: 12:46:43 | demand-backup | Deleting namespace: kuttl-test-cosmic-ghoul === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (2505.05s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/demand-backup (2495.02s) PASS