=== RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://34.136.43.62 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 180 seconds for each step harness.go:372: testsuite: e2e-tests/tests has 28 tests === RUN kuttl/harness === RUN kuttl/harness/one-pod === PAUSE kuttl/harness/one-pod === CONT kuttl/harness/one-pod logger.go:42: 14:54:17 | one-pod | Creating namespace: kuttl-test-guided-fish logger.go:42: 14:54:18 | one-pod/0-deploy-operator | starting test step 0-deploy-operator logger.go:42: 14:54:18 | one-pod/0-deploy-operator | running command: [sh -c set -o errexit set -o xtrace source ../../functions apply_s3_storage_secrets deploy_operator deploy_non_tls_cluster_secrets deploy_tls_cluster_secrets deploy_client deploy_minio] logger.go:42: 14:54:18 | one-pod/0-deploy-operator | + source ../../functions logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ realpath ../../.. logger.go:42: 14:54:18 | one-pod/0-deploy-operator | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:54:18 | one-pod/0-deploy-operator | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/vars.sh logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:54:18 | one-pod/0-deploy-operator | ++++ mktemp -d logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ export TEMP_DIR=/tmp/tmp.C0dqkUoLaf logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ TEMP_DIR=/tmp/tmp.C0dqkUoLaf logger.go:42: 14:54:18 | one-pod/0-deploy-operator | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ export GIT_BRANCH=PR-424 logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ GIT_BRANCH=PR-424 logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ export VERSION=PR-424-70568ae logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ VERSION=PR-424-70568ae logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:54:18 | one-pod/0-deploy-operator | ++++ which gdate logger.go:42: 14:54:18 | one-pod/0-deploy-operator | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-424/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:54:18 | one-pod/0-deploy-operator | ++++ which date logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ date=/usr/bin/date logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ command -v oc logger.go:42: 14:54:18 | one-pod/0-deploy-operator | +++ oc get projects logger.go:42: 14:54:23 | one-pod/0-deploy-operator | error: the server doesn't have a resource type "projects" logger.go:42: 14:54:23 | one-pod/0-deploy-operator | +++ kubectl get nodes logger.go:42: 14:54:23 | one-pod/0-deploy-operator | +++ grep '^minikube' logger.go:42: 14:54:24 | one-pod/0-deploy-operator | ++++ pwd logger.go:42: 14:54:24 | one-pod/0-deploy-operator | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/tests/one-pod logger.go:42: 14:54:24 | one-pod/0-deploy-operator | ++ test_name=one-pod logger.go:42: 14:54:24 | one-pod/0-deploy-operator | + apply_s3_storage_secrets logger.go:42: 14:54:24 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-guided-fish apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf/minio-secret.yml logger.go:42: 14:54:25 | one-pod/0-deploy-operator | secret/minio-secret created logger.go:42: 14:54:25 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-guided-fish apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf/cloud-secret.yml logger.go:42: 14:54:26 | one-pod/0-deploy-operator | secret/aws-s3-secret created logger.go:42: 14:54:27 | one-pod/0-deploy-operator | secret/gcp-cs-secret created logger.go:42: 14:54:27 | one-pod/0-deploy-operator | secret/azure-secret created logger.go:42: 14:54:27 | one-pod/0-deploy-operator | + deploy_operator logger.go:42: 14:54:27 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-guided-fish apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy/crd.yaml logger.go:42: 14:54:28 | one-pod/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlbackups.ps.percona.com serverside-applied logger.go:42: 14:54:28 | one-pod/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlrestores.ps.percona.com serverside-applied logger.go:42: 14:54:29 | one-pod/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqls.ps.percona.com serverside-applied logger.go:42: 14:54:29 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-guided-fish apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy/rbac.yaml logger.go:42: 14:54:31 | one-pod/0-deploy-operator | serviceaccount/percona-server-mysql-operator created logger.go:42: 14:54:31 | one-pod/0-deploy-operator | serviceaccount/percona-server-mysql-operator-orchestrator created logger.go:42: 14:54:31 | one-pod/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 14:54:32 | one-pod/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 14:54:32 | one-pod/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 14:54:32 | one-pod/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 14:54:33 | one-pod/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 14:54:33 | one-pod/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 14:54:33 | one-pod/0-deploy-operator | + yq eval '(select(documentIndex==1).spec.template.spec.containers[] | select(.name=="manager").env[] | select(.name=="DISABLE_TELEMETRY").value) = "true"' logger.go:42: 14:54:33 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-guided-fish apply -f - logger.go:42: 14:54:33 | one-pod/0-deploy-operator | ++ printf 'select(documentIndex==1).spec.template.spec.containers[0].image="%s"' perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:54:33 | one-pod/0-deploy-operator | + yq eval 'select(documentIndex==1).spec.template.spec.containers[0].image="perconalab/percona-server-mysql-operator:PR-424-70568ae"' /mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy/operator.yaml logger.go:42: 14:54:34 | one-pod/0-deploy-operator | configmap/percona-server-mysql-operator-config created logger.go:42: 14:54:35 | one-pod/0-deploy-operator | deployment.apps/percona-server-mysql-operator created logger.go:42: 14:54:35 | one-pod/0-deploy-operator | + deploy_non_tls_cluster_secrets logger.go:42: 14:54:35 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-guided-fish apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf/secrets.yaml logger.go:42: 14:54:36 | one-pod/0-deploy-operator | secret/test-secrets created logger.go:42: 14:54:36 | one-pod/0-deploy-operator | + deploy_tls_cluster_secrets logger.go:42: 14:54:36 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-guided-fish apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf/ssl-secret.yaml logger.go:42: 14:54:37 | one-pod/0-deploy-operator | secret/test-ssl created logger.go:42: 14:54:37 | one-pod/0-deploy-operator | + deploy_client logger.go:42: 14:54:37 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-guided-fish apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf/client.yaml logger.go:42: 14:54:38 | one-pod/0-deploy-operator | pod/mysql-client created logger.go:42: 14:54:38 | one-pod/0-deploy-operator | + deploy_minio logger.go:42: 14:54:38 | one-pod/0-deploy-operator | ++ kubectl -n kuttl-test-guided-fish get secret minio-secret -o 'jsonpath={.data.AWS_ACCESS_KEY_ID}' logger.go:42: 14:54:38 | one-pod/0-deploy-operator | ++ base64 -d logger.go:42: 14:54:39 | one-pod/0-deploy-operator | + local accessKey=some-access-key logger.go:42: 14:54:39 | one-pod/0-deploy-operator | ++ kubectl -n kuttl-test-guided-fish get secret minio-secret -o 'jsonpath={.data.AWS_SECRET_ACCESS_KEY}' logger.go:42: 14:54:39 | one-pod/0-deploy-operator | ++ base64 -d logger.go:42: 14:54:40 | one-pod/0-deploy-operator | + local secretKey=some-secret-key logger.go:42: 14:54:40 | one-pod/0-deploy-operator | + helm uninstall -n kuttl-test-guided-fish minio-service logger.go:42: 14:54:40 | one-pod/0-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-424/kubeconfig logger.go:42: 14:54:40 | one-pod/0-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-424/kubeconfig logger.go:42: 14:54:40 | one-pod/0-deploy-operator | Error: uninstall: Release not loaded: minio-service: release: not found logger.go:42: 14:54:40 | one-pod/0-deploy-operator | + : logger.go:42: 14:54:40 | one-pod/0-deploy-operator | + helm repo remove minio logger.go:42: 14:54:40 | one-pod/0-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-424/kubeconfig logger.go:42: 14:54:40 | one-pod/0-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-424/kubeconfig logger.go:42: 14:54:40 | one-pod/0-deploy-operator | Error: no repositories configured logger.go:42: 14:54:40 | one-pod/0-deploy-operator | + : logger.go:42: 14:54:40 | one-pod/0-deploy-operator | + helm repo add minio https://helm.min.io/ logger.go:42: 14:54:40 | one-pod/0-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-424/kubeconfig logger.go:42: 14:54:40 | one-pod/0-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-424/kubeconfig logger.go:42: 14:54:40 | one-pod/0-deploy-operator | "minio" has been added to your repositories logger.go:42: 14:54:40 | one-pod/0-deploy-operator | ++ printf %q some-access-key logger.go:42: 14:54:40 | one-pod/0-deploy-operator | ++ printf %q some-secret-key logger.go:42: 14:54:40 | one-pod/0-deploy-operator | + retry 10 60 helm install minio-service -n kuttl-test-guided-fish --version 8.0.5 --set accessKey=some-access-key --set secretKey=some-secret-key --set service.type=ClusterIP --set configPathmc=/tmp/.minio/ --set persistence.size=2G --set environment.MINIO_REGION=us-east-1 --set environment.MINIO_HTTP_TRACE=/tmp/trace.log --set securityContext.enabled=false minio/minio logger.go:42: 14:54:40 | one-pod/0-deploy-operator | + local max=10 logger.go:42: 14:54:40 | one-pod/0-deploy-operator | + local delay=60 logger.go:42: 14:54:40 | one-pod/0-deploy-operator | + shift 2 logger.go:42: 14:54:40 | one-pod/0-deploy-operator | + local n=1 logger.go:42: 14:54:40 | one-pod/0-deploy-operator | + helm install minio-service -n kuttl-test-guided-fish --version 8.0.5 --set accessKey=some-access-key --set secretKey=some-secret-key --set service.type=ClusterIP --set configPathmc=/tmp/.minio/ --set persistence.size=2G --set environment.MINIO_REGION=us-east-1 --set environment.MINIO_HTTP_TRACE=/tmp/trace.log --set securityContext.enabled=false minio/minio logger.go:42: 14:54:40 | one-pod/0-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-424/kubeconfig logger.go:42: 14:54:40 | one-pod/0-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-424/kubeconfig logger.go:42: 14:54:45 | one-pod/0-deploy-operator | NAME: minio-service logger.go:42: 14:54:45 | one-pod/0-deploy-operator | LAST DEPLOYED: Thu Aug 17 14:54:42 2023 logger.go:42: 14:54:45 | one-pod/0-deploy-operator | NAMESPACE: kuttl-test-guided-fish logger.go:42: 14:54:45 | one-pod/0-deploy-operator | STATUS: deployed logger.go:42: 14:54:45 | one-pod/0-deploy-operator | REVISION: 1 logger.go:42: 14:54:45 | one-pod/0-deploy-operator | TEST SUITE: None logger.go:42: 14:54:45 | one-pod/0-deploy-operator | NOTES: logger.go:42: 14:54:45 | one-pod/0-deploy-operator | Minio can be accessed via port 9000 on the following DNS name from within your cluster: logger.go:42: 14:54:45 | one-pod/0-deploy-operator | minio-service.kuttl-test-guided-fish.svc.cluster.local logger.go:42: 14:54:45 | one-pod/0-deploy-operator | logger.go:42: 14:54:45 | one-pod/0-deploy-operator | To access Minio from localhost, run the below commands: logger.go:42: 14:54:45 | one-pod/0-deploy-operator | logger.go:42: 14:54:45 | one-pod/0-deploy-operator | 1. export POD_NAME=$(kubectl get pods --namespace kuttl-test-guided-fish -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}") logger.go:42: 14:54:45 | one-pod/0-deploy-operator | logger.go:42: 14:54:45 | one-pod/0-deploy-operator | 2. kubectl port-forward $POD_NAME 9000 --namespace kuttl-test-guided-fish logger.go:42: 14:54:45 | one-pod/0-deploy-operator | logger.go:42: 14:54:45 | one-pod/0-deploy-operator | Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/ logger.go:42: 14:54:45 | one-pod/0-deploy-operator | logger.go:42: 14:54:45 | one-pod/0-deploy-operator | You can now access Minio server on http://localhost:9000. Follow the below steps to connect to Minio server with mc client: logger.go:42: 14:54:45 | one-pod/0-deploy-operator | logger.go:42: 14:54:45 | one-pod/0-deploy-operator | 1. Download the Minio mc client - https://docs.minio.io/docs/minio-client-quickstart-guide logger.go:42: 14:54:45 | one-pod/0-deploy-operator | logger.go:42: 14:54:45 | one-pod/0-deploy-operator | 2. Get the ACCESS_KEY=$(kubectl get secret minio-service -o jsonpath="{.data.accesskey}" | base64 --decode) and the SECRET_KEY=$(kubectl get secret minio-service -o jsonpath="{.data.secretkey}" | base64 --decode) logger.go:42: 14:54:45 | one-pod/0-deploy-operator | logger.go:42: 14:54:45 | one-pod/0-deploy-operator | 3. mc alias set minio-service-local http://localhost:9000 "$ACCESS_KEY" "$SECRET_KEY" --api s3v4 logger.go:42: 14:54:45 | one-pod/0-deploy-operator | logger.go:42: 14:54:45 | one-pod/0-deploy-operator | 4. mc ls minio-service-local logger.go:42: 14:54:45 | one-pod/0-deploy-operator | logger.go:42: 14:54:45 | one-pod/0-deploy-operator | Alternately, you can use your browser or the Minio SDK to access the server - https://docs.minio.io/categories/17 logger.go:42: 14:54:45 | one-pod/0-deploy-operator | ++ kubectl -n kuttl-test-guided-fish get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:54:46 | one-pod/0-deploy-operator | + MINIO_POD=minio-service-5fc4dc898-bv7xb logger.go:42: 14:54:46 | one-pod/0-deploy-operator | + wait_pod minio-service-5fc4dc898-bv7xb logger.go:42: 14:54:46 | one-pod/0-deploy-operator | + local pod=minio-service-5fc4dc898-bv7xb logger.go:42: 14:54:46 | one-pod/0-deploy-operator | + set +o xtrace logger.go:42: 14:55:02 | one-pod/0-deploy-operator | minio-service-5fc4dc898-bv7xb.........true logger.go:42: 14:55:02 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-guided-fish run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- bash -c 'AWS_ACCESS_KEY_ID='\''some-access-key'\'' AWS_SECRET_ACCESS_KEY='\''some-secret-key'\'' AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url http://minio-service:9000 s3 mb s3://operator-testing' logger.go:42: 14:55:08 | one-pod/0-deploy-operator | If you don't see a command prompt, try pressing enter. logger.go:42: 14:55:12 | one-pod/0-deploy-operator | pod "aws-cli" deleted logger.go:42: 14:55:14 | one-pod/0-deploy-operator | test step completed 0-deploy-operator logger.go:42: 14:55:14 | one-pod/1-create-cluster | starting test step 1-create-cluster logger.go:42: 14:55:14 | one-pod/1-create-cluster | running command: [sh -c set -o errexit set -o xtrace source ../../functions get_cr \ | yq eval '.spec.mysql.clusterType="async"' - \ | yq eval '.spec.allowUnsafeConfigurations=true' - \ | yq eval '.spec.mysql.size=1' - \ | yq eval '.spec.proxy.haproxy.enabled=true' - \ | yq eval '.spec.proxy.haproxy.size=1' - \ | yq eval '.spec.orchestrator.enabled=true' - \ | yq eval '.spec.orchestrator.size=1' - \ | yq eval '.spec.backup.storages.minio.type="s3"' - \ | yq eval '.spec.backup.storages.minio.s3.bucket="operator-testing"' - \ | yq eval '.spec.backup.storages.minio.s3.credentialsSecret="minio-secret"' - \ | yq eval '.spec.backup.storages.minio.s3.endpointUrl="http://minio-service:9000"' - \ | yq eval '.spec.backup.storages.minio.s3.region="us-east-1"' - \ | kubectl -n "${NAMESPACE}" apply -f -] logger.go:42: 14:55:14 | one-pod/1-create-cluster | + source ../../functions logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ realpath ../../.. logger.go:42: 14:55:14 | one-pod/1-create-cluster | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:55:14 | one-pod/1-create-cluster | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/vars.sh logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:55:14 | one-pod/1-create-cluster | ++++ mktemp -d logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ export TEMP_DIR=/tmp/tmp.8K52RAWQG3 logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ TEMP_DIR=/tmp/tmp.8K52RAWQG3 logger.go:42: 14:55:14 | one-pod/1-create-cluster | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ export GIT_BRANCH=PR-424 logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ GIT_BRANCH=PR-424 logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ export VERSION=PR-424-70568ae logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ VERSION=PR-424-70568ae logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:55:14 | one-pod/1-create-cluster | ++++ which gdate logger.go:42: 14:55:14 | one-pod/1-create-cluster | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-424/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:55:14 | one-pod/1-create-cluster | ++++ which date logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ date=/usr/bin/date logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ command -v oc logger.go:42: 14:55:14 | one-pod/1-create-cluster | +++ oc get projects logger.go:42: 14:55:20 | one-pod/1-create-cluster | error: the server doesn't have a resource type "projects" logger.go:42: 14:55:20 | one-pod/1-create-cluster | +++ grep '^minikube' logger.go:42: 14:55:20 | one-pod/1-create-cluster | +++ kubectl get nodes logger.go:42: 14:55:20 | one-pod/1-create-cluster | ++++ pwd logger.go:42: 14:55:20 | one-pod/1-create-cluster | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/tests/one-pod logger.go:42: 14:55:20 | one-pod/1-create-cluster | ++ test_name=one-pod logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval .spec.allowUnsafeConfigurations=true - logger.go:42: 14:55:20 | one-pod/1-create-cluster | + get_cr logger.go:42: 14:55:20 | one-pod/1-create-cluster | + local name_suffix= logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval .spec.mysql.size=1 - logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval .spec.proxy.haproxy.enabled=true - logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval .spec.proxy.haproxy.size=1 - logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval .spec.orchestrator.enabled=true - logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval '.spec.backup.storages.minio.s3.region="us-east-1"' - logger.go:42: 14:55:20 | one-pod/1-create-cluster | + kubectl -n kuttl-test-guided-fish apply -f - logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval '.spec.backup.storages.minio.s3.endpointUrl="http://minio-service:9000"' - logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval '.spec.backup.storages.minio.s3.credentialsSecret="minio-secret"' - logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval '.spec.upgradeOptions.apply="disabled"' - logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval '.spec.sslSecretName="test-ssl"' - logger.go:42: 14:55:20 | one-pod/1-create-cluster | ++ printf '.spec.mysql.image="%s"' perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval '.spec.mysql.image="perconalab/percona-server-mysql-operator:main-psmysql"' - logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval '.spec.backup.storages.minio.s3.bucket="operator-testing"' - logger.go:42: 14:55:20 | one-pod/1-create-cluster | + '[' -n '' ']' logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval - logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval '.spec.secretsName="test-secrets"' - logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval '.spec.backup.storages.minio.type="s3"' - logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval .spec.orchestrator.size=1 - logger.go:42: 14:55:20 | one-pod/1-create-cluster | ++ printf '.metadata.name="%s"' one-pod logger.go:42: 14:55:20 | one-pod/1-create-cluster | ++ printf '.spec.backup.image="%s"' perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval '.metadata.name="one-pod"' /mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy/cr.yaml logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval '.spec.backup.image="perconalab/percona-server-mysql-operator:main-backup"' - logger.go:42: 14:55:20 | one-pod/1-create-cluster | ++ printf '.spec.orchestrator.image="%s"' perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval '.spec.orchestrator.image="perconalab/percona-server-mysql-operator:main-orchestrator"' - logger.go:42: 14:55:20 | one-pod/1-create-cluster | ++ printf '.spec.proxy.haproxy.image="%s"' perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval '.spec.proxy.haproxy.image="perconalab/percona-server-mysql-operator:main-haproxy"' - logger.go:42: 14:55:20 | one-pod/1-create-cluster | ++ printf '.spec.initImage="%s"' perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval '.spec.initImage="perconalab/percona-server-mysql-operator:PR-424-70568ae"' - logger.go:42: 14:55:20 | one-pod/1-create-cluster | ++ printf '.spec.pmm.image="%s"' perconalab/pmm-client:dev-latest logger.go:42: 14:55:20 | one-pod/1-create-cluster | ++ printf '.spec.proxy.router.image="%s"' perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval '.spec.proxy.router.image="perconalab/percona-server-mysql-operator:main-router"' - logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval '.spec.pmm.image="perconalab/pmm-client:dev-latest"' - logger.go:42: 14:55:20 | one-pod/1-create-cluster | ++ printf '.spec.toolkit.image="%s"' perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:55:20 | one-pod/1-create-cluster | + yq eval '.spec.toolkit.image="perconalab/percona-server-mysql-operator:main-toolkit"' - logger.go:42: 14:55:22 | one-pod/1-create-cluster | perconaservermysql.ps.percona.com/one-pod created logger.go:42: 14:56:24 | one-pod/1-create-cluster | test step completed 1-create-cluster logger.go:42: 14:56:24 | one-pod/2-write-data | starting test step 2-write-data logger.go:42: 14:56:24 | one-pod/2-write-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password" run_mysql \ "INSERT myDB.myTable (id) VALUES (100500)" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password"] logger.go:42: 14:56:24 | one-pod/2-write-data | + source ../../functions logger.go:42: 14:56:24 | one-pod/2-write-data | +++ realpath ../../.. logger.go:42: 14:56:24 | one-pod/2-write-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:56:24 | one-pod/2-write-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/vars.sh logger.go:42: 14:56:24 | one-pod/2-write-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:56:24 | one-pod/2-write-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:56:24 | one-pod/2-write-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:56:24 | one-pod/2-write-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:56:24 | one-pod/2-write-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:56:24 | one-pod/2-write-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:56:24 | one-pod/2-write-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:56:24 | one-pod/2-write-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:56:24 | one-pod/2-write-data | ++++ mktemp -d logger.go:42: 14:56:24 | one-pod/2-write-data | +++ export TEMP_DIR=/tmp/tmp.oaTeSDbspw logger.go:42: 14:56:24 | one-pod/2-write-data | +++ TEMP_DIR=/tmp/tmp.oaTeSDbspw logger.go:42: 14:56:24 | one-pod/2-write-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:56:24 | one-pod/2-write-data | +++ export GIT_BRANCH=PR-424 logger.go:42: 14:56:24 | one-pod/2-write-data | +++ GIT_BRANCH=PR-424 logger.go:42: 14:56:24 | one-pod/2-write-data | +++ export VERSION=PR-424-70568ae logger.go:42: 14:56:24 | one-pod/2-write-data | +++ VERSION=PR-424-70568ae logger.go:42: 14:56:24 | one-pod/2-write-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:56:24 | one-pod/2-write-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:56:24 | one-pod/2-write-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:56:24 | one-pod/2-write-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:56:24 | one-pod/2-write-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:56:24 | one-pod/2-write-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:56:24 | one-pod/2-write-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:56:24 | one-pod/2-write-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:56:24 | one-pod/2-write-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:56:24 | one-pod/2-write-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:56:24 | one-pod/2-write-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:56:24 | one-pod/2-write-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:56:24 | one-pod/2-write-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:56:24 | one-pod/2-write-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:56:24 | one-pod/2-write-data | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:56:24 | one-pod/2-write-data | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:56:24 | one-pod/2-write-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:56:24 | one-pod/2-write-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:56:24 | one-pod/2-write-data | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:56:24 | one-pod/2-write-data | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:56:24 | one-pod/2-write-data | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:56:24 | one-pod/2-write-data | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:56:24 | one-pod/2-write-data | ++++ which gdate logger.go:42: 14:56:24 | one-pod/2-write-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-424/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:56:24 | one-pod/2-write-data | ++++ which date logger.go:42: 14:56:24 | one-pod/2-write-data | +++ date=/usr/bin/date logger.go:42: 14:56:24 | one-pod/2-write-data | +++ command -v oc logger.go:42: 14:56:24 | one-pod/2-write-data | +++ oc get projects logger.go:42: 14:56:29 | one-pod/2-write-data | error: the server doesn't have a resource type "projects" logger.go:42: 14:56:29 | one-pod/2-write-data | +++ kubectl get nodes logger.go:42: 14:56:29 | one-pod/2-write-data | +++ grep '^minikube' logger.go:42: 14:56:30 | one-pod/2-write-data | ++++ pwd logger.go:42: 14:56:30 | one-pod/2-write-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/tests/one-pod logger.go:42: 14:56:30 | one-pod/2-write-data | ++ test_name=one-pod logger.go:42: 14:56:30 | one-pod/2-write-data | +++ get_cluster_name logger.go:42: 14:56:30 | one-pod/2-write-data | +++ kubectl -n kuttl-test-guided-fish get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:56:31 | one-pod/2-write-data | ++ get_haproxy_svc one-pod logger.go:42: 14:56:31 | one-pod/2-write-data | ++ local cluster=one-pod logger.go:42: 14:56:31 | one-pod/2-write-data | ++ echo one-pod-haproxy logger.go:42: 14:56:31 | one-pod/2-write-data | + run_mysql 'CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' '-h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:56:31 | one-pod/2-write-data | + local 'command=CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' logger.go:42: 14:56:31 | one-pod/2-write-data | + local 'uri=-h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:56:31 | one-pod/2-write-data | + local pod= logger.go:42: 14:56:31 | one-pod/2-write-data | ++ get_client_pod logger.go:42: 14:56:31 | one-pod/2-write-data | ++ kubectl -n kuttl-test-guided-fish get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:56:31 | one-pod/2-write-data | + client_pod=mysql-client logger.go:42: 14:56:31 | one-pod/2-write-data | + wait_pod mysql-client logger.go:42: 14:56:31 | one-pod/2-write-data | + local pod=mysql-client logger.go:42: 14:56:31 | one-pod/2-write-data | + set +o xtrace logger.go:42: 14:56:32 | one-pod/2-write-data | mysql-clienttrue logger.go:42: 14:56:32 | one-pod/2-write-data | + sed -e 's/mysql: //' logger.go:42: 14:56:32 | one-pod/2-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:56:32 | one-pod/2-write-data | + kubectl -n kuttl-test-guided-fish exec mysql-client -- bash -c 'printf '\''%s\n'\'' "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" | mysql -sN -h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:56:34 | one-pod/2-write-data | + : logger.go:42: 14:56:34 | one-pod/2-write-data | +++ get_cluster_name logger.go:42: 14:56:34 | one-pod/2-write-data | +++ kubectl -n kuttl-test-guided-fish get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:56:34 | one-pod/2-write-data | ++ get_haproxy_svc one-pod logger.go:42: 14:56:34 | one-pod/2-write-data | ++ local cluster=one-pod logger.go:42: 14:56:34 | one-pod/2-write-data | ++ echo one-pod-haproxy logger.go:42: 14:56:34 | one-pod/2-write-data | + run_mysql 'INSERT myDB.myTable (id) VALUES (100500)' '-h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:56:34 | one-pod/2-write-data | + local 'command=INSERT myDB.myTable (id) VALUES (100500)' logger.go:42: 14:56:34 | one-pod/2-write-data | + local 'uri=-h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:56:34 | one-pod/2-write-data | + local pod= logger.go:42: 14:56:34 | one-pod/2-write-data | ++ get_client_pod logger.go:42: 14:56:34 | one-pod/2-write-data | ++ kubectl -n kuttl-test-guided-fish get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:56:35 | one-pod/2-write-data | + client_pod=mysql-client logger.go:42: 14:56:35 | one-pod/2-write-data | + wait_pod mysql-client logger.go:42: 14:56:35 | one-pod/2-write-data | + local pod=mysql-client logger.go:42: 14:56:35 | one-pod/2-write-data | + set +o xtrace logger.go:42: 14:56:36 | one-pod/2-write-data | mysql-clienttrue logger.go:42: 14:56:36 | one-pod/2-write-data | + sed -e 's/mysql: //' logger.go:42: 14:56:36 | one-pod/2-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:56:36 | one-pod/2-write-data | + kubectl -n kuttl-test-guided-fish exec mysql-client -- bash -c 'printf '\''%s\n'\'' "INSERT myDB.myTable (id) VALUES (100500)" | mysql -sN -h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:56:37 | one-pod/2-write-data | + : logger.go:42: 14:56:38 | one-pod/2-write-data | test step completed 2-write-data logger.go:42: 14:56:38 | one-pod/3-create-backup-minio | starting test step 3-create-backup-minio logger.go:42: 14:56:40 | one-pod/3-create-backup-minio | PerconaServerMySQLBackup:kuttl-test-guided-fish/one-pod-minio created logger.go:42: 14:56:56 | one-pod/3-create-backup-minio | test step completed 3-create-backup-minio logger.go:42: 14:56:56 | one-pod/4-delete-data | starting test step 4-delete-data logger.go:42: 14:56:56 | one-pod/4-delete-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "TRUNCATE TABLE myDB.myTable" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password" data=$(run_mysql "SELECT * FROM myDB.myTable" "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 04-delete-data-minio --from-literal=data="${data}"] logger.go:42: 14:56:56 | one-pod/4-delete-data | + source ../../functions logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ realpath ../../.. logger.go:42: 14:56:56 | one-pod/4-delete-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:56:56 | one-pod/4-delete-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/vars.sh logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:56:56 | one-pod/4-delete-data | ++++ mktemp -d logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ export TEMP_DIR=/tmp/tmp.jKyY09HSAn logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ TEMP_DIR=/tmp/tmp.jKyY09HSAn logger.go:42: 14:56:56 | one-pod/4-delete-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ export GIT_BRANCH=PR-424 logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ GIT_BRANCH=PR-424 logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ export VERSION=PR-424-70568ae logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ VERSION=PR-424-70568ae logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:56:56 | one-pod/4-delete-data | ++++ which gdate logger.go:42: 14:56:56 | one-pod/4-delete-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-424/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:56:56 | one-pod/4-delete-data | ++++ which date logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ date=/usr/bin/date logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ command -v oc logger.go:42: 14:56:56 | one-pod/4-delete-data | +++ oc get projects logger.go:42: 14:57:01 | one-pod/4-delete-data | error: the server doesn't have a resource type "projects" logger.go:42: 14:57:01 | one-pod/4-delete-data | +++ kubectl get nodes logger.go:42: 14:57:01 | one-pod/4-delete-data | +++ grep '^minikube' logger.go:42: 14:57:02 | one-pod/4-delete-data | ++++ pwd logger.go:42: 14:57:02 | one-pod/4-delete-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/tests/one-pod logger.go:42: 14:57:02 | one-pod/4-delete-data | ++ test_name=one-pod logger.go:42: 14:57:02 | one-pod/4-delete-data | +++ get_cluster_name logger.go:42: 14:57:02 | one-pod/4-delete-data | +++ kubectl -n kuttl-test-guided-fish get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:57:03 | one-pod/4-delete-data | ++ get_haproxy_svc one-pod logger.go:42: 14:57:03 | one-pod/4-delete-data | ++ local cluster=one-pod logger.go:42: 14:57:03 | one-pod/4-delete-data | ++ echo one-pod-haproxy logger.go:42: 14:57:03 | one-pod/4-delete-data | + run_mysql 'TRUNCATE TABLE myDB.myTable' '-h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:57:03 | one-pod/4-delete-data | + local 'command=TRUNCATE TABLE myDB.myTable' logger.go:42: 14:57:03 | one-pod/4-delete-data | + local 'uri=-h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:57:03 | one-pod/4-delete-data | + local pod= logger.go:42: 14:57:03 | one-pod/4-delete-data | ++ get_client_pod logger.go:42: 14:57:03 | one-pod/4-delete-data | ++ kubectl -n kuttl-test-guided-fish get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:57:03 | one-pod/4-delete-data | + client_pod=mysql-client logger.go:42: 14:57:03 | one-pod/4-delete-data | + wait_pod mysql-client logger.go:42: 14:57:03 | one-pod/4-delete-data | + local pod=mysql-client logger.go:42: 14:57:03 | one-pod/4-delete-data | + set +o xtrace logger.go:42: 14:57:04 | one-pod/4-delete-data | mysql-clienttrue logger.go:42: 14:57:04 | one-pod/4-delete-data | + kubectl -n kuttl-test-guided-fish exec mysql-client -- bash -c 'printf '\''%s\n'\'' "TRUNCATE TABLE myDB.myTable" | mysql -sN -h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:57:04 | one-pod/4-delete-data | + sed -e 's/mysql: //' logger.go:42: 14:57:04 | one-pod/4-delete-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:57:06 | one-pod/4-delete-data | + : logger.go:42: 14:57:06 | one-pod/4-delete-data | ++++ get_cluster_name logger.go:42: 14:57:06 | one-pod/4-delete-data | ++++ kubectl -n kuttl-test-guided-fish get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:57:06 | one-pod/4-delete-data | +++ get_haproxy_svc one-pod logger.go:42: 14:57:06 | one-pod/4-delete-data | +++ local cluster=one-pod logger.go:42: 14:57:06 | one-pod/4-delete-data | +++ echo one-pod-haproxy logger.go:42: 14:57:06 | one-pod/4-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:57:06 | one-pod/4-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:57:06 | one-pod/4-delete-data | ++ local 'uri=-h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:57:06 | one-pod/4-delete-data | ++ local pod= logger.go:42: 14:57:06 | one-pod/4-delete-data | +++ get_client_pod logger.go:42: 14:57:06 | one-pod/4-delete-data | +++ kubectl -n kuttl-test-guided-fish get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:57:07 | one-pod/4-delete-data | ++ client_pod=mysql-client logger.go:42: 14:57:07 | one-pod/4-delete-data | ++ wait_pod mysql-client logger.go:42: 14:57:07 | one-pod/4-delete-data | ++ local pod=mysql-client logger.go:42: 14:57:07 | one-pod/4-delete-data | ++ set +o xtrace logger.go:42: 14:57:08 | one-pod/4-delete-data | mysql-clienttrue logger.go:42: 14:57:08 | one-pod/4-delete-data | ++ kubectl -n kuttl-test-guided-fish exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:57:08 | one-pod/4-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 14:57:08 | one-pod/4-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:57:09 | one-pod/4-delete-data | ++ : logger.go:42: 14:57:09 | one-pod/4-delete-data | + data= logger.go:42: 14:57:09 | one-pod/4-delete-data | + kubectl create configmap -n kuttl-test-guided-fish 04-delete-data-minio --from-literal=data= logger.go:42: 14:57:10 | one-pod/4-delete-data | configmap/04-delete-data-minio created logger.go:42: 14:57:11 | one-pod/4-delete-data | test step completed 4-delete-data logger.go:42: 14:57:11 | one-pod/5-restore-from-minio | starting test step 5-restore-from-minio logger.go:42: 14:57:13 | one-pod/5-restore-from-minio | PerconaServerMySQLRestore:kuttl-test-guided-fish/one-pod-restore-minio created logger.go:42: 14:59:09 | one-pod/5-restore-from-minio | test step completed 5-restore-from-minio logger.go:42: 14:59:09 | one-pod/6-read-data | starting test step 6-read-data logger.go:42: 14:59:09 | one-pod/6-read-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions data=$(run_mysql "SELECT * FROM myDB.myTable" "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 06-read-data-minio --from-literal=data="${data}"] logger.go:42: 14:59:09 | one-pod/6-read-data | + source ../../functions logger.go:42: 14:59:09 | one-pod/6-read-data | +++ realpath ../../.. logger.go:42: 14:59:09 | one-pod/6-read-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:59:09 | one-pod/6-read-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/vars.sh logger.go:42: 14:59:09 | one-pod/6-read-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:59:09 | one-pod/6-read-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:59:09 | one-pod/6-read-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:59:09 | one-pod/6-read-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:59:09 | one-pod/6-read-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:59:09 | one-pod/6-read-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:59:09 | one-pod/6-read-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:59:09 | one-pod/6-read-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:59:09 | one-pod/6-read-data | ++++ mktemp -d logger.go:42: 14:59:09 | one-pod/6-read-data | +++ export TEMP_DIR=/tmp/tmp.zcPxztcdu5 logger.go:42: 14:59:09 | one-pod/6-read-data | +++ TEMP_DIR=/tmp/tmp.zcPxztcdu5 logger.go:42: 14:59:09 | one-pod/6-read-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:59:09 | one-pod/6-read-data | +++ export GIT_BRANCH=PR-424 logger.go:42: 14:59:09 | one-pod/6-read-data | +++ GIT_BRANCH=PR-424 logger.go:42: 14:59:09 | one-pod/6-read-data | +++ export VERSION=PR-424-70568ae logger.go:42: 14:59:09 | one-pod/6-read-data | +++ VERSION=PR-424-70568ae logger.go:42: 14:59:09 | one-pod/6-read-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:59:09 | one-pod/6-read-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:59:09 | one-pod/6-read-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:59:09 | one-pod/6-read-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:59:09 | one-pod/6-read-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:59:09 | one-pod/6-read-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:59:09 | one-pod/6-read-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:59:09 | one-pod/6-read-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:59:09 | one-pod/6-read-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:59:09 | one-pod/6-read-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:59:09 | one-pod/6-read-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:59:09 | one-pod/6-read-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:59:09 | one-pod/6-read-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:59:09 | one-pod/6-read-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:59:09 | one-pod/6-read-data | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:59:09 | one-pod/6-read-data | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:59:09 | one-pod/6-read-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:59:09 | one-pod/6-read-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:59:09 | one-pod/6-read-data | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:59:09 | one-pod/6-read-data | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:59:09 | one-pod/6-read-data | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:59:09 | one-pod/6-read-data | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:59:09 | one-pod/6-read-data | ++++ which gdate logger.go:42: 14:59:09 | one-pod/6-read-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-424/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:59:09 | one-pod/6-read-data | ++++ which date logger.go:42: 14:59:09 | one-pod/6-read-data | +++ date=/usr/bin/date logger.go:42: 14:59:09 | one-pod/6-read-data | +++ command -v oc logger.go:42: 14:59:09 | one-pod/6-read-data | +++ oc get projects logger.go:42: 14:59:14 | one-pod/6-read-data | error: the server doesn't have a resource type "projects" logger.go:42: 14:59:14 | one-pod/6-read-data | +++ kubectl get nodes logger.go:42: 14:59:14 | one-pod/6-read-data | +++ grep '^minikube' logger.go:42: 14:59:15 | one-pod/6-read-data | ++++ pwd logger.go:42: 14:59:15 | one-pod/6-read-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/tests/one-pod logger.go:42: 14:59:15 | one-pod/6-read-data | ++ test_name=one-pod logger.go:42: 14:59:15 | one-pod/6-read-data | ++++ get_cluster_name logger.go:42: 14:59:15 | one-pod/6-read-data | ++++ kubectl -n kuttl-test-guided-fish get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:59:16 | one-pod/6-read-data | +++ get_haproxy_svc one-pod logger.go:42: 14:59:16 | one-pod/6-read-data | +++ local cluster=one-pod logger.go:42: 14:59:16 | one-pod/6-read-data | +++ echo one-pod-haproxy logger.go:42: 14:59:16 | one-pod/6-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:59:16 | one-pod/6-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:59:16 | one-pod/6-read-data | ++ local 'uri=-h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:59:16 | one-pod/6-read-data | ++ local pod= logger.go:42: 14:59:16 | one-pod/6-read-data | +++ get_client_pod logger.go:42: 14:59:16 | one-pod/6-read-data | +++ kubectl -n kuttl-test-guided-fish get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:59:17 | one-pod/6-read-data | ++ client_pod=mysql-client logger.go:42: 14:59:17 | one-pod/6-read-data | ++ wait_pod mysql-client logger.go:42: 14:59:17 | one-pod/6-read-data | ++ local pod=mysql-client logger.go:42: 14:59:17 | one-pod/6-read-data | ++ set +o xtrace logger.go:42: 14:59:18 | one-pod/6-read-data | mysql-clienttrue logger.go:42: 14:59:18 | one-pod/6-read-data | ++ kubectl -n kuttl-test-guided-fish exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:59:18 | one-pod/6-read-data | ++ sed -e 's/mysql: //' logger.go:42: 14:59:18 | one-pod/6-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:59:19 | one-pod/6-read-data | + data=100500 logger.go:42: 14:59:19 | one-pod/6-read-data | + kubectl create configmap -n kuttl-test-guided-fish 06-read-data-minio --from-literal=data=100500 logger.go:42: 14:59:20 | one-pod/6-read-data | configmap/06-read-data-minio created logger.go:42: 14:59:22 | one-pod/6-read-data | test step completed 6-read-data logger.go:42: 14:59:22 | one-pod/7-drop-finalizer | starting test step 7-drop-finalizer logger.go:42: 14:59:23 | one-pod/7-drop-finalizer | PerconaServerMySQL:kuttl-test-guided-fish/one-pod updated logger.go:42: 14:59:23 | one-pod/7-drop-finalizer | test step completed 7-drop-finalizer logger.go:42: 14:59:23 | one-pod | one-pod events from ns kuttl-test-guided-fish: logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:54:35 +0000 UTC Normal Pod percona-server-mysql-operator-6b56d66f99-7hd24 Scheduled Successfully assigned kuttl-test-guided-fish/percona-server-mysql-operator-6b56d66f99-7hd24 to gke-jen-ps-424-70568ae-7-default-pool-565e2e31-6ccq logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:54:35 +0000 UTC Normal ReplicaSet.apps percona-server-mysql-operator-6b56d66f99 SuccessfulCreate Created pod: percona-server-mysql-operator-6b56d66f99-7hd24 logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:54:35 +0000 UTC Normal Deployment.apps percona-server-mysql-operator ScalingReplicaSet Scaled up replica set percona-server-mysql-operator-6b56d66f99 to 1 logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:54:37 +0000 UTC Normal Pod percona-server-mysql-operator-6b56d66f99-7hd24.spec.containers{manager} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-424-70568ae" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:54:37 +0000 UTC Normal Pod percona-server-mysql-operator-6b56d66f99-7hd24.spec.containers{manager} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-424-70568ae" in 424.926169ms logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:54:37 +0000 UTC Normal Pod percona-server-mysql-operator-6b56d66f99-7hd24.spec.containers{manager} Created Created container manager logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:54:37 +0000 UTC Normal Pod percona-server-mysql-operator-6b56d66f99-7hd24.spec.containers{manager} Started Started container manager logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:54:38 +0000 UTC Normal Lease.coordination.k8s.io 08db2feb.percona.com LeaderElection percona-server-mysql-operator-6b56d66f99-7hd24_87e6abe1-6ae4-4b91-8a00-1623569b2a4a became leader logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:54:38 +0000 UTC Normal Pod mysql-client Scheduled Successfully assigned kuttl-test-guided-fish/mysql-client to gke-jen-ps-424-70568ae-7-default-pool-565e2e31-6ccq logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:54:39 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Pulled Container image "percona/percona-server:8.0.25" already present on machine logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:54:39 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Created Created container mysql-client logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:54:39 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Started Started container mysql-client logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:54:44 +0000 UTC Normal PersistentVolumeClaim minio-service WaitForFirstConsumer waiting for first consumer to be created before binding logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:54:45 +0000 UTC Normal ReplicaSet.apps minio-service-5fc4dc898 SuccessfulCreate Created pod: minio-service-5fc4dc898-bv7xb logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:54:45 +0000 UTC Normal Deployment.apps minio-service ScalingReplicaSet Scaled up replica set minio-service-5fc4dc898 to 1 logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:54:45 +0000 UTC Normal PersistentVolumeClaim minio-service Provisioning External provisioner is provisioning volume for claim "kuttl-test-guided-fish/minio-service" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:54:45 +0000 UTC Normal PersistentVolumeClaim minio-service ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:54:49 +0000 UTC Normal PersistentVolumeClaim minio-service ProvisioningSucceeded Successfully provisioned volume pvc-feed4811-f8ab-49e1-85c3-5763a999f35e logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:54:50 +0000 UTC Normal Pod minio-service-5fc4dc898-bv7xb Scheduled Successfully assigned kuttl-test-guided-fish/minio-service-5fc4dc898-bv7xb to gke-jen-ps-424-70568ae-7-default-pool-565e2e31-kv8q logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:54:58 +0000 UTC Normal Pod minio-service-5fc4dc898-bv7xb SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-feed4811-f8ab-49e1-85c3-5763a999f35e" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:54:59 +0000 UTC Normal Pod minio-service-5fc4dc898-bv7xb.spec.containers{minio} Pulling Pulling image "minio/minio:RELEASE.2020-11-19T23-48-16Z" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:01 +0000 UTC Normal Pod minio-service-5fc4dc898-bv7xb.spec.containers{minio} Pulled Successfully pulled image "minio/minio:RELEASE.2020-11-19T23-48-16Z" in 2.252638431s logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:01 +0000 UTC Normal Pod minio-service-5fc4dc898-bv7xb.spec.containers{minio} Created Created container minio logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:02 +0000 UTC Normal Pod minio-service-5fc4dc898-bv7xb.spec.containers{minio} Started Started container minio logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:03 +0000 UTC Normal Pod aws-cli Scheduled Successfully assigned kuttl-test-guided-fish/aws-cli to gke-jen-ps-424-70568ae-7-default-pool-565e2e31-6ks2 logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:04 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Pulling Pulling image "perconalab/awscli" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:08 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Pulled Successfully pulled image "perconalab/awscli" in 4.010733732s logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:08 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Created Created container aws-cli logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:08 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Started Started container aws-cli logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:24 +0000 UTC Normal PersistentVolumeClaim datadir-one-pod-mysql-0 WaitForFirstConsumer waiting for first consumer to be created before binding logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:24 +0000 UTC Normal PersistentVolumeClaim datadir-one-pod-mysql-0 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:24 +0000 UTC Normal PersistentVolumeClaim datadir-one-pod-mysql-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-guided-fish/datadir-one-pod-mysql-0" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:24 +0000 UTC Normal StatefulSet.apps one-pod-mysql SuccessfulCreate create Claim datadir-one-pod-mysql-0 Pod one-pod-mysql-0 in StatefulSet one-pod-mysql success logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:24 +0000 UTC Normal StatefulSet.apps one-pod-mysql SuccessfulCreate create Pod one-pod-mysql-0 in StatefulSet one-pod-mysql successful logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:26 +0000 UTC Normal Pod one-pod-orc-0 Scheduled Successfully assigned kuttl-test-guided-fish/one-pod-orc-0 to gke-jen-ps-424-70568ae-7-default-pool-565e2e31-6ccq logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:26 +0000 UTC Normal StatefulSet.apps one-pod-orc SuccessfulCreate create Pod one-pod-orc-0 in StatefulSet one-pod-orc successful logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:27 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-424-70568ae" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:28 +0000 UTC Normal PersistentVolumeClaim datadir-one-pod-mysql-0 ProvisioningSucceeded Successfully provisioned volume pvc-d92c7bb5-e768-4e24-ba15-828db20d2591 logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:28 +0000 UTC Normal Pod one-pod-mysql-0 Scheduled Successfully assigned kuttl-test-guided-fish/one-pod-mysql-0 to gke-jen-ps-424-70568ae-7-default-pool-565e2e31-6ks2 logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:28 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-424-70568ae" in 456.427444ms logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:28 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:28 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:29 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:29 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 383.615962ms logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:29 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Created Created container orc logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:30 +0000 UTC Warning Pod one-pod-mysql-0 FailedMount MountVolume.SetUp failed for volume "config" : failed to sync configmap cache: timed out waiting for the condition logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:30 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Started Started container orc logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:30 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:30 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 370.084848ms logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:30 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:30 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:36 +0000 UTC Normal Pod one-pod-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-d92c7bb5-e768-4e24-ba15-828db20d2591" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:37 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-424-70568ae" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:38 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-424-70568ae" in 374.52644ms logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:38 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:38 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:39 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:40 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 378.564331ms logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:40 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Created Created container mysql logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:40 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Started Started container mysql logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:40 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:40 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 408.103971ms logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:40 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:40 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:40 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:41 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 344.810126ms logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:41 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:55:41 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:56:15 +0000 UTC Normal Pod one-pod-haproxy-0 Scheduled Successfully assigned kuttl-test-guided-fish/one-pod-haproxy-0 to gke-jen-ps-424-70568ae-7-default-pool-565e2e31-6ccq logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:56:15 +0000 UTC Normal StatefulSet.apps one-pod-haproxy SuccessfulCreate create Pod one-pod-haproxy-0 in StatefulSet one-pod-haproxy successful logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:56:16 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-424-70568ae" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:56:17 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-424-70568ae" in 362.43914ms logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:56:17 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:56:17 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:56:18 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:56:19 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 385.193736ms logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:56:19 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Created Created container haproxy logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:56:19 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Started Started container haproxy logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:56:19 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:56:19 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 356.798769ms logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:56:19 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:56:19 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:56:41 +0000 UTC Normal Pod xb-one-pod-minio-minio-rv2t7 Scheduled Successfully assigned kuttl-test-guided-fish/xb-one-pod-minio-minio-rv2t7 to gke-jen-ps-424-70568ae-7-default-pool-565e2e31-6ks2 logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:56:41 +0000 UTC Normal Pod xb-one-pod-minio-minio-rv2t7.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-424-70568ae" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:56:41 +0000 UTC Normal Job.batch xb-one-pod-minio-minio SuccessfulCreate Created pod: xb-one-pod-minio-minio-rv2t7 logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:56:42 +0000 UTC Normal Pod xb-one-pod-minio-minio-rv2t7.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-424-70568ae" in 363.36537ms logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:56:42 +0000 UTC Normal Pod xb-one-pod-minio-minio-rv2t7.spec.initContainers{xtrabackup-init} Created Created container xtrabackup-init logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:56:42 +0000 UTC Normal Pod xb-one-pod-minio-minio-rv2t7.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:56:43 +0000 UTC Normal Pod xb-one-pod-minio-minio-rv2t7.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:56:44 +0000 UTC Normal Pod xb-one-pod-minio-minio-rv2t7.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 382.847755ms logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:56:44 +0000 UTC Normal Pod xb-one-pod-minio-minio-rv2t7.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:56:44 +0000 UTC Normal Pod xb-one-pod-minio-minio-rv2t7.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:56:51 +0000 UTC Normal Job.batch xb-one-pod-minio-minio Completed Job completed logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:57:16 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Killing Stopping container haproxy logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:57:16 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:57:16 +0000 UTC Normal StatefulSet.apps one-pod-haproxy SuccessfulDelete delete Pod one-pod-haproxy-0 in StatefulSet one-pod-haproxy successful logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:57:16 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Killing Stopping container mysql logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:57:16 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Killing Stopping container pt-heartbeat logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:57:16 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Killing Stopping container xtrabackup logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:57:16 +0000 UTC Normal StatefulSet.apps one-pod-mysql SuccessfulDelete delete Pod one-pod-mysql-0 in StatefulSet one-pod-mysql successful logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:57:16 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Killing Stopping container orc logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:57:16 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:57:16 +0000 UTC Normal StatefulSet.apps one-pod-orc SuccessfulDelete delete Pod one-pod-orc-0 in StatefulSet one-pod-orc successful logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:57:17 +0000 UTC Warning Pod one-pod-mysql-0.spec.containers{mysql} Unhealthy Readiness probe failed: 2023/08/17 14:57:17 readiness check failed: connect to db: ping database: dial tcp 10.237.240.25:33062: connect: connection refused logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:57:48 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-7lt4r Scheduled Successfully assigned kuttl-test-guided-fish/xb-restore-one-pod-restore-minio-7lt4r to gke-jen-ps-424-70568ae-7-default-pool-565e2e31-6ks2 logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:57:48 +0000 UTC Normal Job.batch xb-restore-one-pod-restore-minio SuccessfulCreate Created pod: xb-restore-one-pod-restore-minio-7lt4r logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:57:53 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-7lt4r SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-d92c7bb5-e768-4e24-ba15-828db20d2591" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:57:54 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-7lt4r.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-424-70568ae" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:57:54 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-7lt4r.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-424-70568ae" in 415.528498ms logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:57:54 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-7lt4r.spec.initContainers{xtrabackup-init} Created Created container xtrabackup-init logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:57:54 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-7lt4r.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:57:56 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-7lt4r.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:57:56 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-7lt4r.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 397.600417ms logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:57:56 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-7lt4r.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:57:56 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-7lt4r.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:05 +0000 UTC Warning Service one-pod-haproxy ProcessServiceFailed error processing service "kuttl-test-guided-fish/one-pod-haproxy": services "one-pod-haproxy" not found logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:17 +0000 UTC Normal Job.batch xb-restore-one-pod-restore-minio Completed Job completed logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:18 +0000 UTC Normal Pod one-pod-mysql-0 Scheduled Successfully assigned kuttl-test-guided-fish/one-pod-mysql-0 to gke-jen-ps-424-70568ae-7-default-pool-565e2e31-6ks2 logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:18 +0000 UTC Normal Pod one-pod-orc-0 Scheduled Successfully assigned kuttl-test-guided-fish/one-pod-orc-0 to gke-jen-ps-424-70568ae-7-default-pool-565e2e31-6ccq logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:19 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-424-70568ae" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:20 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-424-70568ae" in 400.02991ms logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:20 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:20 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:21 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-424-70568ae" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:21 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-424-70568ae" in 370.400068ms logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:21 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:21 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:22 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:22 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 371.721475ms logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:22 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Created Created container orc logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:22 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Started Started container orc logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:22 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:22 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 355.125884ms logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:22 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:22 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:23 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:23 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 395.792356ms logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:23 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Created Created container mysql logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:23 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Started Started container mysql logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:23 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:23 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 366.333878ms logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:24 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:24 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:24 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:24 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 401.310295ms logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:24 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:58:24 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:59:00 +0000 UTC Normal Pod one-pod-haproxy-0 Scheduled Successfully assigned kuttl-test-guided-fish/one-pod-haproxy-0 to gke-jen-ps-424-70568ae-7-default-pool-565e2e31-6ks2 logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:59:01 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-424-70568ae" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:59:02 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-424-70568ae" in 396.324426ms logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:59:02 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:59:02 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:59:03 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:59:03 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 395.422532ms logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:59:03 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Created Created container haproxy logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:59:03 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Started Started container haproxy logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:59:03 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:59:04 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 363.280894ms logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:59:04 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 14:59:23 | one-pod | 2023-08-17 14:59:04 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 14:59:23 | one-pod | Deleting namespace: kuttl-test-guided-fish === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (349.40s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/one-pod (346.00s) PASS