=== RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://34.133.14.34 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 180 seconds for each step harness.go:372: testsuite: e2e-tests/tests has 26 tests === RUN kuttl/harness === RUN kuttl/harness/one-pod === PAUSE kuttl/harness/one-pod === CONT kuttl/harness/one-pod logger.go:42: 13:05:23 | one-pod | Creating namespace: kuttl-test-liked-quetzal logger.go:42: 13:05:23 | one-pod/0-deploy-operator | starting test step 0-deploy-operator logger.go:42: 13:05:23 | one-pod/0-deploy-operator | running command: [sh -c set -o errexit set -o xtrace source ../../functions apply_s3_storage_secrets deploy_operator deploy_non_tls_cluster_secrets deploy_tls_cluster_secrets deploy_client deploy_minio] logger.go:42: 13:05:23 | one-pod/0-deploy-operator | + source ../../functions logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ realpath ../../.. logger.go:42: 13:05:23 | one-pod/0-deploy-operator | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:05:23 | one-pod/0-deploy-operator | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:05:23 | one-pod/0-deploy-operator | ++++ mktemp -d logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ export TEMP_DIR=/tmp/tmp.v0O6FZQILJ logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ TEMP_DIR=/tmp/tmp.v0O6FZQILJ logger.go:42: 13:05:23 | one-pod/0-deploy-operator | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ export GIT_BRANCH=PR-339 logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ GIT_BRANCH=PR-339 logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ export VERSION=PR-339-3a8483b logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ VERSION=PR-339-3a8483b logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:05:23 | one-pod/0-deploy-operator | ++++ which gdate logger.go:42: 13:05:23 | one-pod/0-deploy-operator | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 13:05:23 | one-pod/0-deploy-operator | ++++ which date logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ date=/usr/bin/date logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ command -v oc logger.go:42: 13:05:23 | one-pod/0-deploy-operator | +++ oc get projects logger.go:42: 13:05:29 | one-pod/0-deploy-operator | error: the server doesn't have a resource type "projects" logger.go:42: 13:05:29 | one-pod/0-deploy-operator | +++ kubectl get nodes logger.go:42: 13:05:29 | one-pod/0-deploy-operator | +++ grep '^minikube' logger.go:42: 13:05:30 | one-pod/0-deploy-operator | ++++ pwd logger.go:42: 13:05:30 | one-pod/0-deploy-operator | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/one-pod logger.go:42: 13:05:30 | one-pod/0-deploy-operator | ++ test_name=one-pod logger.go:42: 13:05:30 | one-pod/0-deploy-operator | + apply_s3_storage_secrets logger.go:42: 13:05:30 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-liked-quetzal apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/minio-secret.yml logger.go:42: 13:05:31 | one-pod/0-deploy-operator | secret/minio-secret created logger.go:42: 13:05:31 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-liked-quetzal apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/cloud-secret.yml logger.go:42: 13:05:32 | one-pod/0-deploy-operator | secret/aws-s3-secret created logger.go:42: 13:05:33 | one-pod/0-deploy-operator | secret/gcp-cs-secret created logger.go:42: 13:05:33 | one-pod/0-deploy-operator | secret/azure-secret created logger.go:42: 13:05:33 | one-pod/0-deploy-operator | + deploy_operator logger.go:42: 13:05:33 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-liked-quetzal apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/crd.yaml logger.go:42: 13:05:34 | one-pod/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlbackups.ps.percona.com serverside-applied logger.go:42: 13:05:35 | one-pod/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlrestores.ps.percona.com serverside-applied logger.go:42: 13:05:36 | one-pod/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqls.ps.percona.com serverside-applied logger.go:42: 13:05:36 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-liked-quetzal apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/rbac.yaml logger.go:42: 13:05:37 | one-pod/0-deploy-operator | serviceaccount/percona-server-mysql-operator created logger.go:42: 13:05:37 | one-pod/0-deploy-operator | serviceaccount/percona-server-mysql-operator-orchestrator created logger.go:42: 13:05:37 | one-pod/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 13:05:38 | one-pod/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 13:05:38 | one-pod/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 13:05:38 | one-pod/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 13:05:39 | one-pod/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 13:05:39 | one-pod/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 13:05:39 | one-pod/0-deploy-operator | + yq eval '(select(documentIndex==1).spec.template.spec.containers[] | select(.name=="manager").env[] | select(.name=="DISABLE_TELEMETRY").value) = "true"' logger.go:42: 13:05:39 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-liked-quetzal apply -f - logger.go:42: 13:05:39 | one-pod/0-deploy-operator | ++ printf 'select(documentIndex==1).spec.template.spec.containers[0].image="%s"' perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:05:39 | one-pod/0-deploy-operator | + yq eval 'select(documentIndex==1).spec.template.spec.containers[0].image="perconalab/percona-server-mysql-operator:PR-339-3a8483b"' /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/operator.yaml logger.go:42: 13:05:40 | one-pod/0-deploy-operator | configmap/percona-server-mysql-operator-config created logger.go:42: 13:05:41 | one-pod/0-deploy-operator | deployment.apps/percona-server-mysql-operator created logger.go:42: 13:05:41 | one-pod/0-deploy-operator | + deploy_non_tls_cluster_secrets logger.go:42: 13:05:41 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-liked-quetzal apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/secrets.yaml logger.go:42: 13:05:42 | one-pod/0-deploy-operator | secret/test-secrets created logger.go:42: 13:05:42 | one-pod/0-deploy-operator | + deploy_tls_cluster_secrets logger.go:42: 13:05:42 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-liked-quetzal apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/ssl-secret.yaml logger.go:42: 13:05:43 | one-pod/0-deploy-operator | secret/test-ssl created logger.go:42: 13:05:43 | one-pod/0-deploy-operator | + deploy_client logger.go:42: 13:05:43 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-liked-quetzal apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/client.yaml logger.go:42: 13:05:45 | one-pod/0-deploy-operator | pod/mysql-client created logger.go:42: 13:05:45 | one-pod/0-deploy-operator | + deploy_minio logger.go:42: 13:05:45 | one-pod/0-deploy-operator | ++ kubectl -n kuttl-test-liked-quetzal get secret minio-secret -o 'jsonpath={.data.AWS_ACCESS_KEY_ID}' logger.go:42: 13:05:45 | one-pod/0-deploy-operator | ++ base64 -d logger.go:42: 13:05:45 | one-pod/0-deploy-operator | + accessKey=some-access-key logger.go:42: 13:05:45 | one-pod/0-deploy-operator | ++ kubectl -n kuttl-test-liked-quetzal get secret minio-secret -o 'jsonpath={.data.AWS_SECRET_ACCESS_KEY}' logger.go:42: 13:05:45 | one-pod/0-deploy-operator | ++ base64 -d logger.go:42: 13:05:46 | one-pod/0-deploy-operator | + secretKey=some-secret-key logger.go:42: 13:05:46 | one-pod/0-deploy-operator | + helm uninstall -n kuttl-test-liked-quetzal minio-service logger.go:42: 13:05:46 | one-pod/0-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 13:05:46 | one-pod/0-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 13:05:46 | one-pod/0-deploy-operator | Error: uninstall: Release not loaded: minio-service: release: not found logger.go:42: 13:05:46 | one-pod/0-deploy-operator | + : logger.go:42: 13:05:46 | one-pod/0-deploy-operator | + helm repo remove minio logger.go:42: 13:05:46 | one-pod/0-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 13:05:46 | one-pod/0-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 13:05:46 | one-pod/0-deploy-operator | "minio" has been removed from your repositories logger.go:42: 13:05:46 | one-pod/0-deploy-operator | + helm repo add minio https://helm.min.io/ logger.go:42: 13:05:47 | one-pod/0-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 13:05:47 | one-pod/0-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 13:05:47 | one-pod/0-deploy-operator | "minio" has been added to your repositories logger.go:42: 13:05:47 | one-pod/0-deploy-operator | ++ printf %q some-access-key logger.go:42: 13:05:47 | one-pod/0-deploy-operator | ++ printf %q some-secret-key logger.go:42: 13:05:47 | one-pod/0-deploy-operator | + retry 10 60 helm install minio-service -n kuttl-test-liked-quetzal --version 8.0.5 --set accessKey=some-access-key --set secretKey=some-secret-key --set service.type=ClusterIP --set configPathmc=/tmp/.minio/ --set persistence.size=2G --set environment.MINIO_REGION=us-east-1 --set environment.MINIO_HTTP_TRACE=/tmp/trace.log --set securityContext.enabled=false minio/minio logger.go:42: 13:05:47 | one-pod/0-deploy-operator | + local max=10 logger.go:42: 13:05:47 | one-pod/0-deploy-operator | + local delay=60 logger.go:42: 13:05:47 | one-pod/0-deploy-operator | + shift 2 logger.go:42: 13:05:47 | one-pod/0-deploy-operator | + local n=1 logger.go:42: 13:05:47 | one-pod/0-deploy-operator | + helm install minio-service -n kuttl-test-liked-quetzal --version 8.0.5 --set accessKey=some-access-key --set secretKey=some-secret-key --set service.type=ClusterIP --set configPathmc=/tmp/.minio/ --set persistence.size=2G --set environment.MINIO_REGION=us-east-1 --set environment.MINIO_HTTP_TRACE=/tmp/trace.log --set securityContext.enabled=false minio/minio logger.go:42: 13:05:47 | one-pod/0-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 13:05:47 | one-pod/0-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 13:05:51 | one-pod/0-deploy-operator | NAME: minio-service logger.go:42: 13:05:51 | one-pod/0-deploy-operator | LAST DEPLOYED: Wed Aug 9 13:05:48 2023 logger.go:42: 13:05:51 | one-pod/0-deploy-operator | NAMESPACE: kuttl-test-liked-quetzal logger.go:42: 13:05:51 | one-pod/0-deploy-operator | STATUS: deployed logger.go:42: 13:05:51 | one-pod/0-deploy-operator | REVISION: 1 logger.go:42: 13:05:51 | one-pod/0-deploy-operator | TEST SUITE: None logger.go:42: 13:05:51 | one-pod/0-deploy-operator | NOTES: logger.go:42: 13:05:51 | one-pod/0-deploy-operator | Minio can be accessed via port 9000 on the following DNS name from within your cluster: logger.go:42: 13:05:51 | one-pod/0-deploy-operator | minio-service.kuttl-test-liked-quetzal.svc.cluster.local logger.go:42: 13:05:51 | one-pod/0-deploy-operator | logger.go:42: 13:05:51 | one-pod/0-deploy-operator | To access Minio from localhost, run the below commands: logger.go:42: 13:05:51 | one-pod/0-deploy-operator | logger.go:42: 13:05:51 | one-pod/0-deploy-operator | 1. export POD_NAME=$(kubectl get pods --namespace kuttl-test-liked-quetzal -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}") logger.go:42: 13:05:51 | one-pod/0-deploy-operator | logger.go:42: 13:05:51 | one-pod/0-deploy-operator | 2. kubectl port-forward $POD_NAME 9000 --namespace kuttl-test-liked-quetzal logger.go:42: 13:05:51 | one-pod/0-deploy-operator | logger.go:42: 13:05:51 | one-pod/0-deploy-operator | Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/ logger.go:42: 13:05:51 | one-pod/0-deploy-operator | logger.go:42: 13:05:51 | one-pod/0-deploy-operator | You can now access Minio server on http://localhost:9000. Follow the below steps to connect to Minio server with mc client: logger.go:42: 13:05:51 | one-pod/0-deploy-operator | logger.go:42: 13:05:51 | one-pod/0-deploy-operator | 1. Download the Minio mc client - https://docs.minio.io/docs/minio-client-quickstart-guide logger.go:42: 13:05:51 | one-pod/0-deploy-operator | logger.go:42: 13:05:51 | one-pod/0-deploy-operator | 2. Get the ACCESS_KEY=$(kubectl get secret minio-service -o jsonpath="{.data.accesskey}" | base64 --decode) and the SECRET_KEY=$(kubectl get secret minio-service -o jsonpath="{.data.secretkey}" | base64 --decode) logger.go:42: 13:05:51 | one-pod/0-deploy-operator | logger.go:42: 13:05:51 | one-pod/0-deploy-operator | 3. mc alias set minio-service-local http://localhost:9000 "$ACCESS_KEY" "$SECRET_KEY" --api s3v4 logger.go:42: 13:05:51 | one-pod/0-deploy-operator | logger.go:42: 13:05:51 | one-pod/0-deploy-operator | 4. mc ls minio-service-local logger.go:42: 13:05:51 | one-pod/0-deploy-operator | logger.go:42: 13:05:51 | one-pod/0-deploy-operator | Alternately, you can use your browser or the Minio SDK to access the server - https://docs.minio.io/categories/17 logger.go:42: 13:05:51 | one-pod/0-deploy-operator | ++ kubectl -n kuttl-test-liked-quetzal get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:05:51 | one-pod/0-deploy-operator | + MINIO_POD=minio-service-5fc4dc898-64kpv logger.go:42: 13:05:51 | one-pod/0-deploy-operator | + wait_pod minio-service-5fc4dc898-64kpv logger.go:42: 13:05:51 | one-pod/0-deploy-operator | + local pod=minio-service-5fc4dc898-64kpv logger.go:42: 13:05:51 | one-pod/0-deploy-operator | + set +o xtrace logger.go:42: 13:06:08 | one-pod/0-deploy-operator | minio-service-5fc4dc898-64kpv.........true logger.go:42: 13:06:08 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-liked-quetzal run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- bash -c 'AWS_ACCESS_KEY_ID='\''some-access-key'\'' AWS_SECRET_ACCESS_KEY='\''some-secret-key'\'' AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url http://minio-service:9000 s3 mb s3://operator-testing' logger.go:42: 13:06:10 | one-pod/0-deploy-operator | If you don't see a command prompt, try pressing enter. logger.go:42: 13:06:11 | one-pod/0-deploy-operator | Error attaching, falling back to logs: unable to upgrade connection: container aws-cli not found in pod aws-cli_kuttl-test-liked-quetzal logger.go:42: 13:06:11 | one-pod/0-deploy-operator | make_bucket: operator-testing logger.go:42: 13:06:14 | one-pod/0-deploy-operator | pod "aws-cli" deleted logger.go:42: 13:06:15 | one-pod/0-deploy-operator | test step completed 0-deploy-operator logger.go:42: 13:06:15 | one-pod/1-create-cluster | starting test step 1-create-cluster logger.go:42: 13:06:15 | one-pod/1-create-cluster | running command: [sh -c set -o errexit set -o xtrace source ../../functions get_cr \ | yq eval '.spec.mysql.clusterType="async"' - \ | yq eval '.spec.allowUnsafeConfigurations=true' - \ | yq eval '.spec.mysql.size=1' - \ | yq eval '.spec.proxy.haproxy.enabled=true' - \ | yq eval '.spec.proxy.haproxy.size=1' - \ | yq eval '.spec.orchestrator.enabled=true' - \ | yq eval '.spec.orchestrator.size=1' - \ | yq eval '.spec.backup.storages.minio.type="s3"' - \ | yq eval '.spec.backup.storages.minio.s3.bucket="operator-testing"' - \ | yq eval '.spec.backup.storages.minio.s3.credentialsSecret="minio-secret"' - \ | yq eval '.spec.backup.storages.minio.s3.endpointUrl="http://minio-service:9000"' - \ | yq eval '.spec.backup.storages.minio.s3.region="us-east-1"' - \ | kubectl -n "${NAMESPACE}" apply -f -] logger.go:42: 13:06:15 | one-pod/1-create-cluster | + source ../../functions logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ realpath ../../.. logger.go:42: 13:06:15 | one-pod/1-create-cluster | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:06:15 | one-pod/1-create-cluster | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:06:15 | one-pod/1-create-cluster | ++++ mktemp -d logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ export TEMP_DIR=/tmp/tmp.QNHFZ146FK logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ TEMP_DIR=/tmp/tmp.QNHFZ146FK logger.go:42: 13:06:15 | one-pod/1-create-cluster | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ export GIT_BRANCH=PR-339 logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ GIT_BRANCH=PR-339 logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ export VERSION=PR-339-3a8483b logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ VERSION=PR-339-3a8483b logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:06:15 | one-pod/1-create-cluster | ++++ which gdate logger.go:42: 13:06:15 | one-pod/1-create-cluster | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 13:06:15 | one-pod/1-create-cluster | ++++ which date logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ date=/usr/bin/date logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ command -v oc logger.go:42: 13:06:15 | one-pod/1-create-cluster | +++ oc get projects logger.go:42: 13:06:21 | one-pod/1-create-cluster | error: the server doesn't have a resource type "projects" logger.go:42: 13:06:21 | one-pod/1-create-cluster | +++ kubectl get nodes logger.go:42: 13:06:21 | one-pod/1-create-cluster | +++ grep '^minikube' logger.go:42: 13:06:22 | one-pod/1-create-cluster | ++++ pwd logger.go:42: 13:06:22 | one-pod/1-create-cluster | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/one-pod logger.go:42: 13:06:22 | one-pod/1-create-cluster | ++ test_name=one-pod logger.go:42: 13:06:22 | one-pod/1-create-cluster | + get_cr logger.go:42: 13:06:22 | one-pod/1-create-cluster | + local name_suffix= logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval .spec.allowUnsafeConfigurations=true - logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval .spec.orchestrator.size=1 - logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval '.spec.backup.storages.minio.s3.region="us-east-1"' - logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval '.spec.backup.storages.minio.s3.credentialsSecret="minio-secret"' - logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval '.spec.backup.storages.minio.s3.bucket="operator-testing"' - logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval '.spec.backup.storages.minio.s3.endpointUrl="http://minio-service:9000"' - logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval '.spec.backup.storages.minio.type="s3"' - logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval .spec.orchestrator.enabled=true - logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval .spec.proxy.haproxy.enabled=true - logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval .spec.mysql.size=1 - logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval '.spec.sslSecretName="test-ssl"' - logger.go:42: 13:06:22 | one-pod/1-create-cluster | + kubectl -n kuttl-test-liked-quetzal apply -f - logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval .spec.proxy.haproxy.size=1 - logger.go:42: 13:06:22 | one-pod/1-create-cluster | ++ printf '.spec.initImage="%s"' perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval '.spec.initImage="perconalab/percona-server-mysql-operator:PR-339-3a8483b"' - logger.go:42: 13:06:22 | one-pod/1-create-cluster | + '[' -n '' ']' logger.go:42: 13:06:22 | one-pod/1-create-cluster | ++ printf '.spec.mysql.image="%s"' perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval '.spec.mysql.image="perconalab/percona-server-mysql-operator:main-psmysql"' - logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval '.spec.secretsName="test-secrets"' - logger.go:42: 13:06:22 | one-pod/1-create-cluster | ++ printf '.spec.backup.image="%s"' perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval '.spec.upgradeOptions.apply="disabled"' - logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval '.spec.backup.image="perconalab/percona-server-mysql-operator:main-backup"' - logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval - logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 13:06:22 | one-pod/1-create-cluster | ++ printf '.metadata.name="%s"' one-pod logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval '.metadata.name="one-pod"' /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/cr.yaml logger.go:42: 13:06:22 | one-pod/1-create-cluster | ++ printf '.spec.pmm.image="%s"' perconalab/pmm-client:dev-latest logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval '.spec.pmm.image="perconalab/pmm-client:dev-latest"' - logger.go:42: 13:06:22 | one-pod/1-create-cluster | ++ printf '.spec.orchestrator.image="%s"' perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval '.spec.orchestrator.image="perconalab/percona-server-mysql-operator:main-orchestrator"' - logger.go:42: 13:06:22 | one-pod/1-create-cluster | ++ printf '.spec.proxy.router.image="%s"' perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval '.spec.proxy.router.image="perconalab/percona-server-mysql-operator:main-router"' - logger.go:42: 13:06:22 | one-pod/1-create-cluster | ++ printf '.spec.proxy.haproxy.image="%s"' perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval '.spec.proxy.haproxy.image="perconalab/percona-server-mysql-operator:main-haproxy"' - logger.go:42: 13:06:22 | one-pod/1-create-cluster | ++ printf '.spec.toolkit.image="%s"' perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:06:22 | one-pod/1-create-cluster | + yq eval '.spec.toolkit.image="perconalab/percona-server-mysql-operator:main-toolkit"' - logger.go:42: 13:06:23 | one-pod/1-create-cluster | perconaservermysql.ps.percona.com/one-pod created logger.go:42: 13:07:25 | one-pod/1-create-cluster | test step completed 1-create-cluster logger.go:42: 13:07:25 | one-pod/2-write-data | starting test step 2-write-data logger.go:42: 13:07:25 | one-pod/2-write-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password" run_mysql \ "INSERT myDB.myTable (id) VALUES (100500)" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password"] logger.go:42: 13:07:25 | one-pod/2-write-data | + source ../../functions logger.go:42: 13:07:25 | one-pod/2-write-data | +++ realpath ../../.. logger.go:42: 13:07:25 | one-pod/2-write-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:07:25 | one-pod/2-write-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 13:07:25 | one-pod/2-write-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:07:25 | one-pod/2-write-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:07:25 | one-pod/2-write-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:07:25 | one-pod/2-write-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:07:25 | one-pod/2-write-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:07:25 | one-pod/2-write-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:07:25 | one-pod/2-write-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:07:25 | one-pod/2-write-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:07:25 | one-pod/2-write-data | ++++ mktemp -d logger.go:42: 13:07:25 | one-pod/2-write-data | +++ export TEMP_DIR=/tmp/tmp.UbAY4MFrji logger.go:42: 13:07:25 | one-pod/2-write-data | +++ TEMP_DIR=/tmp/tmp.UbAY4MFrji logger.go:42: 13:07:25 | one-pod/2-write-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 13:07:25 | one-pod/2-write-data | +++ export GIT_BRANCH=PR-339 logger.go:42: 13:07:25 | one-pod/2-write-data | +++ GIT_BRANCH=PR-339 logger.go:42: 13:07:25 | one-pod/2-write-data | +++ export VERSION=PR-339-3a8483b logger.go:42: 13:07:25 | one-pod/2-write-data | +++ VERSION=PR-339-3a8483b logger.go:42: 13:07:25 | one-pod/2-write-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:07:25 | one-pod/2-write-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:07:25 | one-pod/2-write-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:07:25 | one-pod/2-write-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:07:25 | one-pod/2-write-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:07:25 | one-pod/2-write-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:07:25 | one-pod/2-write-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:07:25 | one-pod/2-write-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:07:25 | one-pod/2-write-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:07:25 | one-pod/2-write-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:07:25 | one-pod/2-write-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:07:25 | one-pod/2-write-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:07:25 | one-pod/2-write-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:07:25 | one-pod/2-write-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:07:25 | one-pod/2-write-data | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:07:25 | one-pod/2-write-data | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:07:25 | one-pod/2-write-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:07:25 | one-pod/2-write-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:07:25 | one-pod/2-write-data | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:07:25 | one-pod/2-write-data | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:07:25 | one-pod/2-write-data | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:07:25 | one-pod/2-write-data | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:07:25 | one-pod/2-write-data | ++++ which gdate logger.go:42: 13:07:25 | one-pod/2-write-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 13:07:25 | one-pod/2-write-data | ++++ which date logger.go:42: 13:07:25 | one-pod/2-write-data | +++ date=/usr/bin/date logger.go:42: 13:07:25 | one-pod/2-write-data | +++ command -v oc logger.go:42: 13:07:25 | one-pod/2-write-data | +++ oc get projects logger.go:42: 13:07:30 | one-pod/2-write-data | error: the server doesn't have a resource type "projects" logger.go:42: 13:07:30 | one-pod/2-write-data | +++ kubectl get nodes logger.go:42: 13:07:30 | one-pod/2-write-data | +++ grep '^minikube' logger.go:42: 13:07:31 | one-pod/2-write-data | ++++ pwd logger.go:42: 13:07:31 | one-pod/2-write-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/one-pod logger.go:42: 13:07:31 | one-pod/2-write-data | ++ test_name=one-pod logger.go:42: 13:07:31 | one-pod/2-write-data | +++ get_cluster_name logger.go:42: 13:07:31 | one-pod/2-write-data | +++ kubectl -n kuttl-test-liked-quetzal get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 13:07:32 | one-pod/2-write-data | ++ get_haproxy_svc one-pod logger.go:42: 13:07:32 | one-pod/2-write-data | ++ local cluster=one-pod logger.go:42: 13:07:32 | one-pod/2-write-data | ++ echo one-pod-haproxy logger.go:42: 13:07:32 | one-pod/2-write-data | + run_mysql 'CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' '-h one-pod-haproxy -uroot -proot_password' logger.go:42: 13:07:32 | one-pod/2-write-data | + local 'command=CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' logger.go:42: 13:07:32 | one-pod/2-write-data | + local 'uri=-h one-pod-haproxy -uroot -proot_password' logger.go:42: 13:07:32 | one-pod/2-write-data | + local pod= logger.go:42: 13:07:32 | one-pod/2-write-data | ++ get_client_pod logger.go:42: 13:07:32 | one-pod/2-write-data | ++ kubectl -n kuttl-test-liked-quetzal get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:07:33 | one-pod/2-write-data | + client_pod=mysql-client logger.go:42: 13:07:33 | one-pod/2-write-data | + wait_pod mysql-client logger.go:42: 13:07:33 | one-pod/2-write-data | + local pod=mysql-client logger.go:42: 13:07:33 | one-pod/2-write-data | + set +o xtrace logger.go:42: 13:07:33 | one-pod/2-write-data | mysql-clienttrue logger.go:42: 13:07:33 | one-pod/2-write-data | + sed -e 's/mysql: //' logger.go:42: 13:07:33 | one-pod/2-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:07:33 | one-pod/2-write-data | + kubectl -n kuttl-test-liked-quetzal exec mysql-client -- bash -c 'printf '\''%s\n'\'' "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" | mysql -sN -h one-pod-haproxy -uroot -proot_password' logger.go:42: 13:07:35 | one-pod/2-write-data | + : logger.go:42: 13:07:35 | one-pod/2-write-data | +++ get_cluster_name logger.go:42: 13:07:35 | one-pod/2-write-data | +++ kubectl -n kuttl-test-liked-quetzal get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 13:07:36 | one-pod/2-write-data | ++ get_haproxy_svc one-pod logger.go:42: 13:07:36 | one-pod/2-write-data | ++ local cluster=one-pod logger.go:42: 13:07:36 | one-pod/2-write-data | ++ echo one-pod-haproxy logger.go:42: 13:07:36 | one-pod/2-write-data | + run_mysql 'INSERT myDB.myTable (id) VALUES (100500)' '-h one-pod-haproxy -uroot -proot_password' logger.go:42: 13:07:36 | one-pod/2-write-data | + local 'command=INSERT myDB.myTable (id) VALUES (100500)' logger.go:42: 13:07:36 | one-pod/2-write-data | + local 'uri=-h one-pod-haproxy -uroot -proot_password' logger.go:42: 13:07:36 | one-pod/2-write-data | + local pod= logger.go:42: 13:07:36 | one-pod/2-write-data | ++ get_client_pod logger.go:42: 13:07:36 | one-pod/2-write-data | ++ kubectl -n kuttl-test-liked-quetzal get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:07:36 | one-pod/2-write-data | + client_pod=mysql-client logger.go:42: 13:07:36 | one-pod/2-write-data | + wait_pod mysql-client logger.go:42: 13:07:36 | one-pod/2-write-data | + local pod=mysql-client logger.go:42: 13:07:36 | one-pod/2-write-data | + set +o xtrace logger.go:42: 13:07:37 | one-pod/2-write-data | mysql-clienttrue logger.go:42: 13:07:37 | one-pod/2-write-data | + sed -e 's/mysql: //' logger.go:42: 13:07:37 | one-pod/2-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:07:37 | one-pod/2-write-data | + kubectl -n kuttl-test-liked-quetzal exec mysql-client -- bash -c 'printf '\''%s\n'\'' "INSERT myDB.myTable (id) VALUES (100500)" | mysql -sN -h one-pod-haproxy -uroot -proot_password' logger.go:42: 13:07:39 | one-pod/2-write-data | + : logger.go:42: 13:07:40 | one-pod/2-write-data | test step completed 2-write-data logger.go:42: 13:07:40 | one-pod/3-create-backup-minio | starting test step 3-create-backup-minio logger.go:42: 13:07:41 | one-pod/3-create-backup-minio | PerconaServerMySQLBackup:kuttl-test-liked-quetzal/one-pod-minio created logger.go:42: 13:07:52 | one-pod/3-create-backup-minio | test step completed 3-create-backup-minio logger.go:42: 13:07:52 | one-pod/4-delete-data | starting test step 4-delete-data logger.go:42: 13:07:52 | one-pod/4-delete-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "TRUNCATE TABLE myDB.myTable" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password" data=$(run_mysql "SELECT * FROM myDB.myTable" "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 04-delete-data-minio --from-literal=data="${data}"] logger.go:42: 13:07:52 | one-pod/4-delete-data | + source ../../functions logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ realpath ../../.. logger.go:42: 13:07:52 | one-pod/4-delete-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:07:52 | one-pod/4-delete-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:07:52 | one-pod/4-delete-data | ++++ mktemp -d logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ export TEMP_DIR=/tmp/tmp.XDDrv8Weuy logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ TEMP_DIR=/tmp/tmp.XDDrv8Weuy logger.go:42: 13:07:52 | one-pod/4-delete-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ export GIT_BRANCH=PR-339 logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ GIT_BRANCH=PR-339 logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ export VERSION=PR-339-3a8483b logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ VERSION=PR-339-3a8483b logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:07:52 | one-pod/4-delete-data | ++++ which gdate logger.go:42: 13:07:52 | one-pod/4-delete-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 13:07:52 | one-pod/4-delete-data | ++++ which date logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ date=/usr/bin/date logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ command -v oc logger.go:42: 13:07:52 | one-pod/4-delete-data | +++ oc get projects logger.go:42: 13:07:58 | one-pod/4-delete-data | error: the server doesn't have a resource type "projects" logger.go:42: 13:07:58 | one-pod/4-delete-data | +++ kubectl get nodes logger.go:42: 13:07:58 | one-pod/4-delete-data | +++ grep '^minikube' logger.go:42: 13:07:59 | one-pod/4-delete-data | ++++ pwd logger.go:42: 13:07:59 | one-pod/4-delete-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/one-pod logger.go:42: 13:07:59 | one-pod/4-delete-data | ++ test_name=one-pod logger.go:42: 13:07:59 | one-pod/4-delete-data | +++ get_cluster_name logger.go:42: 13:07:59 | one-pod/4-delete-data | +++ kubectl -n kuttl-test-liked-quetzal get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 13:07:59 | one-pod/4-delete-data | ++ get_haproxy_svc one-pod logger.go:42: 13:07:59 | one-pod/4-delete-data | ++ local cluster=one-pod logger.go:42: 13:07:59 | one-pod/4-delete-data | ++ echo one-pod-haproxy logger.go:42: 13:07:59 | one-pod/4-delete-data | + run_mysql 'TRUNCATE TABLE myDB.myTable' '-h one-pod-haproxy -uroot -proot_password' logger.go:42: 13:07:59 | one-pod/4-delete-data | + local 'command=TRUNCATE TABLE myDB.myTable' logger.go:42: 13:07:59 | one-pod/4-delete-data | + local 'uri=-h one-pod-haproxy -uroot -proot_password' logger.go:42: 13:07:59 | one-pod/4-delete-data | + local pod= logger.go:42: 13:07:59 | one-pod/4-delete-data | ++ get_client_pod logger.go:42: 13:07:59 | one-pod/4-delete-data | ++ kubectl -n kuttl-test-liked-quetzal get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:08:00 | one-pod/4-delete-data | + client_pod=mysql-client logger.go:42: 13:08:00 | one-pod/4-delete-data | + wait_pod mysql-client logger.go:42: 13:08:00 | one-pod/4-delete-data | + local pod=mysql-client logger.go:42: 13:08:00 | one-pod/4-delete-data | + set +o xtrace logger.go:42: 13:08:01 | one-pod/4-delete-data | mysql-clienttrue logger.go:42: 13:08:01 | one-pod/4-delete-data | + kubectl -n kuttl-test-liked-quetzal exec mysql-client -- bash -c 'printf '\''%s\n'\'' "TRUNCATE TABLE myDB.myTable" | mysql -sN -h one-pod-haproxy -uroot -proot_password' logger.go:42: 13:08:01 | one-pod/4-delete-data | + sed -e 's/mysql: //' logger.go:42: 13:08:01 | one-pod/4-delete-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:08:02 | one-pod/4-delete-data | + : logger.go:42: 13:08:02 | one-pod/4-delete-data | ++++ get_cluster_name logger.go:42: 13:08:02 | one-pod/4-delete-data | ++++ kubectl -n kuttl-test-liked-quetzal get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 13:08:03 | one-pod/4-delete-data | +++ get_haproxy_svc one-pod logger.go:42: 13:08:03 | one-pod/4-delete-data | +++ local cluster=one-pod logger.go:42: 13:08:03 | one-pod/4-delete-data | +++ echo one-pod-haproxy logger.go:42: 13:08:03 | one-pod/4-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h one-pod-haproxy -uroot -proot_password' logger.go:42: 13:08:03 | one-pod/4-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 13:08:03 | one-pod/4-delete-data | ++ local 'uri=-h one-pod-haproxy -uroot -proot_password' logger.go:42: 13:08:03 | one-pod/4-delete-data | ++ local pod= logger.go:42: 13:08:03 | one-pod/4-delete-data | +++ get_client_pod logger.go:42: 13:08:03 | one-pod/4-delete-data | +++ kubectl -n kuttl-test-liked-quetzal get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:08:04 | one-pod/4-delete-data | ++ client_pod=mysql-client logger.go:42: 13:08:04 | one-pod/4-delete-data | ++ wait_pod mysql-client logger.go:42: 13:08:04 | one-pod/4-delete-data | ++ local pod=mysql-client logger.go:42: 13:08:04 | one-pod/4-delete-data | ++ set +o xtrace logger.go:42: 13:08:05 | one-pod/4-delete-data | mysql-clienttrue logger.go:42: 13:08:05 | one-pod/4-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 13:08:05 | one-pod/4-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:08:05 | one-pod/4-delete-data | ++ kubectl -n kuttl-test-liked-quetzal exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h one-pod-haproxy -uroot -proot_password' logger.go:42: 13:08:06 | one-pod/4-delete-data | ++ : logger.go:42: 13:08:06 | one-pod/4-delete-data | + data= logger.go:42: 13:08:06 | one-pod/4-delete-data | + kubectl create configmap -n kuttl-test-liked-quetzal 04-delete-data-minio --from-literal=data= logger.go:42: 13:08:07 | one-pod/4-delete-data | configmap/04-delete-data-minio created logger.go:42: 13:08:08 | one-pod/4-delete-data | test step completed 4-delete-data logger.go:42: 13:08:08 | one-pod/5-restore-from-minio | starting test step 5-restore-from-minio logger.go:42: 13:08:10 | one-pod/5-restore-from-minio | PerconaServerMySQLRestore:kuttl-test-liked-quetzal/one-pod-restore-minio created logger.go:42: 13:10:17 | one-pod/5-restore-from-minio | test step completed 5-restore-from-minio logger.go:42: 13:10:17 | one-pod/6-read-data | starting test step 6-read-data logger.go:42: 13:10:17 | one-pod/6-read-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions data=$(run_mysql "SELECT * FROM myDB.myTable" "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 06-read-data-minio --from-literal=data="${data}"] logger.go:42: 13:10:17 | one-pod/6-read-data | + source ../../functions logger.go:42: 13:10:17 | one-pod/6-read-data | +++ realpath ../../.. logger.go:42: 13:10:17 | one-pod/6-read-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:10:17 | one-pod/6-read-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 13:10:17 | one-pod/6-read-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:10:17 | one-pod/6-read-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:10:17 | one-pod/6-read-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:10:17 | one-pod/6-read-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:10:17 | one-pod/6-read-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:10:17 | one-pod/6-read-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:10:17 | one-pod/6-read-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:10:17 | one-pod/6-read-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:10:17 | one-pod/6-read-data | ++++ mktemp -d logger.go:42: 13:10:17 | one-pod/6-read-data | +++ export TEMP_DIR=/tmp/tmp.i0JNdjFddj logger.go:42: 13:10:17 | one-pod/6-read-data | +++ TEMP_DIR=/tmp/tmp.i0JNdjFddj logger.go:42: 13:10:17 | one-pod/6-read-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 13:10:17 | one-pod/6-read-data | +++ export GIT_BRANCH=PR-339 logger.go:42: 13:10:17 | one-pod/6-read-data | +++ GIT_BRANCH=PR-339 logger.go:42: 13:10:17 | one-pod/6-read-data | +++ export VERSION=PR-339-3a8483b logger.go:42: 13:10:17 | one-pod/6-read-data | +++ VERSION=PR-339-3a8483b logger.go:42: 13:10:17 | one-pod/6-read-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:10:17 | one-pod/6-read-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:10:17 | one-pod/6-read-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:10:17 | one-pod/6-read-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:10:17 | one-pod/6-read-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:10:17 | one-pod/6-read-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:10:17 | one-pod/6-read-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:10:17 | one-pod/6-read-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:10:17 | one-pod/6-read-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:10:17 | one-pod/6-read-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:10:17 | one-pod/6-read-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:10:17 | one-pod/6-read-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:10:17 | one-pod/6-read-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:10:17 | one-pod/6-read-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:10:17 | one-pod/6-read-data | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:10:17 | one-pod/6-read-data | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:10:17 | one-pod/6-read-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:10:17 | one-pod/6-read-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:10:17 | one-pod/6-read-data | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:10:17 | one-pod/6-read-data | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:10:17 | one-pod/6-read-data | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:10:17 | one-pod/6-read-data | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:10:17 | one-pod/6-read-data | ++++ which gdate logger.go:42: 13:10:17 | one-pod/6-read-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 13:10:17 | one-pod/6-read-data | ++++ which date logger.go:42: 13:10:17 | one-pod/6-read-data | +++ date=/usr/bin/date logger.go:42: 13:10:17 | one-pod/6-read-data | +++ command -v oc logger.go:42: 13:10:17 | one-pod/6-read-data | +++ oc get projects logger.go:42: 13:10:23 | one-pod/6-read-data | error: the server doesn't have a resource type "projects" logger.go:42: 13:10:23 | one-pod/6-read-data | +++ kubectl get nodes logger.go:42: 13:10:23 | one-pod/6-read-data | +++ grep '^minikube' logger.go:42: 13:10:24 | one-pod/6-read-data | ++++ pwd logger.go:42: 13:10:24 | one-pod/6-read-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/one-pod logger.go:42: 13:10:24 | one-pod/6-read-data | ++ test_name=one-pod logger.go:42: 13:10:24 | one-pod/6-read-data | ++++ get_cluster_name logger.go:42: 13:10:24 | one-pod/6-read-data | ++++ kubectl -n kuttl-test-liked-quetzal get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 13:10:24 | one-pod/6-read-data | +++ get_haproxy_svc one-pod logger.go:42: 13:10:24 | one-pod/6-read-data | +++ local cluster=one-pod logger.go:42: 13:10:24 | one-pod/6-read-data | +++ echo one-pod-haproxy logger.go:42: 13:10:24 | one-pod/6-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h one-pod-haproxy -uroot -proot_password' logger.go:42: 13:10:24 | one-pod/6-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 13:10:24 | one-pod/6-read-data | ++ local 'uri=-h one-pod-haproxy -uroot -proot_password' logger.go:42: 13:10:24 | one-pod/6-read-data | ++ local pod= logger.go:42: 13:10:24 | one-pod/6-read-data | +++ get_client_pod logger.go:42: 13:10:24 | one-pod/6-read-data | +++ kubectl -n kuttl-test-liked-quetzal get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:10:25 | one-pod/6-read-data | ++ client_pod=mysql-client logger.go:42: 13:10:25 | one-pod/6-read-data | ++ wait_pod mysql-client logger.go:42: 13:10:25 | one-pod/6-read-data | ++ local pod=mysql-client logger.go:42: 13:10:25 | one-pod/6-read-data | ++ set +o xtrace logger.go:42: 13:10:26 | one-pod/6-read-data | mysql-clienttrue logger.go:42: 13:10:26 | one-pod/6-read-data | ++ kubectl -n kuttl-test-liked-quetzal exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h one-pod-haproxy -uroot -proot_password' logger.go:42: 13:10:26 | one-pod/6-read-data | ++ sed -e 's/mysql: //' logger.go:42: 13:10:26 | one-pod/6-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:10:27 | one-pod/6-read-data | + data=100500 logger.go:42: 13:10:27 | one-pod/6-read-data | + kubectl create configmap -n kuttl-test-liked-quetzal 06-read-data-minio --from-literal=data=100500 logger.go:42: 13:10:28 | one-pod/6-read-data | configmap/06-read-data-minio created logger.go:42: 13:10:30 | one-pod/6-read-data | test step completed 6-read-data logger.go:42: 13:10:30 | one-pod/7-drop-finalizer | starting test step 7-drop-finalizer logger.go:42: 13:10:31 | one-pod/7-drop-finalizer | PerconaServerMySQL:kuttl-test-liked-quetzal/one-pod updated logger.go:42: 13:10:31 | one-pod/7-drop-finalizer | test step completed 7-drop-finalizer logger.go:42: 13:10:31 | one-pod | one-pod events from ns kuttl-test-liked-quetzal: logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:05:41 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-5w55k Scheduled Successfully assigned kuttl-test-liked-quetzal/percona-server-mysql-operator-6dcd78d857-5w55k to gke-jen-ps-339-3a8483b-1-default-pool-092d3615-1n52 logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:05:41 +0000 UTC Normal ReplicaSet.apps percona-server-mysql-operator-6dcd78d857 SuccessfulCreate Created pod: percona-server-mysql-operator-6dcd78d857-5w55k logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:05:41 +0000 UTC Normal Deployment.apps percona-server-mysql-operator ScalingReplicaSet Scaled up replica set percona-server-mysql-operator-6dcd78d857 to 1 logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:05:44 +0000 UTC Normal Lease.coordination.k8s.io 08db2feb.percona.com LeaderElection percona-server-mysql-operator-6dcd78d857-5w55k_d334f402-861b-4376-a1a0-519ca71d6449 became leader logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:05:44 +0000 UTC Normal Pod mysql-client Scheduled Successfully assigned kuttl-test-liked-quetzal/mysql-client to gke-jen-ps-339-3a8483b-1-default-pool-092d3615-1n52 logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:05:44 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-5w55k.spec.containers{manager} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:05:44 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-5w55k.spec.containers{manager} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 435.595529ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:05:44 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-5w55k.spec.containers{manager} Created Created container manager logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:05:44 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-5w55k.spec.containers{manager} Started Started container manager logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:05:45 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Pulled Container image "percona/percona-server:8.0.25" already present on machine logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:05:45 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Created Created container mysql-client logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:05:45 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Started Started container mysql-client logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:05:50 +0000 UTC Normal PersistentVolumeClaim minio-service WaitForFirstConsumer waiting for first consumer to be created before binding logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:05:51 +0000 UTC Normal ReplicaSet.apps minio-service-5fc4dc898 SuccessfulCreate Created pod: minio-service-5fc4dc898-64kpv logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:05:51 +0000 UTC Normal Deployment.apps minio-service ScalingReplicaSet Scaled up replica set minio-service-5fc4dc898 to 1 logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:05:51 +0000 UTC Normal PersistentVolumeClaim minio-service ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:05:51 +0000 UTC Normal PersistentVolumeClaim minio-service Provisioning External provisioner is provisioning volume for claim "kuttl-test-liked-quetzal/minio-service" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:05:54 +0000 UTC Normal PersistentVolumeClaim minio-service ProvisioningSucceeded Successfully provisioned volume pvc-40b29594-0bee-4270-8a7c-904a2924c36e logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:05:55 +0000 UTC Normal Pod minio-service-5fc4dc898-64kpv Scheduled Successfully assigned kuttl-test-liked-quetzal/minio-service-5fc4dc898-64kpv to gke-jen-ps-339-3a8483b-1-default-pool-092d3615-1n52 logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:05:59 +0000 UTC Normal Pod minio-service-5fc4dc898-64kpv SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-40b29594-0bee-4270-8a7c-904a2924c36e" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:05 +0000 UTC Normal Pod minio-service-5fc4dc898-64kpv.spec.containers{minio} Pulled Container image "minio/minio:RELEASE.2020-11-19T23-48-16Z" already present on machine logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:05 +0000 UTC Normal Pod minio-service-5fc4dc898-64kpv.spec.containers{minio} Created Created container minio logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:05 +0000 UTC Normal Pod minio-service-5fc4dc898-64kpv.spec.containers{minio} Started Started container minio logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:08 +0000 UTC Normal Pod aws-cli Scheduled Successfully assigned kuttl-test-liked-quetzal/aws-cli to gke-jen-ps-339-3a8483b-1-default-pool-092d3615-k278 logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:09 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Pulling Pulling image "perconalab/awscli" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:10 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Pulled Successfully pulled image "perconalab/awscli" in 395.6295ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:10 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Created Created container aws-cli logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:10 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Started Started container aws-cli logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:27 +0000 UTC Normal PersistentVolumeClaim datadir-one-pod-mysql-0 WaitForFirstConsumer waiting for first consumer to be created before binding logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:27 +0000 UTC Normal PersistentVolumeClaim datadir-one-pod-mysql-0 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:27 +0000 UTC Normal PersistentVolumeClaim datadir-one-pod-mysql-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-liked-quetzal/datadir-one-pod-mysql-0" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:27 +0000 UTC Normal StatefulSet.apps one-pod-mysql SuccessfulCreate create Claim datadir-one-pod-mysql-0 Pod one-pod-mysql-0 in StatefulSet one-pod-mysql success logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:27 +0000 UTC Normal StatefulSet.apps one-pod-mysql SuccessfulCreate create Pod one-pod-mysql-0 in StatefulSet one-pod-mysql successful logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:27 +0000 UTC Normal Pod one-pod-orc-0 Scheduled Successfully assigned kuttl-test-liked-quetzal/one-pod-orc-0 to gke-jen-ps-339-3a8483b-1-default-pool-092d3615-2s1q logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:27 +0000 UTC Normal StatefulSet.apps one-pod-orc SuccessfulCreate create Pod one-pod-orc-0 in StatefulSet one-pod-orc successful logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:28 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:28 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 422.160541ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:28 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:28 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:30 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:30 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 352.036818ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:30 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Created Created container orc logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:30 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Started Started container orc logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:30 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:31 +0000 UTC Normal PersistentVolumeClaim datadir-one-pod-mysql-0 ProvisioningSucceeded Successfully provisioned volume pvc-ce037db2-5b70-4c74-969f-cb6030cb298a logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:31 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 361.652068ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:31 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:31 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:32 +0000 UTC Normal Pod one-pod-mysql-0 Scheduled Successfully assigned kuttl-test-liked-quetzal/one-pod-mysql-0 to gke-jen-ps-339-3a8483b-1-default-pool-092d3615-k278 logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:33 +0000 UTC Warning Pod one-pod-mysql-0 FailedMount MountVolume.SetUp failed for volume "config" : failed to sync configmap cache: timed out waiting for the condition logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:39 +0000 UTC Normal Pod one-pod-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-ce037db2-5b70-4c74-969f-cb6030cb298a" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:41 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:41 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 415.92813ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:41 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:41 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:43 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:44 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 422.27918ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:44 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Created Created container mysql logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:44 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Started Started container mysql logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:44 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:44 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 362.753512ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:44 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:44 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:44 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:45 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 390.185869ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:45 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:06:45 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:07:19 +0000 UTC Normal Pod one-pod-haproxy-0 Scheduled Successfully assigned kuttl-test-liked-quetzal/one-pod-haproxy-0 to gke-jen-ps-339-3a8483b-1-default-pool-092d3615-k278 logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:07:19 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:07:19 +0000 UTC Normal StatefulSet.apps one-pod-haproxy SuccessfulCreate create Pod one-pod-haproxy-0 in StatefulSet one-pod-haproxy successful logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:07:20 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 397.843204ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:07:20 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:07:20 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:07:21 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:07:22 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 400.577452ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:07:22 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Created Created container haproxy logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:07:22 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Started Started container haproxy logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:07:22 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:07:22 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 408.274887ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:07:22 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:07:22 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:07:41 +0000 UTC Normal Pod xb-one-pod-minio-minio-sb8l8 Scheduled Successfully assigned kuttl-test-liked-quetzal/xb-one-pod-minio-minio-sb8l8 to gke-jen-ps-339-3a8483b-1-default-pool-092d3615-2s1q logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:07:41 +0000 UTC Normal Job.batch xb-one-pod-minio-minio SuccessfulCreate Created pod: xb-one-pod-minio-minio-sb8l8 logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:07:42 +0000 UTC Normal Pod xb-one-pod-minio-minio-sb8l8.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:07:42 +0000 UTC Normal Pod xb-one-pod-minio-minio-sb8l8.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 342.345843ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:07:42 +0000 UTC Normal Pod xb-one-pod-minio-minio-sb8l8.spec.initContainers{xtrabackup-init} Created Created container xtrabackup-init logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:07:42 +0000 UTC Normal Pod xb-one-pod-minio-minio-sb8l8.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:07:44 +0000 UTC Normal Pod xb-one-pod-minio-minio-sb8l8.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:07:44 +0000 UTC Normal Pod xb-one-pod-minio-minio-sb8l8.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 420.164312ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:07:44 +0000 UTC Normal Pod xb-one-pod-minio-minio-sb8l8.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:07:44 +0000 UTC Normal Pod xb-one-pod-minio-minio-sb8l8.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:07:51 +0000 UTC Normal Job.batch xb-one-pod-minio-minio Completed Job completed logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:08:13 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Killing Stopping container haproxy logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:08:13 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:08:13 +0000 UTC Normal StatefulSet.apps one-pod-haproxy SuccessfulDelete delete Pod one-pod-haproxy-0 in StatefulSet one-pod-haproxy successful logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:08:13 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Killing Stopping container mysql logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:08:13 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Killing Stopping container xtrabackup logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:08:13 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Killing Stopping container pt-heartbeat logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:08:13 +0000 UTC Normal StatefulSet.apps one-pod-mysql SuccessfulDelete delete Pod one-pod-mysql-0 in StatefulSet one-pod-mysql successful logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:08:13 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Killing Stopping container orc logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:08:13 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:08:13 +0000 UTC Normal StatefulSet.apps one-pod-orc SuccessfulDelete delete Pod one-pod-orc-0 in StatefulSet one-pod-orc successful logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:08:16 +0000 UTC Warning Pod one-pod-mysql-0.spec.containers{mysql} Unhealthy Readiness probe failed: 2023/08/09 13:08:16 readiness check failed: connect to db: ping database: dial tcp 10.117.161.37:33062: connect: connection refused logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:08:26 +0000 UTC Warning Service one-pod-haproxy ProcessServiceFailed error processing service "kuttl-test-liked-quetzal/one-pod-haproxy": services "one-pod-haproxy" not found logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:08:45 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-bvbvw Scheduled Successfully assigned kuttl-test-liked-quetzal/xb-restore-one-pod-restore-minio-bvbvw to gke-jen-ps-339-3a8483b-1-default-pool-092d3615-k278 logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:08:45 +0000 UTC Normal Job.batch xb-restore-one-pod-restore-minio SuccessfulCreate Created pod: xb-restore-one-pod-restore-minio-bvbvw logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:08:49 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-bvbvw SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-ce037db2-5b70-4c74-969f-cb6030cb298a" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:08:52 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-bvbvw.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:08:52 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-bvbvw.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 411.714359ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:08:52 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-bvbvw.spec.initContainers{xtrabackup-init} Created Created container xtrabackup-init logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:08:53 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-bvbvw.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:08:55 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-bvbvw.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:08:55 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-bvbvw.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 462.005081ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:08:55 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-bvbvw.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:08:55 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-bvbvw.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:16 +0000 UTC Normal Job.batch xb-restore-one-pod-restore-minio Completed Job completed logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:20 +0000 UTC Normal Pod one-pod-mysql-0 Scheduled Successfully assigned kuttl-test-liked-quetzal/one-pod-mysql-0 to gke-jen-ps-339-3a8483b-1-default-pool-092d3615-k278 logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:20 +0000 UTC Normal Pod one-pod-orc-0 Scheduled Successfully assigned kuttl-test-liked-quetzal/one-pod-orc-0 to gke-jen-ps-339-3a8483b-1-default-pool-092d3615-2s1q logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:21 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:22 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 629.236288ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:22 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:22 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:23 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:23 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 397.821457ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:23 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:23 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:24 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:25 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:25 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 378.116088ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:25 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Created Created container mysql logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:25 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Started Started container mysql logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:25 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:25 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 359.666345ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:25 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Created Created container orc logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:25 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Started Started container orc logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:25 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:25 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 384.816395ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:25 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:25 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:26 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 581.461698ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:26 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:26 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:26 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:26 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 358.824353ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:26 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:09:26 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:10:06 +0000 UTC Normal Pod one-pod-haproxy-0 Scheduled Successfully assigned kuttl-test-liked-quetzal/one-pod-haproxy-0 to gke-jen-ps-339-3a8483b-1-default-pool-092d3615-k278 logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:10:07 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:10:07 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 428.684341ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:10:07 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:10:07 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:10:09 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:10:09 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 388.044091ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:10:09 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Created Created container haproxy logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:10:09 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Started Started container haproxy logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:10:09 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:10:10 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 391.402642ms logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:10:10 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 13:10:31 | one-pod | 2023-08-09 13:10:10 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 13:10:32 | one-pod | Deleting namespace: kuttl-test-liked-quetzal === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (350.39s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/one-pod (346.98s) PASS