=== RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://34.170.82.24 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 180 seconds for each step harness.go:372: testsuite: e2e-tests/tests has 30 tests === RUN kuttl/harness === RUN kuttl/harness/one-pod === PAUSE kuttl/harness/one-pod === CONT kuttl/harness/one-pod logger.go:42: 14:42:43 | one-pod | Creating namespace: kuttl-test-popular-koala logger.go:42: 14:42:43 | one-pod/0-deploy-operator | starting test step 0-deploy-operator logger.go:42: 14:42:43 | one-pod/0-deploy-operator | running command: [sh -c set -o errexit set -o xtrace source ../../functions init_temp_dir # do this only in the first TestStep apply_s3_storage_secrets deploy_operator deploy_non_tls_cluster_secrets deploy_tls_cluster_secrets deploy_client deploy_minio] logger.go:42: 14:42:43 | one-pod/0-deploy-operator | + source ../../functions logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ realpath ../../.. logger.go:42: 14:42:43 | one-pod/0-deploy-operator | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:42:43 | one-pod/0-deploy-operator | ++++ pwd logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/one-pod logger.go:42: 14:42:43 | one-pod/0-deploy-operator | ++ test_name=one-pod logger.go:42: 14:42:43 | one-pod/0-deploy-operator | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ export TEMP_DIR=/tmp/kuttl/ps/one-pod logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ TEMP_DIR=/tmp/kuttl/ps/one-pod logger.go:42: 14:42:43 | one-pod/0-deploy-operator | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ export GIT_BRANCH=PR-523 logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ GIT_BRANCH=PR-523 logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ export VERSION=PR-523-f00253e logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ VERSION=PR-523-f00253e logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:42:43 | one-pod/0-deploy-operator | ++++ which gdate logger.go:42: 14:42:43 | one-pod/0-deploy-operator | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:42:43 | one-pod/0-deploy-operator | ++++ which date logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ date=/usr/bin/date logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ command -v oc logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ kubectl get nodes logger.go:42: 14:42:43 | one-pod/0-deploy-operator | +++ grep '^minikube' logger.go:42: 14:42:43 | one-pod/0-deploy-operator | + init_temp_dir logger.go:42: 14:42:43 | one-pod/0-deploy-operator | + rm -rf /tmp/kuttl/ps/one-pod logger.go:42: 14:42:43 | one-pod/0-deploy-operator | + mkdir -p /tmp/kuttl/ps/one-pod logger.go:42: 14:42:43 | one-pod/0-deploy-operator | + apply_s3_storage_secrets logger.go:42: 14:42:43 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-popular-koala apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf/minio-secret.yml logger.go:42: 14:42:44 | one-pod/0-deploy-operator | secret/minio-secret created logger.go:42: 14:42:44 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-popular-koala apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf/cloud-secret.yml logger.go:42: 14:42:45 | one-pod/0-deploy-operator | secret/aws-s3-secret created logger.go:42: 14:42:46 | one-pod/0-deploy-operator | secret/gcp-cs-secret created logger.go:42: 14:42:46 | one-pod/0-deploy-operator | secret/azure-secret created logger.go:42: 14:42:46 | one-pod/0-deploy-operator | + deploy_operator logger.go:42: 14:42:46 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-popular-koala apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy/crd.yaml logger.go:42: 14:42:47 | one-pod/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlbackups.ps.percona.com serverside-applied logger.go:42: 14:42:47 | one-pod/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlrestores.ps.percona.com serverside-applied logger.go:42: 14:42:49 | one-pod/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqls.ps.percona.com serverside-applied logger.go:42: 14:42:49 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-popular-koala apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy/rbac.yaml logger.go:42: 14:42:50 | one-pod/0-deploy-operator | serviceaccount/percona-server-mysql-operator created logger.go:42: 14:42:50 | one-pod/0-deploy-operator | serviceaccount/percona-server-mysql-operator-orchestrator created logger.go:42: 14:42:51 | one-pod/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 14:42:51 | one-pod/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 14:42:51 | one-pod/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 14:42:52 | one-pod/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 14:42:52 | one-pod/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 14:42:52 | one-pod/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 14:42:52 | one-pod/0-deploy-operator | + yq eval '(select(documentIndex==1).spec.template.spec.containers[] | select(.name=="manager").env[] | select(.name=="DISABLE_TELEMETRY").value) = "true"' logger.go:42: 14:42:52 | one-pod/0-deploy-operator | + yq eval '(select(documentIndex==1).spec.template.spec.containers[] | select(.name=="manager").env[] | select(.name=="LOG_LEVEL").value) = "DEBUG"' logger.go:42: 14:42:52 | one-pod/0-deploy-operator | ++ printf 'select(documentIndex==1).spec.template.spec.containers[0].image="%s"' perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:42:52 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-popular-koala apply -f - logger.go:42: 14:42:52 | one-pod/0-deploy-operator | + yq eval 'select(documentIndex==1).spec.template.spec.containers[0].image="perconalab/percona-server-mysql-operator:PR-523-f00253e"' /mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy/operator.yaml logger.go:42: 14:42:53 | one-pod/0-deploy-operator | configmap/percona-server-mysql-operator-config created logger.go:42: 14:42:54 | one-pod/0-deploy-operator | deployment.apps/percona-server-mysql-operator created logger.go:42: 14:42:54 | one-pod/0-deploy-operator | + deploy_non_tls_cluster_secrets logger.go:42: 14:42:54 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-popular-koala apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf/secrets.yaml logger.go:42: 14:42:55 | one-pod/0-deploy-operator | secret/test-secrets created logger.go:42: 14:42:55 | one-pod/0-deploy-operator | + deploy_tls_cluster_secrets logger.go:42: 14:42:55 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-popular-koala apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf/ssl-secret.yaml logger.go:42: 14:42:56 | one-pod/0-deploy-operator | secret/test-ssl created logger.go:42: 14:42:56 | one-pod/0-deploy-operator | + deploy_client logger.go:42: 14:42:56 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-popular-koala apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf/client.yaml logger.go:42: 14:42:56 | one-pod/0-deploy-operator | pod/mysql-client created logger.go:42: 14:42:56 | one-pod/0-deploy-operator | + deploy_minio logger.go:42: 14:42:56 | one-pod/0-deploy-operator | + local access_key logger.go:42: 14:42:56 | one-pod/0-deploy-operator | + local secret_key logger.go:42: 14:42:56 | one-pod/0-deploy-operator | ++ kubectl -n kuttl-test-popular-koala get secret minio-secret -o 'jsonpath={.data.AWS_ACCESS_KEY_ID}' logger.go:42: 14:42:56 | one-pod/0-deploy-operator | ++ base64 -d logger.go:42: 14:42:57 | one-pod/0-deploy-operator | + access_key=some-access-key logger.go:42: 14:42:57 | one-pod/0-deploy-operator | ++ kubectl -n kuttl-test-popular-koala get secret minio-secret -o 'jsonpath={.data.AWS_SECRET_ACCESS_KEY}' logger.go:42: 14:42:57 | one-pod/0-deploy-operator | ++ base64 -d logger.go:42: 14:42:57 | one-pod/0-deploy-operator | + secret_key=some-secret-key logger.go:42: 14:42:57 | one-pod/0-deploy-operator | + helm uninstall -n kuttl-test-popular-koala minio-service logger.go:42: 14:42:57 | one-pod/0-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 14:42:57 | one-pod/0-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 14:42:58 | one-pod/0-deploy-operator | Error: uninstall: Release not loaded: minio-service: release: not found logger.go:42: 14:42:58 | one-pod/0-deploy-operator | + : logger.go:42: 14:42:58 | one-pod/0-deploy-operator | + helm repo remove minio logger.go:42: 14:42:58 | one-pod/0-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 14:42:58 | one-pod/0-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 14:42:58 | one-pod/0-deploy-operator | Error: no repositories configured logger.go:42: 14:42:58 | one-pod/0-deploy-operator | + : logger.go:42: 14:42:58 | one-pod/0-deploy-operator | + helm repo add minio https://charts.min.io/ logger.go:42: 14:42:58 | one-pod/0-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 14:42:58 | one-pod/0-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 14:42:58 | one-pod/0-deploy-operator | "minio" has been added to your repositories logger.go:42: 14:42:58 | one-pod/0-deploy-operator | +++ printf %q some-access-key logger.go:42: 14:42:58 | one-pod/0-deploy-operator | ++ printf %q some-access-key logger.go:42: 14:42:58 | one-pod/0-deploy-operator | +++ printf %q some-secret-key logger.go:42: 14:42:58 | one-pod/0-deploy-operator | ++ printf %q some-secret-key logger.go:42: 14:42:58 | one-pod/0-deploy-operator | + retry 10 60 helm install minio-service -n kuttl-test-popular-koala --version 5.0.14 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/.minio/ --set persistence.size=2G --set securityContext.enabled=false minio/minio logger.go:42: 14:42:58 | one-pod/0-deploy-operator | + local max=10 logger.go:42: 14:42:58 | one-pod/0-deploy-operator | + local delay=60 logger.go:42: 14:42:58 | one-pod/0-deploy-operator | + shift 2 logger.go:42: 14:42:58 | one-pod/0-deploy-operator | + local n=1 logger.go:42: 14:42:58 | one-pod/0-deploy-operator | + helm install minio-service -n kuttl-test-popular-koala --version 5.0.14 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/.minio/ --set persistence.size=2G --set securityContext.enabled=false minio/minio logger.go:42: 14:42:58 | one-pod/0-deploy-operator | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 14:42:58 | one-pod/0-deploy-operator | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 14:44:09 | one-pod/0-deploy-operator | NAME: minio-service logger.go:42: 14:44:09 | one-pod/0-deploy-operator | LAST DEPLOYED: Mon Feb 5 14:42:59 2024 logger.go:42: 14:44:09 | one-pod/0-deploy-operator | NAMESPACE: kuttl-test-popular-koala logger.go:42: 14:44:09 | one-pod/0-deploy-operator | STATUS: deployed logger.go:42: 14:44:09 | one-pod/0-deploy-operator | REVISION: 1 logger.go:42: 14:44:09 | one-pod/0-deploy-operator | TEST SUITE: None logger.go:42: 14:44:09 | one-pod/0-deploy-operator | NOTES: logger.go:42: 14:44:09 | one-pod/0-deploy-operator | MinIO can be accessed via port 9000 on the following DNS name from within your cluster: logger.go:42: 14:44:09 | one-pod/0-deploy-operator | minio-service.kuttl-test-popular-koala.svc.cluster.local logger.go:42: 14:44:09 | one-pod/0-deploy-operator | logger.go:42: 14:44:09 | one-pod/0-deploy-operator | To access MinIO from localhost, run the below commands: logger.go:42: 14:44:09 | one-pod/0-deploy-operator | logger.go:42: 14:44:09 | one-pod/0-deploy-operator | 1. export POD_NAME=$(kubectl get pods --namespace kuttl-test-popular-koala -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}") logger.go:42: 14:44:09 | one-pod/0-deploy-operator | logger.go:42: 14:44:09 | one-pod/0-deploy-operator | 2. kubectl port-forward $POD_NAME 9000 --namespace kuttl-test-popular-koala logger.go:42: 14:44:09 | one-pod/0-deploy-operator | logger.go:42: 14:44:09 | one-pod/0-deploy-operator | Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/ logger.go:42: 14:44:09 | one-pod/0-deploy-operator | logger.go:42: 14:44:09 | one-pod/0-deploy-operator | You can now access MinIO server on http://localhost:9000. Follow the below steps to connect to MinIO server with mc client: logger.go:42: 14:44:09 | one-pod/0-deploy-operator | logger.go:42: 14:44:09 | one-pod/0-deploy-operator | 1. Download the MinIO mc client - https://min.io/docs/minio/linux/reference/minio-mc.html#quickstart logger.go:42: 14:44:09 | one-pod/0-deploy-operator | logger.go:42: 14:44:09 | one-pod/0-deploy-operator | 2. export MC_HOST_minio-service-local=http://$(kubectl get secret --namespace kuttl-test-popular-koala minio-service -o jsonpath="{.data.rootUser}" | base64 --decode):$(kubectl get secret --namespace kuttl-test-popular-koala minio-service -o jsonpath="{.data.rootPassword}" | base64 --decode)@localhost:9000 logger.go:42: 14:44:09 | one-pod/0-deploy-operator | logger.go:42: 14:44:09 | one-pod/0-deploy-operator | 3. mc ls minio-service-local logger.go:42: 14:44:09 | one-pod/0-deploy-operator | ++ kubectl -n kuttl-test-popular-koala get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:44:09 | one-pod/0-deploy-operator | + MINIO_POD=minio-service-57dd49b-927sg logger.go:42: 14:44:09 | one-pod/0-deploy-operator | + wait_pod minio-service-57dd49b-927sg logger.go:42: 14:44:09 | one-pod/0-deploy-operator | + local pod=minio-service-57dd49b-927sg logger.go:42: 14:44:09 | one-pod/0-deploy-operator | + set +o xtrace logger.go:42: 14:44:09 | one-pod/0-deploy-operator | minio-service-57dd49b-927sgtrue logger.go:42: 14:44:09 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-popular-koala run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- bash -c 'AWS_ACCESS_KEY_ID='\''some-access-key'\'' AWS_SECRET_ACCESS_KEY='\''some-secret-key'\'' AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url http://minio-service:9000 s3 mb s3://operator-testing' logger.go:42: 14:44:14 | one-pod/0-deploy-operator | If you don't see a command prompt, try pressing enter. logger.go:42: 14:44:19 | one-pod/0-deploy-operator | make_bucket: operator-testing logger.go:42: 14:44:23 | one-pod/0-deploy-operator | pod "aws-cli" deleted logger.go:42: 14:44:25 | one-pod/0-deploy-operator | test step completed 0-deploy-operator logger.go:42: 14:44:25 | one-pod/1-create-cluster | starting test step 1-create-cluster logger.go:42: 14:44:25 | one-pod/1-create-cluster | running command: [sh -c set -o errexit set -o xtrace source ../../functions get_cr \ | yq eval '.spec.mysql.clusterType="async"' - \ | yq eval '.spec.allowUnsafeConfigurations=true' - \ | yq eval '.spec.mysql.size=1' - \ | yq eval '.spec.proxy.haproxy.enabled=true' - \ | yq eval '.spec.proxy.haproxy.size=1' - \ | yq eval '.spec.orchestrator.enabled=true' - \ | yq eval '.spec.orchestrator.size=1' - \ | yq eval '.spec.backup.storages.minio.type="s3"' - \ | yq eval '.spec.backup.storages.minio.s3.bucket="operator-testing"' - \ | yq eval '.spec.backup.storages.minio.s3.credentialsSecret="minio-secret"' - \ | yq eval '.spec.backup.storages.minio.s3.endpointUrl="http://minio-service:9000"' - \ | yq eval '.spec.backup.storages.minio.s3.region="us-east-1"' - \ | kubectl -n "${NAMESPACE}" apply -f -] logger.go:42: 14:44:25 | one-pod/1-create-cluster | + source ../../functions logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ realpath ../../.. logger.go:42: 14:44:25 | one-pod/1-create-cluster | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:44:25 | one-pod/1-create-cluster | ++++ pwd logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/one-pod logger.go:42: 14:44:25 | one-pod/1-create-cluster | ++ test_name=one-pod logger.go:42: 14:44:25 | one-pod/1-create-cluster | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ export TEMP_DIR=/tmp/kuttl/ps/one-pod logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ TEMP_DIR=/tmp/kuttl/ps/one-pod logger.go:42: 14:44:25 | one-pod/1-create-cluster | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ export GIT_BRANCH=PR-523 logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ GIT_BRANCH=PR-523 logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ export VERSION=PR-523-f00253e logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ VERSION=PR-523-f00253e logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:44:25 | one-pod/1-create-cluster | ++++ which gdate logger.go:42: 14:44:25 | one-pod/1-create-cluster | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:44:25 | one-pod/1-create-cluster | ++++ which date logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ date=/usr/bin/date logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ command -v oc logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ kubectl get nodes logger.go:42: 14:44:25 | one-pod/1-create-cluster | +++ grep '^minikube' logger.go:42: 14:44:25 | one-pod/1-create-cluster | + get_cr logger.go:42: 14:44:25 | one-pod/1-create-cluster | + local name_suffix= logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval .spec.allowUnsafeConfigurations=true - logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval '.spec.backup.storages.minio.s3.credentialsSecret="minio-secret"' - logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval '.spec.backup.storages.minio.s3.bucket="operator-testing"' - logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval '.spec.backup.storages.minio.s3.endpointUrl="http://minio-service:9000"' - logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval '.spec.backup.storages.minio.s3.region="us-east-1"' - logger.go:42: 14:44:25 | one-pod/1-create-cluster | + kubectl -n kuttl-test-popular-koala apply -f - logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval '.spec.backup.storages.minio.type="s3"' - logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval .spec.orchestrator.size=1 - logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval .spec.orchestrator.enabled=true - logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval .spec.mysql.size=1 - logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval '.spec.upgradeOptions.apply="disabled"' - logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval .spec.proxy.haproxy.size=1 - logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval .spec.proxy.haproxy.enabled=true - logger.go:42: 14:44:25 | one-pod/1-create-cluster | + '[' -n '' ']' logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval - logger.go:42: 14:44:25 | one-pod/1-create-cluster | ++ printf '.spec.mysql.image="%s"' perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval '.spec.mysql.image="perconalab/percona-server-mysql-operator:main-psmysql"' - logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval '.spec.secretsName="test-secrets"' - logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval '.spec.sslSecretName="test-ssl"' - logger.go:42: 14:44:25 | one-pod/1-create-cluster | ++ printf '.spec.pmm.image="%s"' perconalab/pmm-client:dev-latest logger.go:42: 14:44:25 | one-pod/1-create-cluster | ++ printf '.spec.initImage="%s"' perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval '.spec.pmm.image="perconalab/pmm-client:dev-latest"' - logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval '.spec.initImage="perconalab/percona-server-mysql-operator:PR-523-f00253e"' - logger.go:42: 14:44:25 | one-pod/1-create-cluster | ++ printf '.spec.toolkit.image="%s"' perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval '.spec.toolkit.image="perconalab/percona-server-mysql-operator:main-toolkit"' - logger.go:42: 14:44:25 | one-pod/1-create-cluster | ++ printf '.spec.proxy.haproxy.image="%s"' perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval '.spec.proxy.haproxy.image="perconalab/percona-server-mysql-operator:main-haproxy"' - logger.go:42: 14:44:25 | one-pod/1-create-cluster | ++ printf '.spec.backup.image="%s"' perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval '.spec.backup.image="perconalab/percona-server-mysql-operator:main-backup"' - logger.go:42: 14:44:25 | one-pod/1-create-cluster | ++ printf '.spec.orchestrator.image="%s"' perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:44:25 | one-pod/1-create-cluster | ++ printf '.spec.proxy.router.image="%s"' perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval '.spec.orchestrator.image="perconalab/percona-server-mysql-operator:main-orchestrator"' - logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval '.spec.proxy.router.image="perconalab/percona-server-mysql-operator:main-router"' - logger.go:42: 14:44:25 | one-pod/1-create-cluster | ++ printf '.metadata.name="%s"' one-pod logger.go:42: 14:44:25 | one-pod/1-create-cluster | + yq eval '.metadata.name="one-pod"' /mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy/cr.yaml logger.go:42: 14:44:26 | one-pod/1-create-cluster | perconaservermysql.ps.percona.com/one-pod created logger.go:42: 14:46:25 | one-pod/1-create-cluster | test step completed 1-create-cluster logger.go:42: 14:46:25 | one-pod/2-write-data | starting test step 2-write-data logger.go:42: 14:46:25 | one-pod/2-write-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions sleep 5 run_mysql \ "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password" run_mysql \ "INSERT myDB.myTable (id) VALUES (100500)" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password"] logger.go:42: 14:46:25 | one-pod/2-write-data | + source ../../functions logger.go:42: 14:46:25 | one-pod/2-write-data | +++ realpath ../../.. logger.go:42: 14:46:25 | one-pod/2-write-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:46:25 | one-pod/2-write-data | ++++ pwd logger.go:42: 14:46:25 | one-pod/2-write-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/one-pod logger.go:42: 14:46:25 | one-pod/2-write-data | ++ test_name=one-pod logger.go:42: 14:46:25 | one-pod/2-write-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 14:46:25 | one-pod/2-write-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:46:25 | one-pod/2-write-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:46:25 | one-pod/2-write-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:46:25 | one-pod/2-write-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:46:25 | one-pod/2-write-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:46:25 | one-pod/2-write-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:46:25 | one-pod/2-write-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:46:25 | one-pod/2-write-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:46:25 | one-pod/2-write-data | +++ export TEMP_DIR=/tmp/kuttl/ps/one-pod logger.go:42: 14:46:25 | one-pod/2-write-data | +++ TEMP_DIR=/tmp/kuttl/ps/one-pod logger.go:42: 14:46:25 | one-pod/2-write-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:46:25 | one-pod/2-write-data | +++ export GIT_BRANCH=PR-523 logger.go:42: 14:46:25 | one-pod/2-write-data | +++ GIT_BRANCH=PR-523 logger.go:42: 14:46:25 | one-pod/2-write-data | +++ export VERSION=PR-523-f00253e logger.go:42: 14:46:25 | one-pod/2-write-data | +++ VERSION=PR-523-f00253e logger.go:42: 14:46:25 | one-pod/2-write-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:46:25 | one-pod/2-write-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:46:25 | one-pod/2-write-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:46:25 | one-pod/2-write-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:46:25 | one-pod/2-write-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:46:25 | one-pod/2-write-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:46:25 | one-pod/2-write-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:46:25 | one-pod/2-write-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:46:25 | one-pod/2-write-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:46:25 | one-pod/2-write-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:46:25 | one-pod/2-write-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:46:25 | one-pod/2-write-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:46:25 | one-pod/2-write-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:46:25 | one-pod/2-write-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:46:25 | one-pod/2-write-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:46:25 | one-pod/2-write-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:46:25 | one-pod/2-write-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:46:25 | one-pod/2-write-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:46:25 | one-pod/2-write-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:46:25 | one-pod/2-write-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:46:25 | one-pod/2-write-data | ++++ which gdate logger.go:42: 14:46:25 | one-pod/2-write-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:46:25 | one-pod/2-write-data | ++++ which date logger.go:42: 14:46:25 | one-pod/2-write-data | +++ date=/usr/bin/date logger.go:42: 14:46:25 | one-pod/2-write-data | +++ command -v oc logger.go:42: 14:46:25 | one-pod/2-write-data | +++ kubectl get nodes logger.go:42: 14:46:25 | one-pod/2-write-data | +++ grep '^minikube' logger.go:42: 14:46:26 | one-pod/2-write-data | + sleep 5 logger.go:42: 14:46:31 | one-pod/2-write-data | +++ get_cluster_name logger.go:42: 14:46:31 | one-pod/2-write-data | +++ kubectl -n kuttl-test-popular-koala get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:46:31 | one-pod/2-write-data | ++ get_haproxy_svc one-pod logger.go:42: 14:46:31 | one-pod/2-write-data | ++ local cluster=one-pod logger.go:42: 14:46:31 | one-pod/2-write-data | ++ echo one-pod-haproxy logger.go:42: 14:46:31 | one-pod/2-write-data | + run_mysql 'CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' '-h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:46:31 | one-pod/2-write-data | + local 'command=CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' logger.go:42: 14:46:31 | one-pod/2-write-data | + local 'uri=-h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:46:31 | one-pod/2-write-data | + local pod= logger.go:42: 14:46:31 | one-pod/2-write-data | ++ get_client_pod logger.go:42: 14:46:31 | one-pod/2-write-data | ++ kubectl -n kuttl-test-popular-koala get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:46:31 | one-pod/2-write-data | + client_pod=mysql-client logger.go:42: 14:46:31 | one-pod/2-write-data | + wait_pod mysql-client logger.go:42: 14:46:31 | one-pod/2-write-data | + local pod=mysql-client logger.go:42: 14:46:31 | one-pod/2-write-data | + set +o xtrace logger.go:42: 14:46:32 | one-pod/2-write-data | mysql-clienttrue logger.go:42: 14:46:32 | one-pod/2-write-data | + kubectl -n kuttl-test-popular-koala exec mysql-client -- bash -c 'printf '\''%s\n'\'' "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" | mysql -sN -h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:46:32 | one-pod/2-write-data | + sed -e 's/mysql: //' logger.go:42: 14:46:32 | one-pod/2-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:46:33 | one-pod/2-write-data | + : logger.go:42: 14:46:33 | one-pod/2-write-data | +++ get_cluster_name logger.go:42: 14:46:33 | one-pod/2-write-data | +++ kubectl -n kuttl-test-popular-koala get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:46:34 | one-pod/2-write-data | ++ get_haproxy_svc one-pod logger.go:42: 14:46:34 | one-pod/2-write-data | ++ local cluster=one-pod logger.go:42: 14:46:34 | one-pod/2-write-data | ++ echo one-pod-haproxy logger.go:42: 14:46:34 | one-pod/2-write-data | + run_mysql 'INSERT myDB.myTable (id) VALUES (100500)' '-h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:46:34 | one-pod/2-write-data | + local 'command=INSERT myDB.myTable (id) VALUES (100500)' logger.go:42: 14:46:34 | one-pod/2-write-data | + local 'uri=-h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:46:34 | one-pod/2-write-data | + local pod= logger.go:42: 14:46:34 | one-pod/2-write-data | ++ get_client_pod logger.go:42: 14:46:34 | one-pod/2-write-data | ++ kubectl -n kuttl-test-popular-koala get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:46:34 | one-pod/2-write-data | + client_pod=mysql-client logger.go:42: 14:46:34 | one-pod/2-write-data | + wait_pod mysql-client logger.go:42: 14:46:34 | one-pod/2-write-data | + local pod=mysql-client logger.go:42: 14:46:34 | one-pod/2-write-data | + set +o xtrace logger.go:42: 14:46:34 | one-pod/2-write-data | mysql-clienttrue logger.go:42: 14:46:34 | one-pod/2-write-data | + kubectl -n kuttl-test-popular-koala exec mysql-client -- bash -c 'printf '\''%s\n'\'' "INSERT myDB.myTable (id) VALUES (100500)" | mysql -sN -h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:46:34 | one-pod/2-write-data | + sed -e 's/mysql: //' logger.go:42: 14:46:34 | one-pod/2-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:46:36 | one-pod/2-write-data | + : logger.go:42: 14:46:36 | one-pod/2-write-data | test step completed 2-write-data logger.go:42: 14:46:36 | one-pod/3-create-backup-minio | starting test step 3-create-backup-minio logger.go:42: 14:46:38 | one-pod/3-create-backup-minio | PerconaServerMySQLBackup:kuttl-test-popular-koala/one-pod-minio created logger.go:42: 14:46:53 | one-pod/3-create-backup-minio | test step completed 3-create-backup-minio logger.go:42: 14:46:53 | one-pod/4-delete-data | starting test step 4-delete-data logger.go:42: 14:46:53 | one-pod/4-delete-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "TRUNCATE TABLE myDB.myTable" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password" data=$(run_mysql "SELECT * FROM myDB.myTable" "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 04-delete-data-minio --from-literal=data="${data}"] logger.go:42: 14:46:53 | one-pod/4-delete-data | + source ../../functions logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ realpath ../../.. logger.go:42: 14:46:53 | one-pod/4-delete-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:46:53 | one-pod/4-delete-data | ++++ pwd logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/one-pod logger.go:42: 14:46:53 | one-pod/4-delete-data | ++ test_name=one-pod logger.go:42: 14:46:53 | one-pod/4-delete-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ export TEMP_DIR=/tmp/kuttl/ps/one-pod logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ TEMP_DIR=/tmp/kuttl/ps/one-pod logger.go:42: 14:46:53 | one-pod/4-delete-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ export GIT_BRANCH=PR-523 logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ GIT_BRANCH=PR-523 logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ export VERSION=PR-523-f00253e logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ VERSION=PR-523-f00253e logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:46:53 | one-pod/4-delete-data | ++++ which gdate logger.go:42: 14:46:53 | one-pod/4-delete-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:46:53 | one-pod/4-delete-data | ++++ which date logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ date=/usr/bin/date logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ command -v oc logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ kubectl get nodes logger.go:42: 14:46:53 | one-pod/4-delete-data | +++ grep '^minikube' logger.go:42: 14:46:54 | one-pod/4-delete-data | +++ get_cluster_name logger.go:42: 14:46:54 | one-pod/4-delete-data | +++ kubectl -n kuttl-test-popular-koala get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:46:54 | one-pod/4-delete-data | ++ get_haproxy_svc one-pod logger.go:42: 14:46:54 | one-pod/4-delete-data | ++ local cluster=one-pod logger.go:42: 14:46:54 | one-pod/4-delete-data | ++ echo one-pod-haproxy logger.go:42: 14:46:54 | one-pod/4-delete-data | + run_mysql 'TRUNCATE TABLE myDB.myTable' '-h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:46:54 | one-pod/4-delete-data | + local 'command=TRUNCATE TABLE myDB.myTable' logger.go:42: 14:46:54 | one-pod/4-delete-data | + local 'uri=-h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:46:54 | one-pod/4-delete-data | + local pod= logger.go:42: 14:46:54 | one-pod/4-delete-data | ++ get_client_pod logger.go:42: 14:46:54 | one-pod/4-delete-data | ++ kubectl -n kuttl-test-popular-koala get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:46:55 | one-pod/4-delete-data | + client_pod=mysql-client logger.go:42: 14:46:55 | one-pod/4-delete-data | + wait_pod mysql-client logger.go:42: 14:46:55 | one-pod/4-delete-data | + local pod=mysql-client logger.go:42: 14:46:55 | one-pod/4-delete-data | + set +o xtrace logger.go:42: 14:46:55 | one-pod/4-delete-data | mysql-clienttrue logger.go:42: 14:46:55 | one-pod/4-delete-data | + sed -e 's/mysql: //' logger.go:42: 14:46:55 | one-pod/4-delete-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:46:55 | one-pod/4-delete-data | + kubectl -n kuttl-test-popular-koala exec mysql-client -- bash -c 'printf '\''%s\n'\'' "TRUNCATE TABLE myDB.myTable" | mysql -sN -h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:46:56 | one-pod/4-delete-data | + : logger.go:42: 14:46:56 | one-pod/4-delete-data | ++++ get_cluster_name logger.go:42: 14:46:56 | one-pod/4-delete-data | ++++ kubectl -n kuttl-test-popular-koala get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:46:57 | one-pod/4-delete-data | +++ get_haproxy_svc one-pod logger.go:42: 14:46:57 | one-pod/4-delete-data | +++ local cluster=one-pod logger.go:42: 14:46:57 | one-pod/4-delete-data | +++ echo one-pod-haproxy logger.go:42: 14:46:57 | one-pod/4-delete-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:46:57 | one-pod/4-delete-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:46:57 | one-pod/4-delete-data | ++ local 'uri=-h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:46:57 | one-pod/4-delete-data | ++ local pod= logger.go:42: 14:46:57 | one-pod/4-delete-data | +++ get_client_pod logger.go:42: 14:46:57 | one-pod/4-delete-data | +++ kubectl -n kuttl-test-popular-koala get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:46:57 | one-pod/4-delete-data | ++ client_pod=mysql-client logger.go:42: 14:46:57 | one-pod/4-delete-data | ++ wait_pod mysql-client logger.go:42: 14:46:57 | one-pod/4-delete-data | ++ local pod=mysql-client logger.go:42: 14:46:57 | one-pod/4-delete-data | ++ set +o xtrace logger.go:42: 14:46:58 | one-pod/4-delete-data | mysql-clienttrue logger.go:42: 14:46:58 | one-pod/4-delete-data | ++ kubectl -n kuttl-test-popular-koala exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:46:58 | one-pod/4-delete-data | ++ sed -e 's/mysql: //' logger.go:42: 14:46:58 | one-pod/4-delete-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:46:59 | one-pod/4-delete-data | ++ : logger.go:42: 14:46:59 | one-pod/4-delete-data | + data= logger.go:42: 14:46:59 | one-pod/4-delete-data | + kubectl create configmap -n kuttl-test-popular-koala 04-delete-data-minio --from-literal=data= logger.go:42: 14:46:59 | one-pod/4-delete-data | configmap/04-delete-data-minio created logger.go:42: 14:47:00 | one-pod/4-delete-data | test step completed 4-delete-data logger.go:42: 14:47:00 | one-pod/5-restore-from-minio | starting test step 5-restore-from-minio logger.go:42: 14:47:01 | one-pod/5-restore-from-minio | PerconaServerMySQLRestore:kuttl-test-popular-koala/one-pod-restore-minio created logger.go:42: 14:49:42 | one-pod/5-restore-from-minio | test step completed 5-restore-from-minio logger.go:42: 14:49:42 | one-pod/6-read-data | starting test step 6-read-data logger.go:42: 14:49:42 | one-pod/6-read-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions sleep 5 data=$(run_mysql "SELECT * FROM myDB.myTable" "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 06-read-data-minio --from-literal=data="${data}"] logger.go:42: 14:49:42 | one-pod/6-read-data | + source ../../functions logger.go:42: 14:49:42 | one-pod/6-read-data | +++ realpath ../../.. logger.go:42: 14:49:42 | one-pod/6-read-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:49:42 | one-pod/6-read-data | ++++ pwd logger.go:42: 14:49:42 | one-pod/6-read-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/one-pod logger.go:42: 14:49:42 | one-pod/6-read-data | ++ test_name=one-pod logger.go:42: 14:49:42 | one-pod/6-read-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 14:49:42 | one-pod/6-read-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:49:42 | one-pod/6-read-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:49:42 | one-pod/6-read-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:49:42 | one-pod/6-read-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:49:42 | one-pod/6-read-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:49:42 | one-pod/6-read-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:49:42 | one-pod/6-read-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:49:42 | one-pod/6-read-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:49:42 | one-pod/6-read-data | +++ export TEMP_DIR=/tmp/kuttl/ps/one-pod logger.go:42: 14:49:42 | one-pod/6-read-data | +++ TEMP_DIR=/tmp/kuttl/ps/one-pod logger.go:42: 14:49:42 | one-pod/6-read-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:49:42 | one-pod/6-read-data | +++ export GIT_BRANCH=PR-523 logger.go:42: 14:49:42 | one-pod/6-read-data | +++ GIT_BRANCH=PR-523 logger.go:42: 14:49:42 | one-pod/6-read-data | +++ export VERSION=PR-523-f00253e logger.go:42: 14:49:42 | one-pod/6-read-data | +++ VERSION=PR-523-f00253e logger.go:42: 14:49:42 | one-pod/6-read-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:49:42 | one-pod/6-read-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:49:42 | one-pod/6-read-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:49:42 | one-pod/6-read-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:49:42 | one-pod/6-read-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:49:42 | one-pod/6-read-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:49:42 | one-pod/6-read-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:49:42 | one-pod/6-read-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:49:42 | one-pod/6-read-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:49:42 | one-pod/6-read-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:49:42 | one-pod/6-read-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:49:42 | one-pod/6-read-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:49:42 | one-pod/6-read-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:49:42 | one-pod/6-read-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:49:42 | one-pod/6-read-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:49:42 | one-pod/6-read-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:49:42 | one-pod/6-read-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:49:42 | one-pod/6-read-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:49:42 | one-pod/6-read-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:49:42 | one-pod/6-read-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:49:42 | one-pod/6-read-data | ++++ which gdate logger.go:42: 14:49:42 | one-pod/6-read-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:49:42 | one-pod/6-read-data | ++++ which date logger.go:42: 14:49:42 | one-pod/6-read-data | +++ date=/usr/bin/date logger.go:42: 14:49:42 | one-pod/6-read-data | +++ command -v oc logger.go:42: 14:49:42 | one-pod/6-read-data | +++ kubectl get nodes logger.go:42: 14:49:42 | one-pod/6-read-data | +++ grep '^minikube' logger.go:42: 14:49:42 | one-pod/6-read-data | + sleep 5 logger.go:42: 14:49:47 | one-pod/6-read-data | ++++ get_cluster_name logger.go:42: 14:49:47 | one-pod/6-read-data | ++++ kubectl -n kuttl-test-popular-koala get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:49:48 | one-pod/6-read-data | +++ get_haproxy_svc one-pod logger.go:42: 14:49:48 | one-pod/6-read-data | +++ local cluster=one-pod logger.go:42: 14:49:48 | one-pod/6-read-data | +++ echo one-pod-haproxy logger.go:42: 14:49:48 | one-pod/6-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:49:48 | one-pod/6-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:49:48 | one-pod/6-read-data | ++ local 'uri=-h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:49:48 | one-pod/6-read-data | ++ local pod= logger.go:42: 14:49:48 | one-pod/6-read-data | +++ get_client_pod logger.go:42: 14:49:48 | one-pod/6-read-data | +++ kubectl -n kuttl-test-popular-koala get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:49:48 | one-pod/6-read-data | ++ client_pod=mysql-client logger.go:42: 14:49:48 | one-pod/6-read-data | ++ wait_pod mysql-client logger.go:42: 14:49:48 | one-pod/6-read-data | ++ local pod=mysql-client logger.go:42: 14:49:48 | one-pod/6-read-data | ++ set +o xtrace logger.go:42: 14:49:48 | one-pod/6-read-data | mysql-clienttrue logger.go:42: 14:49:48 | one-pod/6-read-data | ++ kubectl -n kuttl-test-popular-koala exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h one-pod-haproxy -uroot -proot_password' logger.go:42: 14:49:48 | one-pod/6-read-data | ++ sed -e 's/mysql: //' logger.go:42: 14:49:48 | one-pod/6-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:49:50 | one-pod/6-read-data | + data=100500 logger.go:42: 14:49:50 | one-pod/6-read-data | + kubectl create configmap -n kuttl-test-popular-koala 06-read-data-minio --from-literal=data=100500 logger.go:42: 14:49:50 | one-pod/6-read-data | configmap/06-read-data-minio created logger.go:42: 14:49:52 | one-pod/6-read-data | test step completed 6-read-data logger.go:42: 14:49:52 | one-pod/99-drop-finalizer | starting test step 99-drop-finalizer logger.go:42: 14:49:53 | one-pod/99-drop-finalizer | PerconaServerMySQL:kuttl-test-popular-koala/one-pod updated logger.go:42: 14:49:53 | one-pod/99-drop-finalizer | test step completed 99-drop-finalizer logger.go:42: 14:49:53 | one-pod | one-pod events from ns kuttl-test-popular-koala: logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:42:54 +0000 UTC Normal Pod percona-server-mysql-operator-b4c599bbb-pcqpf Scheduled Successfully assigned kuttl-test-popular-koala/percona-server-mysql-operator-b4c599bbb-pcqpf to gke-jen-ps-523-f00253e-4-default-pool-c341acbb-fvd9 default-scheduler logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:42:54 +0000 UTC Normal ReplicaSet.apps percona-server-mysql-operator-b4c599bbb SuccessfulCreate Created pod: percona-server-mysql-operator-b4c599bbb-pcqpf replicaset-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:42:54 +0000 UTC Normal Deployment.apps percona-server-mysql-operator ScalingReplicaSet Scaled up replica set percona-server-mysql-operator-b4c599bbb to 1 deployment-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:42:56 +0000 UTC Normal Lease.coordination.k8s.io 08db2feb.percona.com LeaderElection percona-server-mysql-operator-b4c599bbb-pcqpf_deff7bde-da14-455d-8392-ca58668f21cc became leader percona-server-mysql-operator-b4c599bbb-pcqpf_deff7bde-da14-455d-8392-ca58668f21cc logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:42:56 +0000 UTC Normal Pod mysql-client Scheduled Successfully assigned kuttl-test-popular-koala/mysql-client to gke-jen-ps-523-f00253e-4-default-pool-c341acbb-fvd9 default-scheduler logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:42:56 +0000 UTC Normal Pod percona-server-mysql-operator-b4c599bbb-pcqpf.spec.containers{manager} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:42:56 +0000 UTC Normal Pod percona-server-mysql-operator-b4c599bbb-pcqpf.spec.containers{manager} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 113.29979ms (113.325645ms including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:42:56 +0000 UTC Normal Pod percona-server-mysql-operator-b4c599bbb-pcqpf.spec.containers{manager} Created Created container manager kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:42:56 +0000 UTC Normal Pod percona-server-mysql-operator-b4c599bbb-pcqpf.spec.containers{manager} Started Started container manager kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:42:57 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Pulled Container image "percona/percona-server:8.0.33" already present on machine kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:42:57 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Created Created container mysql-client kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:42:57 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Started Started container mysql-client kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:43:04 +0000 UTC Normal PersistentVolumeClaim minio-service WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:43:05 +0000 UTC Normal ReplicaSet.apps minio-service-57dd49b SuccessfulCreate Created pod: minio-service-57dd49b-927sg replicaset-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:43:05 +0000 UTC Normal Deployment.apps minio-service ScalingReplicaSet Scaled up replica set minio-service-57dd49b to 1 deployment-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:43:05 +0000 UTC Normal PersistentVolumeClaim minio-service Provisioning External provisioner is provisioning volume for claim "kuttl-test-popular-koala/minio-service" pd.csi.storage.gke.io_gke-3cf2883585f342528d0d-3b6b-72f9-vm_3d282452-5427-4f7b-a92f-30723abd1b96 logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:43:05 +0000 UTC Normal PersistentVolumeClaim minio-service ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:43:06 +0000 UTC Normal Pod minio-service-post-job-96pmw Scheduled Successfully assigned kuttl-test-popular-koala/minio-service-post-job-96pmw to gke-jen-ps-523-f00253e-4-default-pool-c341acbb-fvd9 default-scheduler logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:43:06 +0000 UTC Normal Job.batch minio-service-post-job SuccessfulCreate Created pod: minio-service-post-job-96pmw job-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:43:07 +0000 UTC Normal Pod minio-service-post-job-96pmw.spec.containers{minio-make-user} Pulling Pulling image "quay.io/minio/mc:RELEASE.2023-09-29T16-41-22Z" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:43:09 +0000 UTC Normal PersistentVolumeClaim minio-service ProvisioningSucceeded Successfully provisioned volume pvc-adbad33a-b5d7-4207-8de6-c30a39e525a2 pd.csi.storage.gke.io_gke-3cf2883585f342528d0d-3b6b-72f9-vm_3d282452-5427-4f7b-a92f-30723abd1b96 logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:43:10 +0000 UTC Normal Pod minio-service-57dd49b-927sg Scheduled Successfully assigned kuttl-test-popular-koala/minio-service-57dd49b-927sg to gke-jen-ps-523-f00253e-4-default-pool-c341acbb-fvd9 default-scheduler logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:43:10 +0000 UTC Normal Pod minio-service-post-job-96pmw.spec.containers{minio-make-user} Pulled Successfully pulled image "quay.io/minio/mc:RELEASE.2023-09-29T16-41-22Z" in 3.564674207s (3.564688901s including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:43:10 +0000 UTC Normal Pod minio-service-post-job-96pmw.spec.containers{minio-make-user} Created Created container minio-make-user kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:43:10 +0000 UTC Normal Pod minio-service-post-job-96pmw.spec.containers{minio-make-user} Started Started container minio-make-user kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:43:17 +0000 UTC Normal Pod minio-service-57dd49b-927sg SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-adbad33a-b5d7-4207-8de6-c30a39e525a2" attachdetach-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:43:19 +0000 UTC Normal Pod minio-service-57dd49b-927sg.spec.containers{minio} Pulling Pulling image "quay.io/minio/minio:RELEASE.2023-09-30T07-02-29Z" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:43:23 +0000 UTC Normal Pod minio-service-57dd49b-927sg.spec.containers{minio} Pulled Successfully pulled image "quay.io/minio/minio:RELEASE.2023-09-30T07-02-29Z" in 3.489772901s (3.489793461s including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:43:23 +0000 UTC Normal Pod minio-service-57dd49b-927sg.spec.containers{minio} Created Created container minio kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:43:23 +0000 UTC Normal Pod minio-service-57dd49b-927sg.spec.containers{minio} Started Started container minio kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:08 +0000 UTC Normal Job.batch minio-service-post-job Completed Job completed job-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:10 +0000 UTC Normal Pod aws-cli Scheduled Successfully assigned kuttl-test-popular-koala/aws-cli to gke-jen-ps-523-f00253e-4-default-pool-c341acbb-fvd9 default-scheduler logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:10 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Pulling Pulling image "perconalab/awscli" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:13 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Pulled Successfully pulled image "perconalab/awscli" in 3.005558636s (3.005572606s including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:14 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Created Created container aws-cli kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:14 +0000 UTC Normal Pod aws-cli.spec.containers{aws-cli} Started Started container aws-cli kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:33 +0000 UTC Normal PersistentVolumeClaim datadir-one-pod-mysql-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:33 +0000 UTC Normal PersistentVolumeClaim datadir-one-pod-mysql-0 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:33 +0000 UTC Normal PersistentVolumeClaim datadir-one-pod-mysql-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-popular-koala/datadir-one-pod-mysql-0" pd.csi.storage.gke.io_gke-3cf2883585f342528d0d-3b6b-72f9-vm_3d282452-5427-4f7b-a92f-30723abd1b96 logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:33 +0000 UTC Normal StatefulSet.apps one-pod-mysql SuccessfulCreate create Claim datadir-one-pod-mysql-0 Pod one-pod-mysql-0 in StatefulSet one-pod-mysql success statefulset-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:33 +0000 UTC Normal StatefulSet.apps one-pod-mysql SuccessfulCreate create Pod one-pod-mysql-0 in StatefulSet one-pod-mysql successful statefulset-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:37 +0000 UTC Normal PersistentVolumeClaim datadir-one-pod-mysql-0 ProvisioningSucceeded Successfully provisioned volume pvc-67a3250b-8527-443c-a796-eaef2ef6481c pd.csi.storage.gke.io_gke-3cf2883585f342528d0d-3b6b-72f9-vm_3d282452-5427-4f7b-a92f-30723abd1b96 logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:37 +0000 UTC Normal Pod one-pod-mysql-0 Scheduled Successfully assigned kuttl-test-popular-koala/one-pod-mysql-0 to gke-jen-ps-523-f00253e-4-default-pool-c341acbb-fvd9 default-scheduler logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:37 +0000 UTC Normal Pod one-pod-orc-0 Scheduled Successfully assigned kuttl-test-popular-koala/one-pod-orc-0 to gke-jen-ps-523-f00253e-4-default-pool-c341acbb-nkmv default-scheduler logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:37 +0000 UTC Normal StatefulSet.apps one-pod-orc SuccessfulCreate create Pod one-pod-orc-0 in StatefulSet one-pod-orc successful statefulset-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:38 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:38 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 89.754789ms (89.764299ms including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:38 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orc-init} Created Created container orc-init kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:38 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orc-init} Started Started container orc-init kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:40 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:41 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 1.359067317s (1.359081229s including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:41 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Created Created container orc kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:41 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Started Started container orc kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:41 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:41 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 80.053896ms (80.070636ms including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:41 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:41 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:44 +0000 UTC Normal Pod one-pod-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-67a3250b-8527-443c-a796-eaef2ef6481c" attachdetach-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:46 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:46 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 108.065327ms (108.073139ms including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:46 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:46 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:48 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:48 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 86.967966ms (86.98091ms including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:48 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:48 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:48 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:48 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 95.17227ms (95.187194ms including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:48 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:48 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:48 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:58 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 10.148848093s (10.148861182s including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:58 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Created Created container pt-heartbeat kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:44:59 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Started Started container pt-heartbeat kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:45:16 +0000 UTC Warning Pod one-pod-mysql-0.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:45:16 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:45:21 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 113.63337ms (113.649318ms including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:45:21 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 96.404656ms (96.412283ms including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:45:51 +0000 UTC Normal Pod one-pod-haproxy-0 Scheduled Successfully assigned kuttl-test-popular-koala/one-pod-haproxy-0 to gke-jen-ps-523-f00253e-4-default-pool-c341acbb-w471 default-scheduler logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:45:51 +0000 UTC Normal StatefulSet.apps one-pod-haproxy SuccessfulCreate create Pod one-pod-haproxy-0 in StatefulSet one-pod-haproxy successful statefulset-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:45:52 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:46:06 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 14.441059829s (14.441079586s including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:46:06 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Created Created container haproxy-init kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:46:07 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Started Started container haproxy-init kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:46:15 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:46:18 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 2.998981751s (2.998998774s including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:46:18 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Created Created container haproxy kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:46:18 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Started Started container haproxy kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:46:18 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:46:18 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 85.714338ms (85.747902ms including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:46:18 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:46:18 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:46:38 +0000 UTC Normal Pod xb-one-pod-minio-minio-v5mvr Scheduled Successfully assigned kuttl-test-popular-koala/xb-one-pod-minio-minio-v5mvr to gke-jen-ps-523-f00253e-4-default-pool-c341acbb-nkmv default-scheduler logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:46:38 +0000 UTC Normal Pod xb-one-pod-minio-minio-v5mvr.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:46:38 +0000 UTC Normal Job.batch xb-one-pod-minio-minio SuccessfulCreate Created pod: xb-one-pod-minio-minio-v5mvr job-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:46:39 +0000 UTC Normal Pod xb-one-pod-minio-minio-v5mvr.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 97.886125ms (97.905645ms including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:46:39 +0000 UTC Normal Pod xb-one-pod-minio-minio-v5mvr.spec.initContainers{xtrabackup-init} Created Created container xtrabackup-init kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:46:39 +0000 UTC Normal Pod xb-one-pod-minio-minio-v5mvr.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:46:40 +0000 UTC Normal Pod xb-one-pod-minio-minio-v5mvr.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:46:40 +0000 UTC Normal Pod xb-one-pod-minio-minio-v5mvr.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 78.210298ms (78.223038ms including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:46:40 +0000 UTC Normal Pod xb-one-pod-minio-minio-v5mvr.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:46:40 +0000 UTC Normal Pod xb-one-pod-minio-minio-v5mvr.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:46:48 +0000 UTC Normal Job.batch xb-one-pod-minio-minio Completed Job completed job-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:47:05 +0000 UTC Normal StatefulSet.apps one-pod-mysql SuccessfulDelete delete Pod one-pod-mysql-0 in StatefulSet one-pod-mysql successful statefulset-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:47:05 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Killing Stopping container orc kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:47:05 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Killing Stopping container mysql-monit kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:47:05 +0000 UTC Normal StatefulSet.apps one-pod-orc SuccessfulDelete delete Pod one-pod-orc-0 in StatefulSet one-pod-orc successful statefulset-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:47:06 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Killing Stopping container haproxy kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:47:06 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Killing Stopping container mysql-monit kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:47:06 +0000 UTC Normal StatefulSet.apps one-pod-haproxy SuccessfulDelete delete Pod one-pod-haproxy-0 in StatefulSet one-pod-haproxy successful statefulset-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:47:06 +0000 UTC Warning Endpoints one-pod-haproxy FailedToUpdateEndpoint Failed to update endpoint kuttl-test-popular-koala/one-pod-haproxy: Operation cannot be fulfilled on endpoints "one-pod-haproxy": StorageError: invalid object, Code: 4, Key: /registry/services/endpoints/kuttl-test-popular-koala/one-pod-haproxy, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 3e4ab5ab-58b7-4a53-aedf-1606dee9049a, UID in object meta: endpoint-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:47:37 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-j66vp Scheduled Successfully assigned kuttl-test-popular-koala/xb-restore-one-pod-restore-minio-j66vp to gke-jen-ps-523-f00253e-4-default-pool-c341acbb-fvd9 default-scheduler logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:47:37 +0000 UTC Normal Job.batch xb-restore-one-pod-restore-minio SuccessfulCreate Created pod: xb-restore-one-pod-restore-minio-j66vp job-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:47:45 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-j66vp SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-67a3250b-8527-443c-a796-eaef2ef6481c" attachdetach-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:47:47 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-j66vp.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:47:47 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-j66vp.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 115.314912ms (115.330774ms including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:47:47 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-j66vp.spec.initContainers{xtrabackup-init} Created Created container xtrabackup-init kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:47:47 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-j66vp.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:47:48 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-j66vp.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:47:49 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-j66vp.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 83.959467ms (83.975969ms including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:47:49 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-j66vp.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:47:49 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-j66vp.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:48:08 +0000 UTC Normal Job.batch xb-restore-one-pod-restore-minio Completed Job completed job-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:48:12 +0000 UTC Normal Pod one-pod-mysql-0 Scheduled Successfully assigned kuttl-test-popular-koala/one-pod-mysql-0 to gke-jen-ps-523-f00253e-4-default-pool-c341acbb-w471 default-scheduler logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:48:12 +0000 UTC Normal Pod one-pod-orc-0 Scheduled Successfully assigned kuttl-test-popular-koala/one-pod-orc-0 to gke-jen-ps-523-f00253e-4-default-pool-c341acbb-fvd9 default-scheduler logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:48:12 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:48:12 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 107.229653ms (107.258878ms including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:48:12 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orc-init} Created Created container orc-init kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:48:13 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orc-init} Started Started container orc-init kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:48:15 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:48:16 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 1.423764297s (1.423779799s including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:48:16 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Created Created container orc kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:48:16 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orc} Started Started container orc kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:48:16 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:48:16 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 86.200262ms (86.207148ms including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:48:16 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:48:16 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:48:33 +0000 UTC Normal Pod one-pod-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-67a3250b-8527-443c-a796-eaef2ef6481c" attachdetach-controller logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:48:35 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:48:35 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 162.456631ms (162.4715ms including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:48:35 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:48:35 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:48:36 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:48:57 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 20.263915722s (20.26393124s including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:48:57 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:48:57 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:48:57 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:49:07 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 10.476129936s (10.476144696s including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:49:08 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:49:08 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:49:08 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:49:17 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 8.756130645s (8.756145013s including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:49:17 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Created Created container pt-heartbeat kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:49:17 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Started Started container pt-heartbeat kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:49:33 +0000 UTC Normal Pod one-pod-haproxy-0 Scheduled Successfully assigned kuttl-test-popular-koala/one-pod-haproxy-0 to gke-jen-ps-523-f00253e-4-default-pool-c341acbb-fvd9 default-scheduler logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:49:34 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:49:34 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 102.498915ms (102.528233ms including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:49:34 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Created Created container haproxy-init kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:49:34 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Started Started container haproxy-init kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:49:36 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:49:39 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 3.219273416s (3.219309915s including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:49:39 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Created Created container haproxy kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:49:39 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Started Started container haproxy kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:49:39 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:49:39 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 83.030776ms (83.040474ms including waiting) kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:49:39 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 14:49:53 | one-pod | 2024-02-05 14:49:39 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 14:49:53 | one-pod | Deleting namespace: kuttl-test-popular-koala === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (471.53s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/one-pod (468.84s) PASS