=== RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://35.202.81.226 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 180 seconds for each step harness.go:372: testsuite: e2e-tests/tests has 26 tests === RUN kuttl/harness === RUN kuttl/harness/gr-haproxy === PAUSE kuttl/harness/gr-haproxy === CONT kuttl/harness/gr-haproxy logger.go:42: 12:35:58 | gr-haproxy | Creating namespace: kuttl-test-direct-ray logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | starting test step 0-deploy-operator logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | running command: [sh -c set -o errexit set -o xtrace source ../../functions deploy_operator deploy_non_tls_cluster_secrets deploy_tls_cluster_secrets deploy_client] logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | + source ../../functions logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ realpath ../../.. logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | ++++ mktemp -d logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ export TEMP_DIR=/tmp/tmp.U1boGxWBUA logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ TEMP_DIR=/tmp/tmp.U1boGxWBUA logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ GIT_BRANCH=PR-339 logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ VERSION=PR-339-3a8483b logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | ++++ which gdate logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | ++++ which date logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ date=/usr/bin/date logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ command -v oc logger.go:42: 12:35:58 | gr-haproxy/0-deploy-operator | +++ oc get projects logger.go:42: 12:36:03 | gr-haproxy/0-deploy-operator | error: the server doesn't have a resource type "projects" logger.go:42: 12:36:03 | gr-haproxy/0-deploy-operator | +++ kubectl get nodes logger.go:42: 12:36:03 | gr-haproxy/0-deploy-operator | +++ grep '^minikube' logger.go:42: 12:36:04 | gr-haproxy/0-deploy-operator | ++++ pwd logger.go:42: 12:36:04 | gr-haproxy/0-deploy-operator | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/gr-haproxy logger.go:42: 12:36:04 | gr-haproxy/0-deploy-operator | ++ test_name=gr-haproxy logger.go:42: 12:36:04 | gr-haproxy/0-deploy-operator | + deploy_operator logger.go:42: 12:36:04 | gr-haproxy/0-deploy-operator | + kubectl -n kuttl-test-direct-ray apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/crd.yaml logger.go:42: 12:36:05 | gr-haproxy/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlbackups.ps.percona.com serverside-applied logger.go:42: 12:36:06 | gr-haproxy/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlrestores.ps.percona.com serverside-applied logger.go:42: 12:36:07 | gr-haproxy/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqls.ps.percona.com serverside-applied logger.go:42: 12:36:07 | gr-haproxy/0-deploy-operator | + kubectl -n kuttl-test-direct-ray apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/rbac.yaml logger.go:42: 12:36:08 | gr-haproxy/0-deploy-operator | serviceaccount/percona-server-mysql-operator created logger.go:42: 12:36:08 | gr-haproxy/0-deploy-operator | serviceaccount/percona-server-mysql-operator-orchestrator created logger.go:42: 12:36:09 | gr-haproxy/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 12:36:09 | gr-haproxy/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 12:36:09 | gr-haproxy/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 12:36:10 | gr-haproxy/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 12:36:10 | gr-haproxy/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 12:36:10 | gr-haproxy/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 12:36:10 | gr-haproxy/0-deploy-operator | + kubectl -n kuttl-test-direct-ray apply -f - logger.go:42: 12:36:10 | gr-haproxy/0-deploy-operator | ++ printf 'select(documentIndex==1).spec.template.spec.containers[0].image="%s"' perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:36:10 | gr-haproxy/0-deploy-operator | + yq eval '(select(documentIndex==1).spec.template.spec.containers[] | select(.name=="manager").env[] | select(.name=="DISABLE_TELEMETRY").value) = "true"' logger.go:42: 12:36:10 | gr-haproxy/0-deploy-operator | + yq eval 'select(documentIndex==1).spec.template.spec.containers[0].image="perconalab/percona-server-mysql-operator:PR-339-3a8483b"' /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/operator.yaml logger.go:42: 12:36:12 | gr-haproxy/0-deploy-operator | configmap/percona-server-mysql-operator-config created logger.go:42: 12:36:12 | gr-haproxy/0-deploy-operator | deployment.apps/percona-server-mysql-operator created logger.go:42: 12:36:12 | gr-haproxy/0-deploy-operator | + deploy_non_tls_cluster_secrets logger.go:42: 12:36:12 | gr-haproxy/0-deploy-operator | + kubectl -n kuttl-test-direct-ray apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/secrets.yaml logger.go:42: 12:36:13 | gr-haproxy/0-deploy-operator | secret/test-secrets created logger.go:42: 12:36:13 | gr-haproxy/0-deploy-operator | + deploy_tls_cluster_secrets logger.go:42: 12:36:13 | gr-haproxy/0-deploy-operator | + kubectl -n kuttl-test-direct-ray apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/ssl-secret.yaml logger.go:42: 12:36:15 | gr-haproxy/0-deploy-operator | secret/test-ssl created logger.go:42: 12:36:15 | gr-haproxy/0-deploy-operator | + deploy_client logger.go:42: 12:36:15 | gr-haproxy/0-deploy-operator | + kubectl -n kuttl-test-direct-ray apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/client.yaml logger.go:42: 12:36:16 | gr-haproxy/0-deploy-operator | pod/mysql-client created logger.go:42: 12:36:23 | gr-haproxy/0-deploy-operator | test step completed 0-deploy-operator logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | starting test step 1-create-cluster logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | running command: [sh -c set -o errexit set -o xtrace source ../../functions get_cr \ | yq eval '.spec.mysql.clusterType="group-replication"' - \ | yq eval '.spec.proxy.router.enabled=false' - \ | yq eval '.spec.proxy.haproxy.enabled=true' - \ | kubectl -n "${NAMESPACE}" apply -f -] logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | + source ../../functions logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ realpath ../../.. logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | ++++ mktemp -d logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ export TEMP_DIR=/tmp/tmp.uZ0Kg20bcf logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ TEMP_DIR=/tmp/tmp.uZ0Kg20bcf logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ GIT_BRANCH=PR-339 logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ VERSION=PR-339-3a8483b logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | ++++ which gdate logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | ++++ which date logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ date=/usr/bin/date logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ command -v oc logger.go:42: 12:36:23 | gr-haproxy/1-create-cluster | +++ oc get projects logger.go:42: 12:36:29 | gr-haproxy/1-create-cluster | error: the server doesn't have a resource type "projects" logger.go:42: 12:36:29 | gr-haproxy/1-create-cluster | +++ grep '^minikube' logger.go:42: 12:36:29 | gr-haproxy/1-create-cluster | +++ kubectl get nodes logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | ++++ pwd logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/gr-haproxy logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | ++ test_name=gr-haproxy logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | + yq eval '.spec.mysql.clusterType="group-replication"' - logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | + yq eval .spec.proxy.router.enabled=false - logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | + yq eval .spec.proxy.haproxy.enabled=true - logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | + get_cr logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | + local name_suffix= logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | + yq eval '.spec.secretsName="test-secrets"' - logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | ++ printf '.spec.initImage="%s"' perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | + yq eval '.spec.initImage="perconalab/percona-server-mysql-operator:PR-339-3a8483b"' - logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | + kubectl -n kuttl-test-direct-ray apply -f - logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | + yq eval '.spec.sslSecretName="test-ssl"' - logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | + yq eval '.spec.upgradeOptions.apply="disabled"' - logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | ++ printf '.spec.proxy.router.image="%s"' perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | ++ printf '.spec.orchestrator.image="%s"' perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | + yq eval '.spec.orchestrator.image="perconalab/percona-server-mysql-operator:main-orchestrator"' - logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | ++ printf '.spec.proxy.haproxy.image="%s"' perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | + yq eval '.spec.proxy.router.image="perconalab/percona-server-mysql-operator:main-router"' - logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | + '[' -n '' ']' logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | + yq eval - logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | ++ printf '.metadata.name="%s"' gr-haproxy logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | ++ printf '.spec.pmm.image="%s"' perconalab/pmm-client:dev-latest logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | + yq eval '.spec.pmm.image="perconalab/pmm-client:dev-latest"' - logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | + yq eval '.metadata.name="gr-haproxy"' /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/cr.yaml logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | ++ printf '.spec.toolkit.image="%s"' perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | + yq eval '.spec.proxy.haproxy.image="perconalab/percona-server-mysql-operator:main-haproxy"' - logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | ++ printf '.spec.mysql.image="%s"' perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | + yq eval '.spec.mysql.image="perconalab/percona-server-mysql-operator:main-psmysql"' - logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | + yq eval '.spec.toolkit.image="perconalab/percona-server-mysql-operator:main-toolkit"' - logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | ++ printf '.spec.backup.image="%s"' perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:36:30 | gr-haproxy/1-create-cluster | + yq eval '.spec.backup.image="perconalab/percona-server-mysql-operator:main-backup"' - logger.go:42: 12:36:31 | gr-haproxy/1-create-cluster | perconaservermysql.ps.percona.com/gr-haproxy created logger.go:42: 12:40:09 | gr-haproxy/1-create-cluster | test step completed 1-create-cluster logger.go:42: 12:40:09 | gr-haproxy/2-write-data | starting test step 2-write-data logger.go:42: 12:40:09 | gr-haproxy/2-write-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password" run_mysql \ "INSERT myDB.myTable (id) VALUES (100500)" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password"] logger.go:42: 12:40:09 | gr-haproxy/2-write-data | + source ../../functions logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ realpath ../../.. logger.go:42: 12:40:09 | gr-haproxy/2-write-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:40:09 | gr-haproxy/2-write-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:40:09 | gr-haproxy/2-write-data | ++++ mktemp -d logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ export TEMP_DIR=/tmp/tmp.MUOnCGROCx logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ TEMP_DIR=/tmp/tmp.MUOnCGROCx logger.go:42: 12:40:09 | gr-haproxy/2-write-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ GIT_BRANCH=PR-339 logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ VERSION=PR-339-3a8483b logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:40:09 | gr-haproxy/2-write-data | ++++ which gdate logger.go:42: 12:40:09 | gr-haproxy/2-write-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:40:09 | gr-haproxy/2-write-data | ++++ which date logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ date=/usr/bin/date logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ command -v oc logger.go:42: 12:40:09 | gr-haproxy/2-write-data | +++ oc get projects logger.go:42: 12:40:14 | gr-haproxy/2-write-data | error: the server doesn't have a resource type "projects" logger.go:42: 12:40:14 | gr-haproxy/2-write-data | +++ grep '^minikube' logger.go:42: 12:40:14 | gr-haproxy/2-write-data | +++ kubectl get nodes logger.go:42: 12:40:15 | gr-haproxy/2-write-data | ++++ pwd logger.go:42: 12:40:15 | gr-haproxy/2-write-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/gr-haproxy logger.go:42: 12:40:15 | gr-haproxy/2-write-data | ++ test_name=gr-haproxy logger.go:42: 12:40:15 | gr-haproxy/2-write-data | +++ get_cluster_name logger.go:42: 12:40:15 | gr-haproxy/2-write-data | +++ kubectl -n kuttl-test-direct-ray get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 12:40:16 | gr-haproxy/2-write-data | ++ get_haproxy_svc gr-haproxy logger.go:42: 12:40:16 | gr-haproxy/2-write-data | ++ local cluster=gr-haproxy logger.go:42: 12:40:16 | gr-haproxy/2-write-data | ++ echo gr-haproxy-haproxy logger.go:42: 12:40:16 | gr-haproxy/2-write-data | + run_mysql 'CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' '-h gr-haproxy-haproxy -uroot -proot_password' logger.go:42: 12:40:16 | gr-haproxy/2-write-data | + local 'command=CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' logger.go:42: 12:40:16 | gr-haproxy/2-write-data | + local 'uri=-h gr-haproxy-haproxy -uroot -proot_password' logger.go:42: 12:40:16 | gr-haproxy/2-write-data | + local pod= logger.go:42: 12:40:16 | gr-haproxy/2-write-data | ++ get_client_pod logger.go:42: 12:40:16 | gr-haproxy/2-write-data | ++ kubectl -n kuttl-test-direct-ray get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:40:17 | gr-haproxy/2-write-data | + client_pod=mysql-client logger.go:42: 12:40:17 | gr-haproxy/2-write-data | + wait_pod mysql-client logger.go:42: 12:40:17 | gr-haproxy/2-write-data | + local pod=mysql-client logger.go:42: 12:40:17 | gr-haproxy/2-write-data | + set +o xtrace logger.go:42: 12:40:17 | gr-haproxy/2-write-data | mysql-clienttrue logger.go:42: 12:40:17 | gr-haproxy/2-write-data | + kubectl -n kuttl-test-direct-ray exec mysql-client -- bash -c 'printf '\''%s\n'\'' "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" | mysql -sN -h gr-haproxy-haproxy -uroot -proot_password' logger.go:42: 12:40:17 | gr-haproxy/2-write-data | + sed -e 's/mysql: //' logger.go:42: 12:40:17 | gr-haproxy/2-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:40:19 | gr-haproxy/2-write-data | ERROR 2013 (HY000): Lost connection to MySQL server at 'reading initial communication packet', system error: 0 logger.go:42: 12:40:19 | gr-haproxy/2-write-data | command terminated with exit code 1 logger.go:42: 12:40:19 | gr-haproxy/2-write-data | +++ get_cluster_name logger.go:42: 12:40:19 | gr-haproxy/2-write-data | +++ kubectl -n kuttl-test-direct-ray get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 12:40:19 | gr-haproxy/2-write-data | ++ get_haproxy_svc gr-haproxy logger.go:42: 12:40:19 | gr-haproxy/2-write-data | ++ local cluster=gr-haproxy logger.go:42: 12:40:19 | gr-haproxy/2-write-data | ++ echo gr-haproxy-haproxy logger.go:42: 12:40:19 | gr-haproxy/2-write-data | + run_mysql 'INSERT myDB.myTable (id) VALUES (100500)' '-h gr-haproxy-haproxy -uroot -proot_password' logger.go:42: 12:40:19 | gr-haproxy/2-write-data | + local 'command=INSERT myDB.myTable (id) VALUES (100500)' logger.go:42: 12:40:19 | gr-haproxy/2-write-data | + local 'uri=-h gr-haproxy-haproxy -uroot -proot_password' logger.go:42: 12:40:19 | gr-haproxy/2-write-data | + local pod= logger.go:42: 12:40:19 | gr-haproxy/2-write-data | ++ get_client_pod logger.go:42: 12:40:19 | gr-haproxy/2-write-data | ++ kubectl -n kuttl-test-direct-ray get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:40:20 | gr-haproxy/2-write-data | + client_pod=mysql-client logger.go:42: 12:40:20 | gr-haproxy/2-write-data | + wait_pod mysql-client logger.go:42: 12:40:20 | gr-haproxy/2-write-data | + local pod=mysql-client logger.go:42: 12:40:20 | gr-haproxy/2-write-data | + set +o xtrace logger.go:42: 12:40:21 | gr-haproxy/2-write-data | mysql-clienttrue logger.go:42: 12:40:21 | gr-haproxy/2-write-data | + kubectl -n kuttl-test-direct-ray exec mysql-client -- bash -c 'printf '\''%s\n'\'' "INSERT myDB.myTable (id) VALUES (100500)" | mysql -sN -h gr-haproxy-haproxy -uroot -proot_password' logger.go:42: 12:40:21 | gr-haproxy/2-write-data | + sed -e 's/mysql: //' logger.go:42: 12:40:21 | gr-haproxy/2-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:40:22 | gr-haproxy/2-write-data | ERROR 2013 (HY000): Lost connection to MySQL server at 'reading initial communication packet', system error: 0 logger.go:42: 12:40:22 | gr-haproxy/2-write-data | command terminated with exit code 1 logger.go:42: 12:40:23 | gr-haproxy/2-write-data | test step completed 2-write-data logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | starting test step 3-check-connections logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | running command: [sh -c set -o errexit set -o xtrace source ../../functions args="" primary=$(run_mysql "SELECT count(*) FROM myDB.myTable" "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password") args="${args} --from-literal=primary=${primary}" replica=$(run_mysql "SELECT count(*) FROM myDB.myTable" "-h $(get_haproxy_svc $(get_cluster_name)) -P 3307 -uroot -proot_password") args="${args} --from-literal=replica=${replica}" # uncomment after https://jira.percona.com/browse/K8SPS-284 # proxy_protocol=$(run_mysql "SELECT count(*) FROM myDB.myTable" "-h $(get_haproxy_svc $(get_cluster_name)) -P 3309 -uroot -proot_password") # args="${args} --from-literal=proxy_protocol=${proxy_protocol}" mysqlx=$(run_mysql "SELECT count(*) FROM myDB.myTable" "-h $(get_haproxy_svc $(get_cluster_name)) -P 33060 -uroot -proot_password") args="${args} --from-literal=mysqlx=${mysqlx}" mysql_admin=$(run_mysql "SELECT count(*) FROM myDB.myTable" "-h $(get_haproxy_svc $(get_cluster_name)) -P 33062 -uroot -proot_password") args="${args} --from-literal=mysql_admin=${mysql_admin}" kubectl create configmap -n "${NAMESPACE}" 03-check-connections $args] logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | + source ../../functions logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ realpath ../../.. logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | ++++ mktemp -d logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ export TEMP_DIR=/tmp/tmp.vm9jgOypmL logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ TEMP_DIR=/tmp/tmp.vm9jgOypmL logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ GIT_BRANCH=PR-339 logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ VERSION=PR-339-3a8483b logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | ++++ which gdate logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | ++++ which date logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ date=/usr/bin/date logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ command -v oc logger.go:42: 12:40:23 | gr-haproxy/3-check-connections | +++ oc get projects logger.go:42: 12:40:29 | gr-haproxy/3-check-connections | error: the server doesn't have a resource type "projects" logger.go:42: 12:40:29 | gr-haproxy/3-check-connections | +++ kubectl get nodes logger.go:42: 12:40:29 | gr-haproxy/3-check-connections | +++ grep '^minikube' logger.go:42: 12:40:30 | gr-haproxy/3-check-connections | ++++ pwd logger.go:42: 12:40:30 | gr-haproxy/3-check-connections | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/gr-haproxy logger.go:42: 12:40:30 | gr-haproxy/3-check-connections | ++ test_name=gr-haproxy logger.go:42: 12:40:30 | gr-haproxy/3-check-connections | + args= logger.go:42: 12:40:30 | gr-haproxy/3-check-connections | ++++ get_cluster_name logger.go:42: 12:40:30 | gr-haproxy/3-check-connections | ++++ kubectl -n kuttl-test-direct-ray get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 12:40:30 | gr-haproxy/3-check-connections | +++ get_haproxy_svc gr-haproxy logger.go:42: 12:40:30 | gr-haproxy/3-check-connections | +++ local cluster=gr-haproxy logger.go:42: 12:40:30 | gr-haproxy/3-check-connections | +++ echo gr-haproxy-haproxy logger.go:42: 12:40:30 | gr-haproxy/3-check-connections | ++ run_mysql 'SELECT count(*) FROM myDB.myTable' '-h gr-haproxy-haproxy -uroot -proot_password' logger.go:42: 12:40:30 | gr-haproxy/3-check-connections | ++ local 'command=SELECT count(*) FROM myDB.myTable' logger.go:42: 12:40:30 | gr-haproxy/3-check-connections | ++ local 'uri=-h gr-haproxy-haproxy -uroot -proot_password' logger.go:42: 12:40:30 | gr-haproxy/3-check-connections | ++ local pod= logger.go:42: 12:40:30 | gr-haproxy/3-check-connections | +++ get_client_pod logger.go:42: 12:40:30 | gr-haproxy/3-check-connections | +++ kubectl -n kuttl-test-direct-ray get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:40:31 | gr-haproxy/3-check-connections | ++ client_pod=mysql-client logger.go:42: 12:40:31 | gr-haproxy/3-check-connections | ++ wait_pod mysql-client logger.go:42: 12:40:31 | gr-haproxy/3-check-connections | ++ local pod=mysql-client logger.go:42: 12:40:31 | gr-haproxy/3-check-connections | ++ set +o xtrace logger.go:42: 12:40:32 | gr-haproxy/3-check-connections | mysql-clienttrue logger.go:42: 12:40:32 | gr-haproxy/3-check-connections | ++ kubectl -n kuttl-test-direct-ray exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT count(*) FROM myDB.myTable" | mysql -sN -h gr-haproxy-haproxy -uroot -proot_password' logger.go:42: 12:40:32 | gr-haproxy/3-check-connections | ++ sed -e 's/mysql: //' logger.go:42: 12:40:32 | gr-haproxy/3-check-connections | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:40:33 | gr-haproxy/3-check-connections | + primary='ERROR 2013 (HY000): Lost connection to MySQL server at '\''reading initial communication packet'\'', system error: 0 logger.go:42: 12:40:33 | gr-haproxy/3-check-connections | command terminated with exit code 1' logger.go:42: 12:40:33 | gr-haproxy/3-check-connections | + args=' --from-literal=primary=ERROR 2013 (HY000): Lost connection to MySQL server at '\''reading initial communication packet'\'', system error: 0 logger.go:42: 12:40:33 | gr-haproxy/3-check-connections | command terminated with exit code 1' logger.go:42: 12:40:33 | gr-haproxy/3-check-connections | ++++ get_cluster_name logger.go:42: 12:40:33 | gr-haproxy/3-check-connections | ++++ kubectl -n kuttl-test-direct-ray get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 12:40:34 | gr-haproxy/3-check-connections | +++ get_haproxy_svc gr-haproxy logger.go:42: 12:40:34 | gr-haproxy/3-check-connections | +++ local cluster=gr-haproxy logger.go:42: 12:40:34 | gr-haproxy/3-check-connections | +++ echo gr-haproxy-haproxy logger.go:42: 12:40:34 | gr-haproxy/3-check-connections | ++ run_mysql 'SELECT count(*) FROM myDB.myTable' '-h gr-haproxy-haproxy -P 3307 -uroot -proot_password' logger.go:42: 12:40:34 | gr-haproxy/3-check-connections | ++ local 'command=SELECT count(*) FROM myDB.myTable' logger.go:42: 12:40:34 | gr-haproxy/3-check-connections | ++ local 'uri=-h gr-haproxy-haproxy -P 3307 -uroot -proot_password' logger.go:42: 12:40:34 | gr-haproxy/3-check-connections | ++ local pod= logger.go:42: 12:40:34 | gr-haproxy/3-check-connections | +++ get_client_pod logger.go:42: 12:40:34 | gr-haproxy/3-check-connections | +++ kubectl -n kuttl-test-direct-ray get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:40:35 | gr-haproxy/3-check-connections | ++ client_pod=mysql-client logger.go:42: 12:40:35 | gr-haproxy/3-check-connections | ++ wait_pod mysql-client logger.go:42: 12:40:35 | gr-haproxy/3-check-connections | ++ local pod=mysql-client logger.go:42: 12:40:35 | gr-haproxy/3-check-connections | ++ set +o xtrace logger.go:42: 12:40:35 | gr-haproxy/3-check-connections | mysql-clienttrue logger.go:42: 12:40:35 | gr-haproxy/3-check-connections | ++ kubectl -n kuttl-test-direct-ray exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT count(*) FROM myDB.myTable" | mysql -sN -h gr-haproxy-haproxy -P 3307 -uroot -proot_password' logger.go:42: 12:40:35 | gr-haproxy/3-check-connections | ++ sed -e 's/mysql: //' logger.go:42: 12:40:35 | gr-haproxy/3-check-connections | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:40:37 | gr-haproxy/3-check-connections | + replica='ERROR 2013 (HY000): Lost connection to MySQL server at '\''reading initial communication packet'\'', system error: 0 logger.go:42: 12:40:37 | gr-haproxy/3-check-connections | command terminated with exit code 1' logger.go:42: 12:40:37 | gr-haproxy/3-check-connections | + args=' --from-literal=primary=ERROR 2013 (HY000): Lost connection to MySQL server at '\''reading initial communication packet'\'', system error: 0 logger.go:42: 12:40:37 | gr-haproxy/3-check-connections | command terminated with exit code 1 --from-literal=replica=ERROR 2013 (HY000): Lost connection to MySQL server at '\''reading initial communication packet'\'', system error: 0 logger.go:42: 12:40:37 | gr-haproxy/3-check-connections | command terminated with exit code 1' logger.go:42: 12:40:37 | gr-haproxy/3-check-connections | ++++ get_cluster_name logger.go:42: 12:40:37 | gr-haproxy/3-check-connections | ++++ kubectl -n kuttl-test-direct-ray get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 12:40:38 | gr-haproxy/3-check-connections | +++ get_haproxy_svc gr-haproxy logger.go:42: 12:40:38 | gr-haproxy/3-check-connections | +++ local cluster=gr-haproxy logger.go:42: 12:40:38 | gr-haproxy/3-check-connections | +++ echo gr-haproxy-haproxy logger.go:42: 12:40:38 | gr-haproxy/3-check-connections | ++ run_mysql 'SELECT count(*) FROM myDB.myTable' '-h gr-haproxy-haproxy -P 33060 -uroot -proot_password' logger.go:42: 12:40:38 | gr-haproxy/3-check-connections | ++ local 'command=SELECT count(*) FROM myDB.myTable' logger.go:42: 12:40:38 | gr-haproxy/3-check-connections | ++ local 'uri=-h gr-haproxy-haproxy -P 33060 -uroot -proot_password' logger.go:42: 12:40:38 | gr-haproxy/3-check-connections | ++ local pod= logger.go:42: 12:40:38 | gr-haproxy/3-check-connections | +++ get_client_pod logger.go:42: 12:40:38 | gr-haproxy/3-check-connections | +++ kubectl -n kuttl-test-direct-ray get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:40:38 | gr-haproxy/3-check-connections | ++ client_pod=mysql-client logger.go:42: 12:40:38 | gr-haproxy/3-check-connections | ++ wait_pod mysql-client logger.go:42: 12:40:38 | gr-haproxy/3-check-connections | ++ local pod=mysql-client logger.go:42: 12:40:38 | gr-haproxy/3-check-connections | ++ set +o xtrace logger.go:42: 12:40:39 | gr-haproxy/3-check-connections | mysql-clienttrue logger.go:42: 12:40:39 | gr-haproxy/3-check-connections | ++ kubectl -n kuttl-test-direct-ray exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT count(*) FROM myDB.myTable" | mysql -sN -h gr-haproxy-haproxy -P 33060 -uroot -proot_password' logger.go:42: 12:40:39 | gr-haproxy/3-check-connections | ++ sed -e 's/mysql: //' logger.go:42: 12:40:39 | gr-haproxy/3-check-connections | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:40:41 | gr-haproxy/3-check-connections | + mysqlx='ERROR 2013 (HY000): Lost connection to MySQL server at '\''reading initial communication packet'\'', system error: 0 logger.go:42: 12:40:41 | gr-haproxy/3-check-connections | command terminated with exit code 1' logger.go:42: 12:40:41 | gr-haproxy/3-check-connections | + args=' --from-literal=primary=ERROR 2013 (HY000): Lost connection to MySQL server at '\''reading initial communication packet'\'', system error: 0 logger.go:42: 12:40:41 | gr-haproxy/3-check-connections | command terminated with exit code 1 --from-literal=replica=ERROR 2013 (HY000): Lost connection to MySQL server at '\''reading initial communication packet'\'', system error: 0 logger.go:42: 12:40:41 | gr-haproxy/3-check-connections | command terminated with exit code 1 --from-literal=mysqlx=ERROR 2013 (HY000): Lost connection to MySQL server at '\''reading initial communication packet'\'', system error: 0 logger.go:42: 12:40:41 | gr-haproxy/3-check-connections | command terminated with exit code 1' logger.go:42: 12:40:41 | gr-haproxy/3-check-connections | ++++ get_cluster_name logger.go:42: 12:40:41 | gr-haproxy/3-check-connections | ++++ kubectl -n kuttl-test-direct-ray get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 12:40:41 | gr-haproxy/3-check-connections | +++ get_haproxy_svc gr-haproxy logger.go:42: 12:40:41 | gr-haproxy/3-check-connections | +++ local cluster=gr-haproxy logger.go:42: 12:40:41 | gr-haproxy/3-check-connections | +++ echo gr-haproxy-haproxy logger.go:42: 12:40:41 | gr-haproxy/3-check-connections | ++ run_mysql 'SELECT count(*) FROM myDB.myTable' '-h gr-haproxy-haproxy -P 33062 -uroot -proot_password' logger.go:42: 12:40:41 | gr-haproxy/3-check-connections | ++ local 'command=SELECT count(*) FROM myDB.myTable' logger.go:42: 12:40:41 | gr-haproxy/3-check-connections | ++ local 'uri=-h gr-haproxy-haproxy -P 33062 -uroot -proot_password' logger.go:42: 12:40:41 | gr-haproxy/3-check-connections | ++ local pod= logger.go:42: 12:40:41 | gr-haproxy/3-check-connections | +++ get_client_pod logger.go:42: 12:40:41 | gr-haproxy/3-check-connections | +++ kubectl -n kuttl-test-direct-ray get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 12:40:42 | gr-haproxy/3-check-connections | ++ client_pod=mysql-client logger.go:42: 12:40:42 | gr-haproxy/3-check-connections | ++ wait_pod mysql-client logger.go:42: 12:40:42 | gr-haproxy/3-check-connections | ++ local pod=mysql-client logger.go:42: 12:40:42 | gr-haproxy/3-check-connections | ++ set +o xtrace logger.go:42: 12:40:43 | gr-haproxy/3-check-connections | mysql-clienttrue logger.go:42: 12:40:43 | gr-haproxy/3-check-connections | ++ kubectl -n kuttl-test-direct-ray exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT count(*) FROM myDB.myTable" | mysql -sN -h gr-haproxy-haproxy -P 33062 -uroot -proot_password' logger.go:42: 12:40:43 | gr-haproxy/3-check-connections | ++ sed -e 's/mysql: //' logger.go:42: 12:40:43 | gr-haproxy/3-check-connections | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 12:40:44 | gr-haproxy/3-check-connections | + mysql_admin='ERROR 2013 (HY000): Lost connection to MySQL server at '\''reading initial communication packet'\'', system error: 0 logger.go:42: 12:40:44 | gr-haproxy/3-check-connections | command terminated with exit code 1' logger.go:42: 12:40:44 | gr-haproxy/3-check-connections | + args=' --from-literal=primary=ERROR 2013 (HY000): Lost connection to MySQL server at '\''reading initial communication packet'\'', system error: 0 logger.go:42: 12:40:44 | gr-haproxy/3-check-connections | command terminated with exit code 1 --from-literal=replica=ERROR 2013 (HY000): Lost connection to MySQL server at '\''reading initial communication packet'\'', system error: 0 logger.go:42: 12:40:44 | gr-haproxy/3-check-connections | command terminated with exit code 1 --from-literal=mysqlx=ERROR 2013 (HY000): Lost connection to MySQL server at '\''reading initial communication packet'\'', system error: 0 logger.go:42: 12:40:44 | gr-haproxy/3-check-connections | command terminated with exit code 1 --from-literal=mysql_admin=ERROR 2013 (HY000): Lost connection to MySQL server at '\''reading initial communication packet'\'', system error: 0 logger.go:42: 12:40:44 | gr-haproxy/3-check-connections | command terminated with exit code 1' logger.go:42: 12:40:44 | gr-haproxy/3-check-connections | + kubectl create configmap -n kuttl-test-direct-ray 03-check-connections --from-literal=primary=ERROR 2013 '(HY000):' Lost connection to MySQL server at ''\''reading' initial communication 'packet'\'',' system error: 0 command terminated with exit code 1 --from-literal=replica=ERROR 2013 '(HY000):' Lost connection to MySQL server at ''\''reading' initial communication 'packet'\'',' system error: 0 command terminated with exit code 1 --from-literal=mysqlx=ERROR 2013 '(HY000):' Lost connection to MySQL server at ''\''reading' initial communication 'packet'\'',' system error: 0 command terminated with exit code 1 --from-literal=mysql_admin=ERROR 2013 '(HY000):' Lost connection to MySQL server at ''\''reading' initial communication 'packet'\'',' system error: 0 command terminated with exit code 1 logger.go:42: 12:40:45 | gr-haproxy/3-check-connections | error: exactly one NAME is required, got 85 logger.go:42: 12:40:45 | gr-haproxy/3-check-connections | See 'kubectl create configmap -h' for help and examples case.go:364: failed in step 3-check-connections case.go:366: exit status 1 logger.go:42: 12:40:46 | gr-haproxy | gr-haproxy events from ns kuttl-test-direct-ray: logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:12 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-kbvzg Scheduled Successfully assigned kuttl-test-direct-ray/percona-server-mysql-operator-6dcd78d857-kbvzg to gke-jen-ps-339-3a8483b-1-default-pool-9abd8684-p5sl logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:12 +0000 UTC Normal ReplicaSet.apps percona-server-mysql-operator-6dcd78d857 SuccessfulCreate Created pod: percona-server-mysql-operator-6dcd78d857-kbvzg logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:12 +0000 UTC Normal Deployment.apps percona-server-mysql-operator ScalingReplicaSet Scaled up replica set percona-server-mysql-operator-6dcd78d857 to 1 logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:15 +0000 UTC Normal Lease.coordination.k8s.io 08db2feb.percona.com LeaderElection percona-server-mysql-operator-6dcd78d857-kbvzg_16a6de5d-7cf1-4a61-ad63-bef96b782354 became leader logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:15 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-kbvzg.spec.containers{manager} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:15 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-kbvzg.spec.containers{manager} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 402.994532ms logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:15 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-kbvzg.spec.containers{manager} Created Created container manager logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:15 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-kbvzg.spec.containers{manager} Started Started container manager logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:16 +0000 UTC Normal Pod mysql-client Scheduled Successfully assigned kuttl-test-direct-ray/mysql-client to gke-jen-ps-339-3a8483b-1-default-pool-9abd8684-f1l9 logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:17 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Pulling Pulling image "percona/percona-server:8.0.25" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:28 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Pulled Successfully pulled image "percona/percona-server:8.0.25" in 11.10340813s logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:28 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Created Created container mysql-client logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:28 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Started Started container mysql-client logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:33 +0000 UTC Normal PersistentVolumeClaim datadir-gr-haproxy-mysql-0 WaitForFirstConsumer waiting for first consumer to be created before binding logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:33 +0000 UTC Normal StatefulSet.apps gr-haproxy-mysql SuccessfulCreate create Claim datadir-gr-haproxy-mysql-0 Pod gr-haproxy-mysql-0 in StatefulSet gr-haproxy-mysql success logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:34 +0000 UTC Normal PersistentVolumeClaim datadir-gr-haproxy-mysql-0 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:34 +0000 UTC Normal PersistentVolumeClaim datadir-gr-haproxy-mysql-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-direct-ray/datadir-gr-haproxy-mysql-0" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:34 +0000 UTC Normal StatefulSet.apps gr-haproxy-mysql SuccessfulCreate create Pod gr-haproxy-mysql-0 in StatefulSet gr-haproxy-mysql successful logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:38 +0000 UTC Normal PersistentVolumeClaim datadir-gr-haproxy-mysql-0 ProvisioningSucceeded Successfully provisioned volume pvc-8e1fb825-bf54-4576-bc21-9c0898c53fb2 logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:39 +0000 UTC Normal Pod gr-haproxy-mysql-0 Scheduled Successfully assigned kuttl-test-direct-ray/gr-haproxy-mysql-0 to gke-jen-ps-339-3a8483b-1-default-pool-9abd8684-hccm logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:44 +0000 UTC Normal Pod gr-haproxy-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-8e1fb825-bf54-4576-bc21-9c0898c53fb2" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:49 +0000 UTC Normal Pod gr-haproxy-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:49 +0000 UTC Normal Pod gr-haproxy-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 419.483567ms logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:49 +0000 UTC Normal Pod gr-haproxy-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:49 +0000 UTC Normal Pod gr-haproxy-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:51 +0000 UTC Normal Pod gr-haproxy-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:52 +0000 UTC Normal Pod gr-haproxy-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 384.219765ms logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:52 +0000 UTC Normal Pod gr-haproxy-mysql-0.spec.containers{mysql} Created Created container mysql logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:52 +0000 UTC Normal Pod gr-haproxy-mysql-0.spec.containers{mysql} Started Started container mysql logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:52 +0000 UTC Normal Pod gr-haproxy-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:52 +0000 UTC Normal Pod gr-haproxy-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 396.78164ms logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:52 +0000 UTC Normal Pod gr-haproxy-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:36:52 +0000 UTC Normal Pod gr-haproxy-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:23 +0000 UTC Normal PersistentVolumeClaim datadir-gr-haproxy-mysql-1 WaitForFirstConsumer waiting for first consumer to be created before binding logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:23 +0000 UTC Normal PersistentVolumeClaim datadir-gr-haproxy-mysql-1 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:23 +0000 UTC Normal PersistentVolumeClaim datadir-gr-haproxy-mysql-1 Provisioning External provisioner is provisioning volume for claim "kuttl-test-direct-ray/datadir-gr-haproxy-mysql-1" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:23 +0000 UTC Normal StatefulSet.apps gr-haproxy-mysql SuccessfulCreate create Claim datadir-gr-haproxy-mysql-1 Pod gr-haproxy-mysql-1 in StatefulSet gr-haproxy-mysql success logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:23 +0000 UTC Normal StatefulSet.apps gr-haproxy-mysql SuccessfulCreate create Pod gr-haproxy-mysql-1 in StatefulSet gr-haproxy-mysql successful logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:24 +0000 UTC Normal Pod gr-haproxy-haproxy-0 Scheduled Successfully assigned kuttl-test-direct-ray/gr-haproxy-haproxy-0 to gke-jen-ps-339-3a8483b-1-default-pool-9abd8684-f1l9 logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:24 +0000 UTC Normal StatefulSet.apps gr-haproxy-haproxy SuccessfulCreate create Pod gr-haproxy-haproxy-0 in StatefulSet gr-haproxy-haproxy successful logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:25 +0000 UTC Normal Pod gr-haproxy-haproxy-0.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:25 +0000 UTC Normal Pod gr-haproxy-haproxy-0.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 387.527673ms logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:25 +0000 UTC Normal Pod gr-haproxy-haproxy-0.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:26 +0000 UTC Normal Pod gr-haproxy-haproxy-0.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:27 +0000 UTC Normal PersistentVolumeClaim datadir-gr-haproxy-mysql-1 ProvisioningSucceeded Successfully provisioned volume pvc-c77e36ee-d2df-4297-b377-c7b5a1a57778 logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:27 +0000 UTC Normal Pod gr-haproxy-haproxy-0.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:27 +0000 UTC Normal Pod gr-haproxy-mysql-1 Scheduled Successfully assigned kuttl-test-direct-ray/gr-haproxy-mysql-1 to gke-jen-ps-339-3a8483b-1-default-pool-9abd8684-p5sl logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:28 +0000 UTC Normal Pod gr-haproxy-haproxy-0.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 382.42099ms logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:28 +0000 UTC Normal Pod gr-haproxy-haproxy-0.spec.containers{haproxy} Created Created container haproxy logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:28 +0000 UTC Normal Pod gr-haproxy-haproxy-0.spec.containers{haproxy} Started Started container haproxy logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:28 +0000 UTC Normal Pod gr-haproxy-haproxy-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:28 +0000 UTC Normal Pod gr-haproxy-haproxy-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 379.525986ms logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:28 +0000 UTC Normal Pod gr-haproxy-haproxy-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:28 +0000 UTC Normal Pod gr-haproxy-haproxy-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:28 +0000 UTC Normal Pod gr-haproxy-haproxy-1 Scheduled Successfully assigned kuttl-test-direct-ray/gr-haproxy-haproxy-1 to gke-jen-ps-339-3a8483b-1-default-pool-9abd8684-p5sl logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:28 +0000 UTC Normal StatefulSet.apps gr-haproxy-haproxy SuccessfulCreate create Pod gr-haproxy-haproxy-1 in StatefulSet gr-haproxy-haproxy successful logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:29 +0000 UTC Normal Pod gr-haproxy-haproxy-1.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:29 +0000 UTC Normal Pod gr-haproxy-haproxy-1.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 390.04565ms logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:29 +0000 UTC Normal Pod gr-haproxy-haproxy-1.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:30 +0000 UTC Normal Pod gr-haproxy-haproxy-1.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:32 +0000 UTC Normal Pod gr-haproxy-haproxy-1.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:32 +0000 UTC Normal Pod gr-haproxy-haproxy-1.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 397.030967ms logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:32 +0000 UTC Normal Pod gr-haproxy-haproxy-1.spec.containers{haproxy} Created Created container haproxy logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:32 +0000 UTC Normal Pod gr-haproxy-haproxy-1.spec.containers{haproxy} Started Started container haproxy logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:32 +0000 UTC Normal Pod gr-haproxy-haproxy-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:32 +0000 UTC Normal Pod gr-haproxy-mysql-1 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-c77e36ee-d2df-4297-b377-c7b5a1a57778" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:33 +0000 UTC Normal Pod gr-haproxy-haproxy-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 405.306789ms logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:33 +0000 UTC Normal Pod gr-haproxy-haproxy-1.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:33 +0000 UTC Normal Pod gr-haproxy-haproxy-1.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:34 +0000 UTC Normal Pod gr-haproxy-haproxy-2 Scheduled Successfully assigned kuttl-test-direct-ray/gr-haproxy-haproxy-2 to gke-jen-ps-339-3a8483b-1-default-pool-9abd8684-hccm logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:34 +0000 UTC Normal StatefulSet.apps gr-haproxy-haproxy SuccessfulCreate create Pod gr-haproxy-haproxy-2 in StatefulSet gr-haproxy-haproxy successful logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:34 +0000 UTC Normal Pod gr-haproxy-mysql-1.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:34 +0000 UTC Normal Pod gr-haproxy-mysql-1.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 364.006717ms logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:34 +0000 UTC Normal Pod gr-haproxy-mysql-1.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:34 +0000 UTC Normal Pod gr-haproxy-mysql-1.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:35 +0000 UTC Normal Pod gr-haproxy-haproxy-2.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:35 +0000 UTC Normal Pod gr-haproxy-haproxy-2.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 440.947461ms logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:35 +0000 UTC Normal Pod gr-haproxy-haproxy-2.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:35 +0000 UTC Normal Pod gr-haproxy-haproxy-2.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:36 +0000 UTC Normal Pod gr-haproxy-mysql-1.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:36 +0000 UTC Normal Pod gr-haproxy-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 389.229588ms logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:36 +0000 UTC Normal Pod gr-haproxy-mysql-1.spec.containers{mysql} Created Created container mysql logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:36 +0000 UTC Normal Pod gr-haproxy-mysql-1.spec.containers{mysql} Started Started container mysql logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:36 +0000 UTC Normal Pod gr-haproxy-mysql-1.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:37 +0000 UTC Normal Pod gr-haproxy-mysql-1.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 394.420534ms logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:37 +0000 UTC Normal Pod gr-haproxy-mysql-1.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:37 +0000 UTC Normal Pod gr-haproxy-mysql-1.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:38 +0000 UTC Normal Pod gr-haproxy-haproxy-2.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:38 +0000 UTC Normal Pod gr-haproxy-haproxy-2.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 389.649189ms logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:38 +0000 UTC Normal Pod gr-haproxy-haproxy-2.spec.containers{haproxy} Created Created container haproxy logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:38 +0000 UTC Normal Pod gr-haproxy-haproxy-2.spec.containers{haproxy} Started Started container haproxy logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:38 +0000 UTC Normal Pod gr-haproxy-haproxy-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:38 +0000 UTC Normal Pod gr-haproxy-haproxy-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 378.441659ms logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:38 +0000 UTC Normal Pod gr-haproxy-haproxy-2.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:37:39 +0000 UTC Normal Pod gr-haproxy-haproxy-2.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:38:06 +0000 UTC Warning Pod gr-haproxy-mysql-1.spec.containers{mysql} Unhealthy Startup probe failed: logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:38:06 +0000 UTC Normal Pod gr-haproxy-mysql-1.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:38:07 +0000 UTC Normal Pod gr-haproxy-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 417.109735ms logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:38:39 +0000 UTC Normal PersistentVolumeClaim datadir-gr-haproxy-mysql-2 WaitForFirstConsumer waiting for first consumer to be created before binding logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:38:39 +0000 UTC Normal PersistentVolumeClaim datadir-gr-haproxy-mysql-2 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:38:39 +0000 UTC Normal PersistentVolumeClaim datadir-gr-haproxy-mysql-2 Provisioning External provisioner is provisioning volume for claim "kuttl-test-direct-ray/datadir-gr-haproxy-mysql-2" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:38:39 +0000 UTC Normal StatefulSet.apps gr-haproxy-mysql SuccessfulCreate create Claim datadir-gr-haproxy-mysql-2 Pod gr-haproxy-mysql-2 in StatefulSet gr-haproxy-mysql success logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:38:39 +0000 UTC Normal StatefulSet.apps gr-haproxy-mysql SuccessfulCreate create Pod gr-haproxy-mysql-2 in StatefulSet gr-haproxy-mysql successful logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:38:43 +0000 UTC Normal PersistentVolumeClaim datadir-gr-haproxy-mysql-2 ProvisioningSucceeded Successfully provisioned volume pvc-489c7437-3224-4ed8-ab1f-64df89cea6b5 logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:38:43 +0000 UTC Normal Pod gr-haproxy-mysql-2 Scheduled Successfully assigned kuttl-test-direct-ray/gr-haproxy-mysql-2 to gke-jen-ps-339-3a8483b-1-default-pool-9abd8684-f1l9 logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:38:50 +0000 UTC Normal Pod gr-haproxy-mysql-2 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-489c7437-3224-4ed8-ab1f-64df89cea6b5" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:38:54 +0000 UTC Normal Pod gr-haproxy-mysql-2.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:38:55 +0000 UTC Normal Pod gr-haproxy-mysql-2.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 457.357772ms logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:38:55 +0000 UTC Normal Pod gr-haproxy-mysql-2.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:38:55 +0000 UTC Normal Pod gr-haproxy-mysql-2.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:38:56 +0000 UTC Normal Pod gr-haproxy-mysql-2.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:38:57 +0000 UTC Normal Pod gr-haproxy-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 383.612388ms logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:38:57 +0000 UTC Normal Pod gr-haproxy-mysql-2.spec.containers{mysql} Created Created container mysql logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:38:57 +0000 UTC Normal Pod gr-haproxy-mysql-2.spec.containers{mysql} Started Started container mysql logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:38:57 +0000 UTC Normal Pod gr-haproxy-mysql-2.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:38:57 +0000 UTC Normal Pod gr-haproxy-mysql-2.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 354.595766ms logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:38:57 +0000 UTC Normal Pod gr-haproxy-mysql-2.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:38:57 +0000 UTC Normal Pod gr-haproxy-mysql-2.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:39:28 +0000 UTC Warning Pod gr-haproxy-mysql-2.spec.containers{mysql} Unhealthy Startup probe failed: logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:39:28 +0000 UTC Normal Pod gr-haproxy-mysql-2.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted logger.go:42: 12:40:46 | gr-haproxy | 2023-08-09 12:39:28 +0000 UTC Normal Pod gr-haproxy-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 412.30187ms logger.go:42: 12:40:46 | gr-haproxy | Deleting namespace: kuttl-test-direct-ray case.go:114: context deadline exceeded === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (471.52s) --- FAIL: kuttl/harness (0.00s) --- FAIL: kuttl/harness/gr-haproxy (468.10s) FAIL