=== RUN   kuttl
    harness.go:462: starting setup
    harness.go:252: running tests using configured kubeconfig.
    harness.go:275: Successful connection to cluster at: https://34.133.14.34
    harness.go:360: running tests
    harness.go:73: going to run test suite with timeout of 180 seconds for each step
    harness.go:372: testsuite: e2e-tests/tests has 26 tests
=== RUN   kuttl/harness
=== RUN   kuttl/harness/haproxy
=== PAUSE kuttl/harness/haproxy
=== CONT  kuttl/harness/haproxy
    logger.go:42: 12:45:50 | haproxy | Creating namespace: kuttl-test-live-kingfish
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | starting test step 0-deploy-operator
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        deploy_operator
        deploy_non_tls_cluster_secrets
        deploy_tls_cluster_secrets
        deploy_client]
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | + source ../../functions
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ realpath ../../..
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | ++++ mktemp -d
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ export TEMP_DIR=/tmp/tmp.QLDg2zJSDI
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ TEMP_DIR=/tmp/tmp.QLDg2zJSDI
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ export GIT_BRANCH=PR-339
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ GIT_BRANCH=PR-339
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ export VERSION=PR-339-3a8483b
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ VERSION=PR-339-3a8483b
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ export PMM_SERVER_VERSION=9.9.9
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ PMM_SERVER_VERSION=9.9.9
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ export IMAGE_PMM_SERVER_TAG=dev-latest
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ IMAGE_PMM_SERVER_TAG=dev-latest
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | ++++ which gdate
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | ++++ which date
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ date=/usr/bin/date
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ command -v oc
    logger.go:42: 12:45:50 | haproxy/0-deploy-operator | +++ oc get projects
    logger.go:42: 12:45:56 | haproxy/0-deploy-operator | error: the server doesn't have a resource type "projects"
    logger.go:42: 12:45:56 | haproxy/0-deploy-operator | +++ kubectl get nodes
    logger.go:42: 12:45:56 | haproxy/0-deploy-operator | +++ grep '^minikube'
    logger.go:42: 12:45:57 | haproxy/0-deploy-operator | ++++ pwd
    logger.go:42: 12:45:57 | haproxy/0-deploy-operator | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/haproxy
    logger.go:42: 12:45:57 | haproxy/0-deploy-operator | ++ test_name=haproxy
    logger.go:42: 12:45:57 | haproxy/0-deploy-operator | + deploy_operator
    logger.go:42: 12:45:57 | haproxy/0-deploy-operator | + kubectl -n kuttl-test-live-kingfish apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/crd.yaml
    logger.go:42: 12:45:58 | haproxy/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlbackups.ps.percona.com serverside-applied
    logger.go:42: 12:45:58 | haproxy/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlrestores.ps.percona.com serverside-applied
    logger.go:42: 12:45:59 | haproxy/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqls.ps.percona.com serverside-applied
    logger.go:42: 12:45:59 | haproxy/0-deploy-operator | + kubectl -n kuttl-test-live-kingfish apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/rbac.yaml
    logger.go:42: 12:46:00 | haproxy/0-deploy-operator | serviceaccount/percona-server-mysql-operator created
    logger.go:42: 12:46:01 | haproxy/0-deploy-operator | serviceaccount/percona-server-mysql-operator-orchestrator created
    logger.go:42: 12:46:01 | haproxy/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created
    logger.go:42: 12:46:01 | haproxy/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator created
    logger.go:42: 12:46:02 | haproxy/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created
    logger.go:42: 12:46:02 | haproxy/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created
    logger.go:42: 12:46:02 | haproxy/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator created
    logger.go:42: 12:46:03 | haproxy/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created
    logger.go:42: 12:46:03 | haproxy/0-deploy-operator | + kubectl -n kuttl-test-live-kingfish apply -f -
    logger.go:42: 12:46:03 | haproxy/0-deploy-operator | + yq eval '(select(documentIndex==1).spec.template.spec.containers[] | select(.name=="manager").env[] | select(.name=="DISABLE_TELEMETRY").value) = "true"'
    logger.go:42: 12:46:03 | haproxy/0-deploy-operator | ++ printf 'select(documentIndex==1).spec.template.spec.containers[0].image="%s"' perconalab/percona-server-mysql-operator:PR-339-3a8483b
    logger.go:42: 12:46:03 | haproxy/0-deploy-operator | + yq eval 'select(documentIndex==1).spec.template.spec.containers[0].image="perconalab/percona-server-mysql-operator:PR-339-3a8483b"' /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/operator.yaml
    logger.go:42: 12:46:04 | haproxy/0-deploy-operator | configmap/percona-server-mysql-operator-config created
    logger.go:42: 12:46:04 | haproxy/0-deploy-operator | deployment.apps/percona-server-mysql-operator created
    logger.go:42: 12:46:04 | haproxy/0-deploy-operator | + deploy_non_tls_cluster_secrets
    logger.go:42: 12:46:04 | haproxy/0-deploy-operator | + kubectl -n kuttl-test-live-kingfish apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/secrets.yaml
    logger.go:42: 12:46:05 | haproxy/0-deploy-operator | secret/test-secrets created
    logger.go:42: 12:46:05 | haproxy/0-deploy-operator | + deploy_tls_cluster_secrets
    logger.go:42: 12:46:05 | haproxy/0-deploy-operator | + kubectl -n kuttl-test-live-kingfish apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/ssl-secret.yaml
    logger.go:42: 12:46:07 | haproxy/0-deploy-operator | secret/test-ssl created
    logger.go:42: 12:46:07 | haproxy/0-deploy-operator | + deploy_client
    logger.go:42: 12:46:07 | haproxy/0-deploy-operator | + kubectl -n kuttl-test-live-kingfish apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/client.yaml
    logger.go:42: 12:46:08 | haproxy/0-deploy-operator | pod/mysql-client created
    logger.go:42: 12:46:15 | haproxy/0-deploy-operator | test step completed 0-deploy-operator
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | starting test step 1-create-cluster
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        get_cr \
          | yq eval '.spec.allowUnsafeConfigurations=false' - \
          | yq eval '.spec.mysql.clusterType="async"' - \
          | yq eval '.spec.orchestrator.enabled=true' - \
          | yq eval '.spec.proxy.haproxy.enabled=true' - \
          | yq eval '.spec.proxy.haproxy.size=3' - \
          | kubectl -n "${NAMESPACE}" apply -f -]
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | + source ../../functions
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ realpath ../../..
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | ++++ mktemp -d
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ export TEMP_DIR=/tmp/tmp.cfKMFDfdi9
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ TEMP_DIR=/tmp/tmp.cfKMFDfdi9
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ export GIT_BRANCH=PR-339
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ GIT_BRANCH=PR-339
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ export VERSION=PR-339-3a8483b
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ VERSION=PR-339-3a8483b
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ export PMM_SERVER_VERSION=9.9.9
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ PMM_SERVER_VERSION=9.9.9
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ export IMAGE_PMM_SERVER_TAG=dev-latest
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ IMAGE_PMM_SERVER_TAG=dev-latest
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | ++++ which gdate
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | ++++ which date
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ date=/usr/bin/date
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ command -v oc
    logger.go:42: 12:46:15 | haproxy/1-create-cluster | +++ oc get projects
    logger.go:42: 12:46:21 | haproxy/1-create-cluster | error: the server doesn't have a resource type "projects"
    logger.go:42: 12:46:21 | haproxy/1-create-cluster | +++ kubectl get nodes
    logger.go:42: 12:46:21 | haproxy/1-create-cluster | +++ grep '^minikube'
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | ++++ pwd
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/haproxy
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | ++ test_name=haproxy
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | + get_cr
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | + local name_suffix=
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | + yq eval '.spec.mysql.clusterType="async"' -
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | + yq eval '.spec.sslSecretName="test-ssl"' -
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | + yq eval .spec.proxy.haproxy.size=3 -
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | ++ printf '.spec.initImage="%s"' perconalab/percona-server-mysql-operator:PR-339-3a8483b
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | + yq eval .spec.proxy.haproxy.enabled=true -
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | + yq eval .spec.orchestrator.enabled=true -
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | + kubectl -n kuttl-test-live-kingfish apply -f -
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | + yq eval '.spec.upgradeOptions.apply="disabled"' -
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | + yq eval '.spec.initImage="perconalab/percona-server-mysql-operator:PR-339-3a8483b"' -
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | + yq eval '.spec.mysql.clusterType="async"' -
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | + '[' -n '' ']'
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | + yq eval -
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | ++ printf '.spec.toolkit.image="%s"' perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | + yq eval '.spec.toolkit.image="perconalab/percona-server-mysql-operator:main-toolkit"' -
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | ++ printf '.spec.proxy.router.image="%s"' perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | + yq eval '.spec.proxy.router.image="perconalab/percona-server-mysql-operator:main-router"' -
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | + yq eval '.spec.secretsName="test-secrets"' -
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | ++ printf '.spec.pmm.image="%s"' perconalab/pmm-client:dev-latest
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | + yq eval .spec.allowUnsafeConfigurations=false -
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | + yq eval '.spec.pmm.image="perconalab/pmm-client:dev-latest"' -
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | ++ printf '.spec.mysql.image="%s"' perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | + yq eval '.spec.mysql.image="perconalab/percona-server-mysql-operator:main-psmysql"' -
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | ++ printf '.metadata.name="%s"' haproxy
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | ++ printf '.spec.proxy.haproxy.image="%s"' perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | + yq eval '.spec.proxy.haproxy.image="perconalab/percona-server-mysql-operator:main-haproxy"' -
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | + yq eval '.metadata.name="haproxy"' /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/cr.yaml
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | ++ printf '.spec.backup.image="%s"' perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | ++ printf '.spec.orchestrator.image="%s"' perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | + yq eval '.spec.backup.image="perconalab/percona-server-mysql-operator:main-backup"' -
    logger.go:42: 12:46:22 | haproxy/1-create-cluster | + yq eval '.spec.orchestrator.image="perconalab/percona-server-mysql-operator:main-orchestrator"' -
    logger.go:42: 12:46:23 | haproxy/1-create-cluster | perconaservermysql.ps.percona.com/haproxy created
    logger.go:42: 12:49:44 | haproxy/1-create-cluster | test step completed 1-create-cluster
    logger.go:42: 12:49:44 | haproxy/2-write-data | starting test step 2-write-data
    logger.go:42: 12:49:44 | haproxy/2-write-data | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        run_mysql \
        	"CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" \
        	"-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password"
        
        run_mysql \
        	"INSERT myDB.myTable (id) VALUES (100500)" \
        	"-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password"]
    logger.go:42: 12:49:44 | haproxy/2-write-data | + source ../../functions
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ realpath ../../..
    logger.go:42: 12:49:44 | haproxy/2-write-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:49:44 | haproxy/2-write-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf
    logger.go:42: 12:49:44 | haproxy/2-write-data | ++++ mktemp -d
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ export TEMP_DIR=/tmp/tmp.iEizhR10pn
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ TEMP_DIR=/tmp/tmp.iEizhR10pn
    logger.go:42: 12:49:44 | haproxy/2-write-data | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ export GIT_BRANCH=PR-339
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ GIT_BRANCH=PR-339
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ export VERSION=PR-339-3a8483b
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ VERSION=PR-339-3a8483b
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ export PMM_SERVER_VERSION=9.9.9
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ PMM_SERVER_VERSION=9.9.9
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ export IMAGE_PMM_SERVER_TAG=dev-latest
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ IMAGE_PMM_SERVER_TAG=dev-latest
    logger.go:42: 12:49:44 | haproxy/2-write-data | ++++ which gdate
    logger.go:42: 12:49:44 | haproxy/2-write-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 12:49:44 | haproxy/2-write-data | ++++ which date
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ date=/usr/bin/date
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ command -v oc
    logger.go:42: 12:49:44 | haproxy/2-write-data | +++ oc get projects
    logger.go:42: 12:49:50 | haproxy/2-write-data | error: the server doesn't have a resource type "projects"
    logger.go:42: 12:49:50 | haproxy/2-write-data | +++ kubectl get nodes
    logger.go:42: 12:49:50 | haproxy/2-write-data | +++ grep '^minikube'
    logger.go:42: 12:49:50 | haproxy/2-write-data | ++++ pwd
    logger.go:42: 12:49:50 | haproxy/2-write-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/haproxy
    logger.go:42: 12:49:50 | haproxy/2-write-data | ++ test_name=haproxy
    logger.go:42: 12:49:50 | haproxy/2-write-data | +++ get_cluster_name
    logger.go:42: 12:49:50 | haproxy/2-write-data | +++ kubectl -n kuttl-test-live-kingfish get ps -o 'jsonpath={.items[0].metadata.name}'
    logger.go:42: 12:49:51 | haproxy/2-write-data | ++ get_haproxy_svc haproxy
    logger.go:42: 12:49:51 | haproxy/2-write-data | ++ local cluster=haproxy
    logger.go:42: 12:49:51 | haproxy/2-write-data | ++ echo haproxy-haproxy
    logger.go:42: 12:49:51 | haproxy/2-write-data | + run_mysql 'CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' '-h haproxy-haproxy -uroot -proot_password'
    logger.go:42: 12:49:51 | haproxy/2-write-data | + local 'command=CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)'
    logger.go:42: 12:49:51 | haproxy/2-write-data | + local 'uri=-h haproxy-haproxy -uroot -proot_password'
    logger.go:42: 12:49:51 | haproxy/2-write-data | + local pod=
    logger.go:42: 12:49:51 | haproxy/2-write-data | ++ get_client_pod
    logger.go:42: 12:49:51 | haproxy/2-write-data | ++ kubectl -n kuttl-test-live-kingfish get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 12:49:52 | haproxy/2-write-data | + client_pod=mysql-client
    logger.go:42: 12:49:52 | haproxy/2-write-data | + wait_pod mysql-client
    logger.go:42: 12:49:52 | haproxy/2-write-data | + local pod=mysql-client
    logger.go:42: 12:49:52 | haproxy/2-write-data | + set +o xtrace
    logger.go:42: 12:49:53 | haproxy/2-write-data | mysql-clienttrue
    logger.go:42: 12:49:53 | haproxy/2-write-data | + kubectl -n kuttl-test-live-kingfish exec mysql-client -- bash -c 'printf '\''%s\n'\'' "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" | mysql -sN -h haproxy-haproxy -uroot -proot_password'
    logger.go:42: 12:49:53 | haproxy/2-write-data | + sed -e 's/mysql: //'
    logger.go:42: 12:49:53 | haproxy/2-write-data | + grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 12:49:54 | haproxy/2-write-data | + :
    logger.go:42: 12:49:54 | haproxy/2-write-data | +++ get_cluster_name
    logger.go:42: 12:49:54 | haproxy/2-write-data | +++ kubectl -n kuttl-test-live-kingfish get ps -o 'jsonpath={.items[0].metadata.name}'
    logger.go:42: 12:49:55 | haproxy/2-write-data | ++ get_haproxy_svc haproxy
    logger.go:42: 12:49:55 | haproxy/2-write-data | ++ local cluster=haproxy
    logger.go:42: 12:49:55 | haproxy/2-write-data | ++ echo haproxy-haproxy
    logger.go:42: 12:49:55 | haproxy/2-write-data | + run_mysql 'INSERT myDB.myTable (id) VALUES (100500)' '-h haproxy-haproxy -uroot -proot_password'
    logger.go:42: 12:49:55 | haproxy/2-write-data | + local 'command=INSERT myDB.myTable (id) VALUES (100500)'
    logger.go:42: 12:49:55 | haproxy/2-write-data | + local 'uri=-h haproxy-haproxy -uroot -proot_password'
    logger.go:42: 12:49:55 | haproxy/2-write-data | + local pod=
    logger.go:42: 12:49:55 | haproxy/2-write-data | ++ get_client_pod
    logger.go:42: 12:49:55 | haproxy/2-write-data | ++ kubectl -n kuttl-test-live-kingfish get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 12:49:56 | haproxy/2-write-data | + client_pod=mysql-client
    logger.go:42: 12:49:56 | haproxy/2-write-data | + wait_pod mysql-client
    logger.go:42: 12:49:56 | haproxy/2-write-data | + local pod=mysql-client
    logger.go:42: 12:49:56 | haproxy/2-write-data | + set +o xtrace
    logger.go:42: 12:49:56 | haproxy/2-write-data | mysql-clienttrue
    logger.go:42: 12:49:56 | haproxy/2-write-data | + kubectl -n kuttl-test-live-kingfish exec mysql-client -- bash -c 'printf '\''%s\n'\'' "INSERT myDB.myTable (id) VALUES (100500)" | mysql -sN -h haproxy-haproxy -uroot -proot_password'
    logger.go:42: 12:49:56 | haproxy/2-write-data | + sed -e 's/mysql: //'
    logger.go:42: 12:49:56 | haproxy/2-write-data | + grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 12:49:58 | haproxy/2-write-data | + :
    logger.go:42: 12:49:59 | haproxy/2-write-data | test step completed 2-write-data
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | starting test step 3-read-from-primary
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        data=$(run_mysql "SELECT * FROM myDB.myTable" "-h $(get_haproxy_svc $(get_cluster_name)) -P3306 -uroot -proot_password")
        
        kubectl create configmap -n "${NAMESPACE}" 03-read-from-primary --from-literal=data="${data}"]
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | + source ../../functions
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ realpath ../../..
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | ++++ mktemp -d
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ export TEMP_DIR=/tmp/tmp.je9Z958BaK
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ TEMP_DIR=/tmp/tmp.je9Z958BaK
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ export GIT_BRANCH=PR-339
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ GIT_BRANCH=PR-339
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ export VERSION=PR-339-3a8483b
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ VERSION=PR-339-3a8483b
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ export PMM_SERVER_VERSION=9.9.9
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ PMM_SERVER_VERSION=9.9.9
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ export IMAGE_PMM_SERVER_TAG=dev-latest
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ IMAGE_PMM_SERVER_TAG=dev-latest
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | ++++ which gdate
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | ++++ which date
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ date=/usr/bin/date
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ command -v oc
    logger.go:42: 12:49:59 | haproxy/3-read-from-primary | +++ oc get projects
    logger.go:42: 12:50:04 | haproxy/3-read-from-primary | error: the server doesn't have a resource type "projects"
    logger.go:42: 12:50:04 | haproxy/3-read-from-primary | +++ kubectl get nodes
    logger.go:42: 12:50:04 | haproxy/3-read-from-primary | +++ grep '^minikube'
    logger.go:42: 12:50:05 | haproxy/3-read-from-primary | ++++ pwd
    logger.go:42: 12:50:05 | haproxy/3-read-from-primary | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/haproxy
    logger.go:42: 12:50:05 | haproxy/3-read-from-primary | ++ test_name=haproxy
    logger.go:42: 12:50:05 | haproxy/3-read-from-primary | ++++ get_cluster_name
    logger.go:42: 12:50:05 | haproxy/3-read-from-primary | ++++ kubectl -n kuttl-test-live-kingfish get ps -o 'jsonpath={.items[0].metadata.name}'
    logger.go:42: 12:50:06 | haproxy/3-read-from-primary | +++ get_haproxy_svc haproxy
    logger.go:42: 12:50:06 | haproxy/3-read-from-primary | +++ local cluster=haproxy
    logger.go:42: 12:50:06 | haproxy/3-read-from-primary | +++ echo haproxy-haproxy
    logger.go:42: 12:50:06 | haproxy/3-read-from-primary | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h haproxy-haproxy -P3306 -uroot -proot_password'
    logger.go:42: 12:50:06 | haproxy/3-read-from-primary | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 12:50:06 | haproxy/3-read-from-primary | ++ local 'uri=-h haproxy-haproxy -P3306 -uroot -proot_password'
    logger.go:42: 12:50:06 | haproxy/3-read-from-primary | ++ local pod=
    logger.go:42: 12:50:06 | haproxy/3-read-from-primary | +++ get_client_pod
    logger.go:42: 12:50:06 | haproxy/3-read-from-primary | +++ kubectl -n kuttl-test-live-kingfish get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 12:50:07 | haproxy/3-read-from-primary | ++ client_pod=mysql-client
    logger.go:42: 12:50:07 | haproxy/3-read-from-primary | ++ wait_pod mysql-client
    logger.go:42: 12:50:07 | haproxy/3-read-from-primary | ++ local pod=mysql-client
    logger.go:42: 12:50:07 | haproxy/3-read-from-primary | ++ set +o xtrace
    logger.go:42: 12:50:07 | haproxy/3-read-from-primary | mysql-clienttrue
    logger.go:42: 12:50:07 | haproxy/3-read-from-primary | ++ kubectl -n kuttl-test-live-kingfish exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h haproxy-haproxy -P3306 -uroot -proot_password'
    logger.go:42: 12:50:07 | haproxy/3-read-from-primary | ++ sed -e 's/mysql: //'
    logger.go:42: 12:50:07 | haproxy/3-read-from-primary | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 12:50:09 | haproxy/3-read-from-primary | + data=100500
    logger.go:42: 12:50:09 | haproxy/3-read-from-primary | + kubectl create configmap -n kuttl-test-live-kingfish 03-read-from-primary --from-literal=data=100500
    logger.go:42: 12:50:10 | haproxy/3-read-from-primary | configmap/03-read-from-primary created
    logger.go:42: 12:50:11 | haproxy/3-read-from-primary | test step completed 3-read-from-primary
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | starting test step 4-read-from-replicas
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        data=$(run_mysql "SELECT * FROM myDB.myTable" "-h $(get_haproxy_svc $(get_cluster_name)) -P3307 -uroot -proot_password")
        kubectl create configmap -n "${NAMESPACE}" 04-read-from-replicas --from-literal=${test_name}-haproxy-replicas=${data}]
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | + source ../../functions
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ realpath ../../..
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | ++++ mktemp -d
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ export TEMP_DIR=/tmp/tmp.Yaacvwp1DE
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ TEMP_DIR=/tmp/tmp.Yaacvwp1DE
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ export GIT_BRANCH=PR-339
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ GIT_BRANCH=PR-339
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ export VERSION=PR-339-3a8483b
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ VERSION=PR-339-3a8483b
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ export PMM_SERVER_VERSION=9.9.9
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ PMM_SERVER_VERSION=9.9.9
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ export IMAGE_PMM_SERVER_TAG=dev-latest
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ IMAGE_PMM_SERVER_TAG=dev-latest
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | ++++ which gdate
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | ++++ which date
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ date=/usr/bin/date
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ command -v oc
    logger.go:42: 12:50:11 | haproxy/4-read-from-replicas | +++ oc get projects
    logger.go:42: 12:50:16 | haproxy/4-read-from-replicas | error: the server doesn't have a resource type "projects"
    logger.go:42: 12:50:16 | haproxy/4-read-from-replicas | +++ grep '^minikube'
    logger.go:42: 12:50:16 | haproxy/4-read-from-replicas | +++ kubectl get nodes
    logger.go:42: 12:50:17 | haproxy/4-read-from-replicas | ++++ pwd
    logger.go:42: 12:50:17 | haproxy/4-read-from-replicas | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/haproxy
    logger.go:42: 12:50:17 | haproxy/4-read-from-replicas | ++ test_name=haproxy
    logger.go:42: 12:50:17 | haproxy/4-read-from-replicas | ++++ get_cluster_name
    logger.go:42: 12:50:17 | haproxy/4-read-from-replicas | ++++ kubectl -n kuttl-test-live-kingfish get ps -o 'jsonpath={.items[0].metadata.name}'
    logger.go:42: 12:50:18 | haproxy/4-read-from-replicas | +++ get_haproxy_svc haproxy
    logger.go:42: 12:50:18 | haproxy/4-read-from-replicas | +++ local cluster=haproxy
    logger.go:42: 12:50:18 | haproxy/4-read-from-replicas | +++ echo haproxy-haproxy
    logger.go:42: 12:50:18 | haproxy/4-read-from-replicas | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h haproxy-haproxy -P3307 -uroot -proot_password'
    logger.go:42: 12:50:18 | haproxy/4-read-from-replicas | ++ local 'command=SELECT * FROM myDB.myTable'
    logger.go:42: 12:50:18 | haproxy/4-read-from-replicas | ++ local 'uri=-h haproxy-haproxy -P3307 -uroot -proot_password'
    logger.go:42: 12:50:18 | haproxy/4-read-from-replicas | ++ local pod=
    logger.go:42: 12:50:18 | haproxy/4-read-from-replicas | +++ get_client_pod
    logger.go:42: 12:50:18 | haproxy/4-read-from-replicas | +++ kubectl -n kuttl-test-live-kingfish get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 12:50:19 | haproxy/4-read-from-replicas | ++ client_pod=mysql-client
    logger.go:42: 12:50:19 | haproxy/4-read-from-replicas | ++ wait_pod mysql-client
    logger.go:42: 12:50:19 | haproxy/4-read-from-replicas | ++ local pod=mysql-client
    logger.go:42: 12:50:19 | haproxy/4-read-from-replicas | ++ set +o xtrace
    logger.go:42: 12:50:19 | haproxy/4-read-from-replicas | mysql-clienttrue
    logger.go:42: 12:50:19 | haproxy/4-read-from-replicas | ++ sed -e 's/mysql: //'
    logger.go:42: 12:50:19 | haproxy/4-read-from-replicas | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 12:50:19 | haproxy/4-read-from-replicas | ++ kubectl -n kuttl-test-live-kingfish exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h haproxy-haproxy -P3307 -uroot -proot_password'
    logger.go:42: 12:50:21 | haproxy/4-read-from-replicas | + data=100500
    logger.go:42: 12:50:21 | haproxy/4-read-from-replicas | + kubectl create configmap -n kuttl-test-live-kingfish 04-read-from-replicas --from-literal=haproxy-haproxy-replicas=100500
    logger.go:42: 12:50:22 | haproxy/4-read-from-replicas | configmap/04-read-from-replicas created
    logger.go:42: 12:50:23 | haproxy/4-read-from-replicas | test step completed 4-read-from-replicas
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | starting test step 5-check-pods-have-same-primary
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        data=()
        for i in $(seq 0 2); do
        	data+=("$(get_primary_from_haproxy ${test_name}-haproxy-$i)")
        done
        if [ "${data[0]}" != "${data[1]}" -o "${data[1]}" != "${data[2]}" ]; then
          echo "Not all haproxy pods point to same primary: 0: ${data[0]} 1: ${data[1]} 2: ${data[2]}"
          exit 1
        fi]
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | + source ../../functions
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ realpath ../../..
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | ++++ mktemp -d
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ export TEMP_DIR=/tmp/tmp.zTZ3bpqQ2H
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ TEMP_DIR=/tmp/tmp.zTZ3bpqQ2H
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ export GIT_BRANCH=PR-339
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ GIT_BRANCH=PR-339
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ export VERSION=PR-339-3a8483b
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ VERSION=PR-339-3a8483b
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ export PMM_SERVER_VERSION=9.9.9
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ PMM_SERVER_VERSION=9.9.9
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ export IMAGE_PMM_SERVER_TAG=dev-latest
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ IMAGE_PMM_SERVER_TAG=dev-latest
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | ++++ which gdate
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | ++++ which date
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ date=/usr/bin/date
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ command -v oc
    logger.go:42: 12:50:23 | haproxy/5-check-pods-have-same-primary | +++ oc get projects
    logger.go:42: 12:50:28 | haproxy/5-check-pods-have-same-primary | error: the server doesn't have a resource type "projects"
    logger.go:42: 12:50:28 | haproxy/5-check-pods-have-same-primary | +++ kubectl get nodes
    logger.go:42: 12:50:28 | haproxy/5-check-pods-have-same-primary | +++ grep '^minikube'
    logger.go:42: 12:50:29 | haproxy/5-check-pods-have-same-primary | ++++ pwd
    logger.go:42: 12:50:29 | haproxy/5-check-pods-have-same-primary | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/haproxy
    logger.go:42: 12:50:29 | haproxy/5-check-pods-have-same-primary | ++ test_name=haproxy
    logger.go:42: 12:50:29 | haproxy/5-check-pods-have-same-primary | + data=()
    logger.go:42: 12:50:29 | haproxy/5-check-pods-have-same-primary | ++ seq 0 2
    logger.go:42: 12:50:29 | haproxy/5-check-pods-have-same-primary | + for i in '$(seq 0 2)'
    logger.go:42: 12:50:29 | haproxy/5-check-pods-have-same-primary | + data+=("$(get_primary_from_haproxy ${test_name}-haproxy-$i)")
    logger.go:42: 12:50:29 | haproxy/5-check-pods-have-same-primary | ++ get_primary_from_haproxy haproxy-haproxy-0
    logger.go:42: 12:50:29 | haproxy/5-check-pods-have-same-primary | ++ local haproxy_pod=haproxy-haproxy-0
    logger.go:42: 12:50:29 | haproxy/5-check-pods-have-same-primary | +++ kubectl -n kuttl-test-live-kingfish get pods haproxy-haproxy-0 -o 'jsonpath={.status.podIP}'
    logger.go:42: 12:50:30 | haproxy/5-check-pods-have-same-primary | ++ local haproxy_pod_ip=10.117.162.27
    logger.go:42: 12:50:30 | haproxy/5-check-pods-have-same-primary | ++ run_mysql 'SHOW VARIABLES LIKE '\''%hostname%'\'';' '-h 10.117.162.27 -P3306 -uroot -proot_password'
    logger.go:42: 12:50:30 | haproxy/5-check-pods-have-same-primary | ++ awk '{print $2}'
    logger.go:42: 12:50:30 | haproxy/5-check-pods-have-same-primary | ++ local 'command=SHOW VARIABLES LIKE '\''%hostname%'\'';'
    logger.go:42: 12:50:30 | haproxy/5-check-pods-have-same-primary | ++ local 'uri=-h 10.117.162.27 -P3306 -uroot -proot_password'
    logger.go:42: 12:50:30 | haproxy/5-check-pods-have-same-primary | ++ local pod=
    logger.go:42: 12:50:30 | haproxy/5-check-pods-have-same-primary | +++ get_client_pod
    logger.go:42: 12:50:30 | haproxy/5-check-pods-have-same-primary | +++ kubectl -n kuttl-test-live-kingfish get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 12:50:31 | haproxy/5-check-pods-have-same-primary | ++ client_pod=mysql-client
    logger.go:42: 12:50:31 | haproxy/5-check-pods-have-same-primary | ++ wait_pod mysql-client
    logger.go:42: 12:50:31 | haproxy/5-check-pods-have-same-primary | ++ local pod=mysql-client
    logger.go:42: 12:50:31 | haproxy/5-check-pods-have-same-primary | ++ set +o xtrace
    logger.go:42: 12:50:32 | haproxy/5-check-pods-have-same-primary | mysql-clienttrue
    logger.go:42: 12:50:32 | haproxy/5-check-pods-have-same-primary | ++ sed -e 's/mysql: //'
    logger.go:42: 12:50:32 | haproxy/5-check-pods-have-same-primary | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 12:50:32 | haproxy/5-check-pods-have-same-primary | ++ kubectl -n kuttl-test-live-kingfish exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SHOW VARIABLES LIKE '\''%hostname%'\'';" | mysql -sN -h 10.117.162.27 -P3306 -uroot -proot_password'
    logger.go:42: 12:50:33 | haproxy/5-check-pods-have-same-primary | + for i in '$(seq 0 2)'
    logger.go:42: 12:50:33 | haproxy/5-check-pods-have-same-primary | + data+=("$(get_primary_from_haproxy ${test_name}-haproxy-$i)")
    logger.go:42: 12:50:33 | haproxy/5-check-pods-have-same-primary | ++ get_primary_from_haproxy haproxy-haproxy-1
    logger.go:42: 12:50:33 | haproxy/5-check-pods-have-same-primary | ++ local haproxy_pod=haproxy-haproxy-1
    logger.go:42: 12:50:33 | haproxy/5-check-pods-have-same-primary | +++ kubectl -n kuttl-test-live-kingfish get pods haproxy-haproxy-1 -o 'jsonpath={.status.podIP}'
    logger.go:42: 12:50:34 | haproxy/5-check-pods-have-same-primary | ++ local haproxy_pod_ip=10.117.160.24
    logger.go:42: 12:50:34 | haproxy/5-check-pods-have-same-primary | ++ run_mysql 'SHOW VARIABLES LIKE '\''%hostname%'\'';' '-h 10.117.160.24 -P3306 -uroot -proot_password'
    logger.go:42: 12:50:34 | haproxy/5-check-pods-have-same-primary | ++ local 'command=SHOW VARIABLES LIKE '\''%hostname%'\'';'
    logger.go:42: 12:50:34 | haproxy/5-check-pods-have-same-primary | ++ awk '{print $2}'
    logger.go:42: 12:50:34 | haproxy/5-check-pods-have-same-primary | ++ local 'uri=-h 10.117.160.24 -P3306 -uroot -proot_password'
    logger.go:42: 12:50:34 | haproxy/5-check-pods-have-same-primary | ++ local pod=
    logger.go:42: 12:50:34 | haproxy/5-check-pods-have-same-primary | +++ get_client_pod
    logger.go:42: 12:50:34 | haproxy/5-check-pods-have-same-primary | +++ kubectl -n kuttl-test-live-kingfish get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 12:50:35 | haproxy/5-check-pods-have-same-primary | ++ client_pod=mysql-client
    logger.go:42: 12:50:35 | haproxy/5-check-pods-have-same-primary | ++ wait_pod mysql-client
    logger.go:42: 12:50:35 | haproxy/5-check-pods-have-same-primary | ++ local pod=mysql-client
    logger.go:42: 12:50:35 | haproxy/5-check-pods-have-same-primary | ++ set +o xtrace
    logger.go:42: 12:50:35 | haproxy/5-check-pods-have-same-primary | mysql-clienttrue
    logger.go:42: 12:50:35 | haproxy/5-check-pods-have-same-primary | ++ kubectl -n kuttl-test-live-kingfish exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SHOW VARIABLES LIKE '\''%hostname%'\'';" | mysql -sN -h 10.117.160.24 -P3306 -uroot -proot_password'
    logger.go:42: 12:50:35 | haproxy/5-check-pods-have-same-primary | ++ sed -e 's/mysql: //'
    logger.go:42: 12:50:35 | haproxy/5-check-pods-have-same-primary | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 12:50:37 | haproxy/5-check-pods-have-same-primary | + for i in '$(seq 0 2)'
    logger.go:42: 12:50:37 | haproxy/5-check-pods-have-same-primary | + data+=("$(get_primary_from_haproxy ${test_name}-haproxy-$i)")
    logger.go:42: 12:50:37 | haproxy/5-check-pods-have-same-primary | ++ get_primary_from_haproxy haproxy-haproxy-2
    logger.go:42: 12:50:37 | haproxy/5-check-pods-have-same-primary | ++ local haproxy_pod=haproxy-haproxy-2
    logger.go:42: 12:50:37 | haproxy/5-check-pods-have-same-primary | +++ kubectl -n kuttl-test-live-kingfish get pods haproxy-haproxy-2 -o 'jsonpath={.status.podIP}'
    logger.go:42: 12:50:38 | haproxy/5-check-pods-have-same-primary | ++ local haproxy_pod_ip=10.117.161.31
    logger.go:42: 12:50:38 | haproxy/5-check-pods-have-same-primary | ++ run_mysql 'SHOW VARIABLES LIKE '\''%hostname%'\'';' '-h 10.117.161.31 -P3306 -uroot -proot_password'
    logger.go:42: 12:50:38 | haproxy/5-check-pods-have-same-primary | ++ local 'command=SHOW VARIABLES LIKE '\''%hostname%'\'';'
    logger.go:42: 12:50:38 | haproxy/5-check-pods-have-same-primary | ++ local 'uri=-h 10.117.161.31 -P3306 -uroot -proot_password'
    logger.go:42: 12:50:38 | haproxy/5-check-pods-have-same-primary | ++ local pod=
    logger.go:42: 12:50:38 | haproxy/5-check-pods-have-same-primary | ++ awk '{print $2}'
    logger.go:42: 12:50:38 | haproxy/5-check-pods-have-same-primary | +++ get_client_pod
    logger.go:42: 12:50:38 | haproxy/5-check-pods-have-same-primary | +++ kubectl -n kuttl-test-live-kingfish get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 12:50:38 | haproxy/5-check-pods-have-same-primary | ++ client_pod=mysql-client
    logger.go:42: 12:50:38 | haproxy/5-check-pods-have-same-primary | ++ wait_pod mysql-client
    logger.go:42: 12:50:38 | haproxy/5-check-pods-have-same-primary | ++ local pod=mysql-client
    logger.go:42: 12:50:38 | haproxy/5-check-pods-have-same-primary | ++ set +o xtrace
    logger.go:42: 12:50:39 | haproxy/5-check-pods-have-same-primary | mysql-clienttrue
    logger.go:42: 12:50:39 | haproxy/5-check-pods-have-same-primary | ++ kubectl -n kuttl-test-live-kingfish exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SHOW VARIABLES LIKE '\''%hostname%'\'';" | mysql -sN -h 10.117.161.31 -P3306 -uroot -proot_password'
    logger.go:42: 12:50:39 | haproxy/5-check-pods-have-same-primary | ++ sed -e 's/mysql: //'
    logger.go:42: 12:50:39 | haproxy/5-check-pods-have-same-primary | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 12:50:41 | haproxy/5-check-pods-have-same-primary | + '[' haproxy-mysql-0 '!=' haproxy-mysql-0 -o haproxy-mysql-0 '!=' haproxy-mysql-0 ']'
    logger.go:42: 12:50:42 | haproxy/5-check-pods-have-same-primary | test step completed 5-check-pods-have-same-primary
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | starting test step 6-check-label-haproxy-primary
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        primary_pod_from_label="$(get_primary_from_label)"
        primary_pod_from_haproxy="$(get_primary_from_haproxy ${test_name}-haproxy-0)"
        if [ "${primary_pod_from_label}" != "${primary_pod_from_haproxy}" ]; then
          echo "Primary in k8s label (${primary_pod_from_label}) is not set to same pod as in haproxy (${primary_pod_from_haproxy})!"
          exit 1
        fi]
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | + source ../../functions
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ realpath ../../..
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | ++++ mktemp -d
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ export TEMP_DIR=/tmp/tmp.eXdy8Txyds
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ TEMP_DIR=/tmp/tmp.eXdy8Txyds
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ export GIT_BRANCH=PR-339
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ GIT_BRANCH=PR-339
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ export VERSION=PR-339-3a8483b
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ VERSION=PR-339-3a8483b
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ export PMM_SERVER_VERSION=9.9.9
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ PMM_SERVER_VERSION=9.9.9
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ export IMAGE_PMM_SERVER_TAG=dev-latest
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ IMAGE_PMM_SERVER_TAG=dev-latest
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | ++++ which gdate
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | ++++ which date
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ date=/usr/bin/date
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ command -v oc
    logger.go:42: 12:50:42 | haproxy/6-check-label-haproxy-primary | +++ oc get projects
    logger.go:42: 12:50:47 | haproxy/6-check-label-haproxy-primary | error: the server doesn't have a resource type "projects"
    logger.go:42: 12:50:47 | haproxy/6-check-label-haproxy-primary | +++ grep '^minikube'
    logger.go:42: 12:50:47 | haproxy/6-check-label-haproxy-primary | +++ kubectl get nodes
    logger.go:42: 12:50:48 | haproxy/6-check-label-haproxy-primary | ++++ pwd
    logger.go:42: 12:50:48 | haproxy/6-check-label-haproxy-primary | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/haproxy
    logger.go:42: 12:50:48 | haproxy/6-check-label-haproxy-primary | ++ test_name=haproxy
    logger.go:42: 12:50:48 | haproxy/6-check-label-haproxy-primary | ++ get_primary_from_label
    logger.go:42: 12:50:48 | haproxy/6-check-label-haproxy-primary | ++ kubectl -n kuttl-test-live-kingfish get pods -l mysql.percona.com/primary=true '-ojsonpath={.items[0].metadata.name}'
    logger.go:42: 12:50:49 | haproxy/6-check-label-haproxy-primary | + primary_pod_from_label=haproxy-mysql-0
    logger.go:42: 12:50:49 | haproxy/6-check-label-haproxy-primary | ++ get_primary_from_haproxy haproxy-haproxy-0
    logger.go:42: 12:50:49 | haproxy/6-check-label-haproxy-primary | ++ local haproxy_pod=haproxy-haproxy-0
    logger.go:42: 12:50:49 | haproxy/6-check-label-haproxy-primary | +++ kubectl -n kuttl-test-live-kingfish get pods haproxy-haproxy-0 -o 'jsonpath={.status.podIP}'
    logger.go:42: 12:50:49 | haproxy/6-check-label-haproxy-primary | ++ local haproxy_pod_ip=10.117.162.27
    logger.go:42: 12:50:49 | haproxy/6-check-label-haproxy-primary | ++ run_mysql 'SHOW VARIABLES LIKE '\''%hostname%'\'';' '-h 10.117.162.27 -P3306 -uroot -proot_password'
    logger.go:42: 12:50:49 | haproxy/6-check-label-haproxy-primary | ++ awk '{print $2}'
    logger.go:42: 12:50:49 | haproxy/6-check-label-haproxy-primary | ++ local 'command=SHOW VARIABLES LIKE '\''%hostname%'\'';'
    logger.go:42: 12:50:49 | haproxy/6-check-label-haproxy-primary | ++ local 'uri=-h 10.117.162.27 -P3306 -uroot -proot_password'
    logger.go:42: 12:50:49 | haproxy/6-check-label-haproxy-primary | ++ local pod=
    logger.go:42: 12:50:49 | haproxy/6-check-label-haproxy-primary | +++ get_client_pod
    logger.go:42: 12:50:49 | haproxy/6-check-label-haproxy-primary | +++ kubectl -n kuttl-test-live-kingfish get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 12:50:50 | haproxy/6-check-label-haproxy-primary | ++ client_pod=mysql-client
    logger.go:42: 12:50:50 | haproxy/6-check-label-haproxy-primary | ++ wait_pod mysql-client
    logger.go:42: 12:50:50 | haproxy/6-check-label-haproxy-primary | ++ local pod=mysql-client
    logger.go:42: 12:50:50 | haproxy/6-check-label-haproxy-primary | ++ set +o xtrace
    logger.go:42: 12:50:51 | haproxy/6-check-label-haproxy-primary | mysql-clienttrue
    logger.go:42: 12:50:51 | haproxy/6-check-label-haproxy-primary | ++ kubectl -n kuttl-test-live-kingfish exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SHOW VARIABLES LIKE '\''%hostname%'\'';" | mysql -sN -h 10.117.162.27 -P3306 -uroot -proot_password'
    logger.go:42: 12:50:51 | haproxy/6-check-label-haproxy-primary | ++ sed -e 's/mysql: //'
    logger.go:42: 12:50:51 | haproxy/6-check-label-haproxy-primary | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 12:50:52 | haproxy/6-check-label-haproxy-primary | + primary_pod_from_haproxy=haproxy-mysql-0
    logger.go:42: 12:50:52 | haproxy/6-check-label-haproxy-primary | + '[' haproxy-mysql-0 '!=' haproxy-mysql-0 ']'
    logger.go:42: 12:50:53 | haproxy/6-check-label-haproxy-primary | test step completed 6-check-label-haproxy-primary
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | starting test step 7-check-primary-failover
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        primary_pod_from_label="$(get_primary_from_label)"
        
        kubectl -n "${NAMESPACE}" delete pod ${primary_pod_from_label}
        sleep 3
        
        data=()
        for i in $(seq 0 2); do
        	data+=("$(get_primary_from_haproxy ${test_name}-haproxy-$i)")
        done
        if [ "${data[0]}" != "${data[1]}" -o "${data[1]}" != "${data[2]}" ]; then
          echo "Not all haproxy pods point to same primary: 0: ${data[0]} 1: ${data[1]} 2: ${data[2]}"
          exit 1
        fi
        
        primary_pod_from_label="$(get_primary_from_label)"
        primary_pod_from_haproxy="$(get_primary_from_haproxy ${test_name}-haproxy-0)"
        if [ "${primary_pod_from_label}" != "${primary_pod_from_haproxy}" ]; then
          echo "Primary in k8s label (${primary_pod_from_label}) is not set to same pod as in haproxy (${primary_pod_from_haproxy})!"
          exit 1
        fi]
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | + source ../../functions
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ realpath ../../..
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | ++++ mktemp -d
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ export TEMP_DIR=/tmp/tmp.Zv8UyZfZbd
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ TEMP_DIR=/tmp/tmp.Zv8UyZfZbd
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ export GIT_BRANCH=PR-339
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ GIT_BRANCH=PR-339
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ export VERSION=PR-339-3a8483b
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ VERSION=PR-339-3a8483b
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ export PMM_SERVER_VERSION=9.9.9
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ PMM_SERVER_VERSION=9.9.9
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ export IMAGE_PMM_SERVER_TAG=dev-latest
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ IMAGE_PMM_SERVER_TAG=dev-latest
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | ++++ which gdate
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | ++++ which date
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ date=/usr/bin/date
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ command -v oc
    logger.go:42: 12:50:53 | haproxy/7-check-primary-failover | +++ oc get projects
    logger.go:42: 12:50:59 | haproxy/7-check-primary-failover | error: the server doesn't have a resource type "projects"
    logger.go:42: 12:50:59 | haproxy/7-check-primary-failover | +++ kubectl get nodes
    logger.go:42: 12:50:59 | haproxy/7-check-primary-failover | +++ grep '^minikube'
    logger.go:42: 12:51:00 | haproxy/7-check-primary-failover | ++++ pwd
    logger.go:42: 12:51:00 | haproxy/7-check-primary-failover | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/haproxy
    logger.go:42: 12:51:00 | haproxy/7-check-primary-failover | ++ test_name=haproxy
    logger.go:42: 12:51:00 | haproxy/7-check-primary-failover | ++ get_primary_from_label
    logger.go:42: 12:51:00 | haproxy/7-check-primary-failover | ++ kubectl -n kuttl-test-live-kingfish get pods -l mysql.percona.com/primary=true '-ojsonpath={.items[0].metadata.name}'
    logger.go:42: 12:51:00 | haproxy/7-check-primary-failover | + primary_pod_from_label=haproxy-mysql-0
    logger.go:42: 12:51:00 | haproxy/7-check-primary-failover | + kubectl -n kuttl-test-live-kingfish delete pod haproxy-mysql-0
    logger.go:42: 12:51:01 | haproxy/7-check-primary-failover | pod "haproxy-mysql-0" deleted
    logger.go:42: 12:51:21 | haproxy/7-check-primary-failover | + sleep 3
    logger.go:42: 12:51:24 | haproxy/7-check-primary-failover | + data=()
    logger.go:42: 12:51:24 | haproxy/7-check-primary-failover | ++ seq 0 2
    logger.go:42: 12:51:24 | haproxy/7-check-primary-failover | + for i in '$(seq 0 2)'
    logger.go:42: 12:51:24 | haproxy/7-check-primary-failover | + data+=("$(get_primary_from_haproxy ${test_name}-haproxy-$i)")
    logger.go:42: 12:51:24 | haproxy/7-check-primary-failover | ++ get_primary_from_haproxy haproxy-haproxy-0
    logger.go:42: 12:51:24 | haproxy/7-check-primary-failover | ++ local haproxy_pod=haproxy-haproxy-0
    logger.go:42: 12:51:24 | haproxy/7-check-primary-failover | +++ kubectl -n kuttl-test-live-kingfish get pods haproxy-haproxy-0 -o 'jsonpath={.status.podIP}'
    logger.go:42: 12:51:25 | haproxy/7-check-primary-failover | ++ local haproxy_pod_ip=10.117.162.27
    logger.go:42: 12:51:25 | haproxy/7-check-primary-failover | ++ run_mysql 'SHOW VARIABLES LIKE '\''%hostname%'\'';' '-h 10.117.162.27 -P3306 -uroot -proot_password'
    logger.go:42: 12:51:25 | haproxy/7-check-primary-failover | ++ local 'command=SHOW VARIABLES LIKE '\''%hostname%'\'';'
    logger.go:42: 12:51:25 | haproxy/7-check-primary-failover | ++ awk '{print $2}'
    logger.go:42: 12:51:25 | haproxy/7-check-primary-failover | ++ local 'uri=-h 10.117.162.27 -P3306 -uroot -proot_password'
    logger.go:42: 12:51:25 | haproxy/7-check-primary-failover | ++ local pod=
    logger.go:42: 12:51:25 | haproxy/7-check-primary-failover | +++ get_client_pod
    logger.go:42: 12:51:25 | haproxy/7-check-primary-failover | +++ kubectl -n kuttl-test-live-kingfish get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 12:51:26 | haproxy/7-check-primary-failover | ++ client_pod=mysql-client
    logger.go:42: 12:51:26 | haproxy/7-check-primary-failover | ++ wait_pod mysql-client
    logger.go:42: 12:51:26 | haproxy/7-check-primary-failover | ++ local pod=mysql-client
    logger.go:42: 12:51:26 | haproxy/7-check-primary-failover | ++ set +o xtrace
    logger.go:42: 12:51:26 | haproxy/7-check-primary-failover | mysql-clienttrue
    logger.go:42: 12:51:26 | haproxy/7-check-primary-failover | ++ kubectl -n kuttl-test-live-kingfish exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SHOW VARIABLES LIKE '\''%hostname%'\'';" | mysql -sN -h 10.117.162.27 -P3306 -uroot -proot_password'
    logger.go:42: 12:51:26 | haproxy/7-check-primary-failover | ++ sed -e 's/mysql: //'
    logger.go:42: 12:51:26 | haproxy/7-check-primary-failover | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 12:51:28 | haproxy/7-check-primary-failover | + for i in '$(seq 0 2)'
    logger.go:42: 12:51:28 | haproxy/7-check-primary-failover | + data+=("$(get_primary_from_haproxy ${test_name}-haproxy-$i)")
    logger.go:42: 12:51:28 | haproxy/7-check-primary-failover | ++ get_primary_from_haproxy haproxy-haproxy-1
    logger.go:42: 12:51:28 | haproxy/7-check-primary-failover | ++ local haproxy_pod=haproxy-haproxy-1
    logger.go:42: 12:51:28 | haproxy/7-check-primary-failover | +++ kubectl -n kuttl-test-live-kingfish get pods haproxy-haproxy-1 -o 'jsonpath={.status.podIP}'
    logger.go:42: 12:51:29 | haproxy/7-check-primary-failover | ++ local haproxy_pod_ip=10.117.160.24
    logger.go:42: 12:51:29 | haproxy/7-check-primary-failover | ++ run_mysql 'SHOW VARIABLES LIKE '\''%hostname%'\'';' '-h 10.117.160.24 -P3306 -uroot -proot_password'
    logger.go:42: 12:51:29 | haproxy/7-check-primary-failover | ++ local 'command=SHOW VARIABLES LIKE '\''%hostname%'\'';'
    logger.go:42: 12:51:29 | haproxy/7-check-primary-failover | ++ awk '{print $2}'
    logger.go:42: 12:51:29 | haproxy/7-check-primary-failover | ++ local 'uri=-h 10.117.160.24 -P3306 -uroot -proot_password'
    logger.go:42: 12:51:29 | haproxy/7-check-primary-failover | ++ local pod=
    logger.go:42: 12:51:29 | haproxy/7-check-primary-failover | +++ get_client_pod
    logger.go:42: 12:51:29 | haproxy/7-check-primary-failover | +++ kubectl -n kuttl-test-live-kingfish get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 12:51:29 | haproxy/7-check-primary-failover | ++ client_pod=mysql-client
    logger.go:42: 12:51:29 | haproxy/7-check-primary-failover | ++ wait_pod mysql-client
    logger.go:42: 12:51:29 | haproxy/7-check-primary-failover | ++ local pod=mysql-client
    logger.go:42: 12:51:29 | haproxy/7-check-primary-failover | ++ set +o xtrace
    logger.go:42: 12:51:30 | haproxy/7-check-primary-failover | mysql-clienttrue
    logger.go:42: 12:51:30 | haproxy/7-check-primary-failover | ++ kubectl -n kuttl-test-live-kingfish exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SHOW VARIABLES LIKE '\''%hostname%'\'';" | mysql -sN -h 10.117.160.24 -P3306 -uroot -proot_password'
    logger.go:42: 12:51:30 | haproxy/7-check-primary-failover | ++ sed -e 's/mysql: //'
    logger.go:42: 12:51:30 | haproxy/7-check-primary-failover | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 12:51:32 | haproxy/7-check-primary-failover | + for i in '$(seq 0 2)'
    logger.go:42: 12:51:32 | haproxy/7-check-primary-failover | + data+=("$(get_primary_from_haproxy ${test_name}-haproxy-$i)")
    logger.go:42: 12:51:32 | haproxy/7-check-primary-failover | ++ get_primary_from_haproxy haproxy-haproxy-2
    logger.go:42: 12:51:32 | haproxy/7-check-primary-failover | ++ local haproxy_pod=haproxy-haproxy-2
    logger.go:42: 12:51:32 | haproxy/7-check-primary-failover | +++ kubectl -n kuttl-test-live-kingfish get pods haproxy-haproxy-2 -o 'jsonpath={.status.podIP}'
    logger.go:42: 12:51:33 | haproxy/7-check-primary-failover | ++ local haproxy_pod_ip=10.117.161.31
    logger.go:42: 12:51:33 | haproxy/7-check-primary-failover | ++ run_mysql 'SHOW VARIABLES LIKE '\''%hostname%'\'';' '-h 10.117.161.31 -P3306 -uroot -proot_password'
    logger.go:42: 12:51:33 | haproxy/7-check-primary-failover | ++ awk '{print $2}'
    logger.go:42: 12:51:33 | haproxy/7-check-primary-failover | ++ local 'command=SHOW VARIABLES LIKE '\''%hostname%'\'';'
    logger.go:42: 12:51:33 | haproxy/7-check-primary-failover | ++ local 'uri=-h 10.117.161.31 -P3306 -uroot -proot_password'
    logger.go:42: 12:51:33 | haproxy/7-check-primary-failover | ++ local pod=
    logger.go:42: 12:51:33 | haproxy/7-check-primary-failover | +++ get_client_pod
    logger.go:42: 12:51:33 | haproxy/7-check-primary-failover | +++ kubectl -n kuttl-test-live-kingfish get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 12:51:33 | haproxy/7-check-primary-failover | ++ client_pod=mysql-client
    logger.go:42: 12:51:33 | haproxy/7-check-primary-failover | ++ wait_pod mysql-client
    logger.go:42: 12:51:33 | haproxy/7-check-primary-failover | ++ local pod=mysql-client
    logger.go:42: 12:51:33 | haproxy/7-check-primary-failover | ++ set +o xtrace
    logger.go:42: 12:51:34 | haproxy/7-check-primary-failover | mysql-clienttrue
    logger.go:42: 12:51:34 | haproxy/7-check-primary-failover | ++ sed -e 's/mysql: //'
    logger.go:42: 12:51:34 | haproxy/7-check-primary-failover | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 12:51:34 | haproxy/7-check-primary-failover | ++ kubectl -n kuttl-test-live-kingfish exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SHOW VARIABLES LIKE '\''%hostname%'\'';" | mysql -sN -h 10.117.161.31 -P3306 -uroot -proot_password'
    logger.go:42: 12:51:35 | haproxy/7-check-primary-failover | + '[' haproxy-mysql-2 '!=' haproxy-mysql-2 -o haproxy-mysql-2 '!=' haproxy-mysql-2 ']'
    logger.go:42: 12:51:35 | haproxy/7-check-primary-failover | ++ get_primary_from_label
    logger.go:42: 12:51:35 | haproxy/7-check-primary-failover | ++ kubectl -n kuttl-test-live-kingfish get pods -l mysql.percona.com/primary=true '-ojsonpath={.items[0].metadata.name}'
    logger.go:42: 12:51:36 | haproxy/7-check-primary-failover | + primary_pod_from_label=haproxy-mysql-2
    logger.go:42: 12:51:36 | haproxy/7-check-primary-failover | ++ get_primary_from_haproxy haproxy-haproxy-0
    logger.go:42: 12:51:36 | haproxy/7-check-primary-failover | ++ local haproxy_pod=haproxy-haproxy-0
    logger.go:42: 12:51:36 | haproxy/7-check-primary-failover | +++ kubectl -n kuttl-test-live-kingfish get pods haproxy-haproxy-0 -o 'jsonpath={.status.podIP}'
    logger.go:42: 12:51:37 | haproxy/7-check-primary-failover | ++ local haproxy_pod_ip=10.117.162.27
    logger.go:42: 12:51:37 | haproxy/7-check-primary-failover | ++ run_mysql 'SHOW VARIABLES LIKE '\''%hostname%'\'';' '-h 10.117.162.27 -P3306 -uroot -proot_password'
    logger.go:42: 12:51:37 | haproxy/7-check-primary-failover | ++ awk '{print $2}'
    logger.go:42: 12:51:37 | haproxy/7-check-primary-failover | ++ local 'command=SHOW VARIABLES LIKE '\''%hostname%'\'';'
    logger.go:42: 12:51:37 | haproxy/7-check-primary-failover | ++ local 'uri=-h 10.117.162.27 -P3306 -uroot -proot_password'
    logger.go:42: 12:51:37 | haproxy/7-check-primary-failover | ++ local pod=
    logger.go:42: 12:51:37 | haproxy/7-check-primary-failover | +++ get_client_pod
    logger.go:42: 12:51:37 | haproxy/7-check-primary-failover | +++ kubectl -n kuttl-test-live-kingfish get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 12:51:37 | haproxy/7-check-primary-failover | ++ client_pod=mysql-client
    logger.go:42: 12:51:37 | haproxy/7-check-primary-failover | ++ wait_pod mysql-client
    logger.go:42: 12:51:37 | haproxy/7-check-primary-failover | ++ local pod=mysql-client
    logger.go:42: 12:51:37 | haproxy/7-check-primary-failover | ++ set +o xtrace
    logger.go:42: 12:51:38 | haproxy/7-check-primary-failover | mysql-clienttrue
    logger.go:42: 12:51:38 | haproxy/7-check-primary-failover | ++ kubectl -n kuttl-test-live-kingfish exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SHOW VARIABLES LIKE '\''%hostname%'\'';" | mysql -sN -h 10.117.162.27 -P3306 -uroot -proot_password'
    logger.go:42: 12:51:38 | haproxy/7-check-primary-failover | ++ sed -e 's/mysql: //'
    logger.go:42: 12:51:38 | haproxy/7-check-primary-failover | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 12:51:40 | haproxy/7-check-primary-failover | + primary_pod_from_haproxy=haproxy-mysql-2
    logger.go:42: 12:51:40 | haproxy/7-check-primary-failover | + '[' haproxy-mysql-2 '!=' haproxy-mysql-2 ']'
    logger.go:42: 12:51:41 | haproxy/7-check-primary-failover | test step completed 7-check-primary-failover
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | starting test step 8-check-password-leak
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        check_passwords_leak]
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | + source ../../functions
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ realpath ../../..
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | ++++ mktemp -d
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ export TEMP_DIR=/tmp/tmp.Eswj9CARVV
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ TEMP_DIR=/tmp/tmp.Eswj9CARVV
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ export GIT_BRANCH=PR-339
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ GIT_BRANCH=PR-339
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ export VERSION=PR-339-3a8483b
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ VERSION=PR-339-3a8483b
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ export PMM_SERVER_VERSION=9.9.9
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ PMM_SERVER_VERSION=9.9.9
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ export IMAGE_PMM_SERVER_TAG=dev-latest
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ IMAGE_PMM_SERVER_TAG=dev-latest
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | ++++ which gdate
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | ++++ which date
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ date=/usr/bin/date
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ command -v oc
    logger.go:42: 12:51:41 | haproxy/8-check-password-leak | +++ oc get projects
    logger.go:42: 12:51:46 | haproxy/8-check-password-leak | error: the server doesn't have a resource type "projects"
    logger.go:42: 12:51:46 | haproxy/8-check-password-leak | +++ grep '^minikube'
    logger.go:42: 12:51:46 | haproxy/8-check-password-leak | +++ kubectl get nodes
    logger.go:42: 12:51:47 | haproxy/8-check-password-leak | ++++ pwd
    logger.go:42: 12:51:47 | haproxy/8-check-password-leak | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/haproxy
    logger.go:42: 12:51:47 | haproxy/8-check-password-leak | ++ test_name=haproxy
    logger.go:42: 12:51:47 | haproxy/8-check-password-leak | + check_passwords_leak
    logger.go:42: 12:51:47 | haproxy/8-check-password-leak | ++ kubectl get secrets -o json
    logger.go:42: 12:51:47 | haproxy/8-check-password-leak | ++ jq -r '.items[].data | to_entries | .[] | select(.key | (endswith(".crt") or endswith(".key") or endswith(".pub") or endswith(".pem") or endswith(".p12")) | not) | .value'
    logger.go:42: 12:51:48 | haproxy/8-check-password-leak | + secrets=
    logger.go:42: 12:51:48 | haproxy/8-check-password-leak | + passwords=' '
    logger.go:42: 12:51:48 | haproxy/8-check-password-leak | ++ kubectl -n kuttl-test-live-kingfish get pods -o name
    logger.go:42: 12:51:48 | haproxy/8-check-password-leak | ++ awk -F / '{print $2}'
    logger.go:42: 12:51:49 | haproxy/8-check-password-leak | + pods='haproxy-haproxy-0
    logger.go:42: 12:51:49 | haproxy/8-check-password-leak | haproxy-haproxy-1
    logger.go:42: 12:51:49 | haproxy/8-check-password-leak | haproxy-haproxy-2
    logger.go:42: 12:51:49 | haproxy/8-check-password-leak | haproxy-mysql-0
    logger.go:42: 12:51:49 | haproxy/8-check-password-leak | haproxy-mysql-1
    logger.go:42: 12:51:49 | haproxy/8-check-password-leak | haproxy-mysql-2
    logger.go:42: 12:51:49 | haproxy/8-check-password-leak | haproxy-orc-0
    logger.go:42: 12:51:49 | haproxy/8-check-password-leak | haproxy-orc-1
    logger.go:42: 12:51:49 | haproxy/8-check-password-leak | haproxy-orc-2
    logger.go:42: 12:51:49 | haproxy/8-check-password-leak | mysql-client
    logger.go:42: 12:51:49 | haproxy/8-check-password-leak | percona-server-mysql-operator-6dcd78d857-rs599'
    logger.go:42: 12:51:49 | haproxy/8-check-password-leak | + collect_logs kuttl-test-live-kingfish
    logger.go:42: 12:51:49 | haproxy/8-check-password-leak | + NS=kuttl-test-live-kingfish
    logger.go:42: 12:51:49 | haproxy/8-check-password-leak | + for p in '$pods'
    logger.go:42: 12:51:49 | haproxy/8-check-password-leak | ++ kubectl -n kuttl-test-live-kingfish get pod haproxy-haproxy-0 -o 'jsonpath={.spec.containers[*].name}'
    logger.go:42: 12:51:50 | haproxy/8-check-password-leak | + containers='haproxy mysql-monit'
    logger.go:42: 12:51:50 | haproxy/8-check-password-leak | + for c in '$containers'
    logger.go:42: 12:51:50 | haproxy/8-check-password-leak | + [[ haproxy =~ pmm ]]
    logger.go:42: 12:51:50 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-live-kingfish logs haproxy-haproxy-0 -c haproxy
    logger.go:42: 12:51:51 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-haproxy-0-haproxy.txt
    logger.go:42: 12:51:51 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-haproxy-0-haproxy.txt
    logger.go:42: 12:51:51 | haproxy/8-check-password-leak | + for c in '$containers'
    logger.go:42: 12:51:51 | haproxy/8-check-password-leak | + [[ mysql-monit =~ pmm ]]
    logger.go:42: 12:51:51 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-live-kingfish logs haproxy-haproxy-0 -c mysql-monit
    logger.go:42: 12:51:52 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-haproxy-0-mysql-monit.txt
    logger.go:42: 12:51:52 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-haproxy-0-mysql-monit.txt
    logger.go:42: 12:51:52 | haproxy/8-check-password-leak | + echo
    logger.go:42: 12:51:52 | haproxy/8-check-password-leak | 
    logger.go:42: 12:51:52 | haproxy/8-check-password-leak | + for p in '$pods'
    logger.go:42: 12:51:52 | haproxy/8-check-password-leak | ++ kubectl -n kuttl-test-live-kingfish get pod haproxy-haproxy-1 -o 'jsonpath={.spec.containers[*].name}'
    logger.go:42: 12:51:53 | haproxy/8-check-password-leak | + containers='haproxy mysql-monit'
    logger.go:42: 12:51:53 | haproxy/8-check-password-leak | + for c in '$containers'
    logger.go:42: 12:51:53 | haproxy/8-check-password-leak | + [[ haproxy =~ pmm ]]
    logger.go:42: 12:51:53 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-live-kingfish logs haproxy-haproxy-1 -c haproxy
    logger.go:42: 12:51:54 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-haproxy-1-haproxy.txt
    logger.go:42: 12:51:54 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-haproxy-1-haproxy.txt
    logger.go:42: 12:51:54 | haproxy/8-check-password-leak | + for c in '$containers'
    logger.go:42: 12:51:54 | haproxy/8-check-password-leak | + [[ mysql-monit =~ pmm ]]
    logger.go:42: 12:51:54 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-live-kingfish logs haproxy-haproxy-1 -c mysql-monit
    logger.go:42: 12:51:55 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-haproxy-1-mysql-monit.txt
    logger.go:42: 12:51:55 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-haproxy-1-mysql-monit.txt
    logger.go:42: 12:51:55 | haproxy/8-check-password-leak | + echo
    logger.go:42: 12:51:55 | haproxy/8-check-password-leak | 
    logger.go:42: 12:51:55 | haproxy/8-check-password-leak | + for p in '$pods'
    logger.go:42: 12:51:55 | haproxy/8-check-password-leak | ++ kubectl -n kuttl-test-live-kingfish get pod haproxy-haproxy-2 -o 'jsonpath={.spec.containers[*].name}'
    logger.go:42: 12:51:56 | haproxy/8-check-password-leak | + containers='haproxy mysql-monit'
    logger.go:42: 12:51:56 | haproxy/8-check-password-leak | + for c in '$containers'
    logger.go:42: 12:51:56 | haproxy/8-check-password-leak | + [[ haproxy =~ pmm ]]
    logger.go:42: 12:51:56 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-live-kingfish logs haproxy-haproxy-2 -c haproxy
    logger.go:42: 12:51:57 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-haproxy-2-haproxy.txt
    logger.go:42: 12:51:57 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-haproxy-2-haproxy.txt
    logger.go:42: 12:51:57 | haproxy/8-check-password-leak | + for c in '$containers'
    logger.go:42: 12:51:57 | haproxy/8-check-password-leak | + [[ mysql-monit =~ pmm ]]
    logger.go:42: 12:51:57 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-live-kingfish logs haproxy-haproxy-2 -c mysql-monit
    logger.go:42: 12:51:58 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-haproxy-2-mysql-monit.txt
    logger.go:42: 12:51:58 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-haproxy-2-mysql-monit.txt
    logger.go:42: 12:51:58 | haproxy/8-check-password-leak | + echo
    logger.go:42: 12:51:58 | haproxy/8-check-password-leak | 
    logger.go:42: 12:51:58 | haproxy/8-check-password-leak | + for p in '$pods'
    logger.go:42: 12:51:58 | haproxy/8-check-password-leak | ++ kubectl -n kuttl-test-live-kingfish get pod haproxy-mysql-0 -o 'jsonpath={.spec.containers[*].name}'
    logger.go:42: 12:51:59 | haproxy/8-check-password-leak | + containers='mysql xtrabackup pt-heartbeat'
    logger.go:42: 12:51:59 | haproxy/8-check-password-leak | + for c in '$containers'
    logger.go:42: 12:51:59 | haproxy/8-check-password-leak | + [[ mysql =~ pmm ]]
    logger.go:42: 12:51:59 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-live-kingfish logs haproxy-mysql-0 -c mysql
    logger.go:42: 12:52:00 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-mysql-0-mysql.txt
    logger.go:42: 12:52:00 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-mysql-0-mysql.txt
    logger.go:42: 12:52:00 | haproxy/8-check-password-leak | + for c in '$containers'
    logger.go:42: 12:52:00 | haproxy/8-check-password-leak | + [[ xtrabackup =~ pmm ]]
    logger.go:42: 12:52:00 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-live-kingfish logs haproxy-mysql-0 -c xtrabackup
    logger.go:42: 12:52:01 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-mysql-0-xtrabackup.txt
    logger.go:42: 12:52:01 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-mysql-0-xtrabackup.txt
    logger.go:42: 12:52:01 | haproxy/8-check-password-leak | + for c in '$containers'
    logger.go:42: 12:52:01 | haproxy/8-check-password-leak | + [[ pt-heartbeat =~ pmm ]]
    logger.go:42: 12:52:01 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-live-kingfish logs haproxy-mysql-0 -c pt-heartbeat
    logger.go:42: 12:52:01 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-mysql-0-pt-heartbeat.txt
    logger.go:42: 12:52:01 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-mysql-0-pt-heartbeat.txt
    logger.go:42: 12:52:01 | haproxy/8-check-password-leak | + echo
    logger.go:42: 12:52:01 | haproxy/8-check-password-leak | 
    logger.go:42: 12:52:01 | haproxy/8-check-password-leak | + for p in '$pods'
    logger.go:42: 12:52:01 | haproxy/8-check-password-leak | ++ kubectl -n kuttl-test-live-kingfish get pod haproxy-mysql-1 -o 'jsonpath={.spec.containers[*].name}'
    logger.go:42: 12:52:02 | haproxy/8-check-password-leak | + containers='mysql xtrabackup pt-heartbeat'
    logger.go:42: 12:52:02 | haproxy/8-check-password-leak | + for c in '$containers'
    logger.go:42: 12:52:02 | haproxy/8-check-password-leak | + [[ mysql =~ pmm ]]
    logger.go:42: 12:52:02 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-live-kingfish logs haproxy-mysql-1 -c mysql
    logger.go:42: 12:52:03 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-mysql-1-mysql.txt
    logger.go:42: 12:52:03 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-mysql-1-mysql.txt
    logger.go:42: 12:52:03 | haproxy/8-check-password-leak | + for c in '$containers'
    logger.go:42: 12:52:03 | haproxy/8-check-password-leak | + [[ xtrabackup =~ pmm ]]
    logger.go:42: 12:52:03 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-live-kingfish logs haproxy-mysql-1 -c xtrabackup
    logger.go:42: 12:52:04 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-mysql-1-xtrabackup.txt
    logger.go:42: 12:52:04 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-mysql-1-xtrabackup.txt
    logger.go:42: 12:52:04 | haproxy/8-check-password-leak | + for c in '$containers'
    logger.go:42: 12:52:04 | haproxy/8-check-password-leak | + [[ pt-heartbeat =~ pmm ]]
    logger.go:42: 12:52:04 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-live-kingfish logs haproxy-mysql-1 -c pt-heartbeat
    logger.go:42: 12:52:05 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-mysql-1-pt-heartbeat.txt
    logger.go:42: 12:52:05 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-mysql-1-pt-heartbeat.txt
    logger.go:42: 12:52:05 | haproxy/8-check-password-leak | + echo
    logger.go:42: 12:52:05 | haproxy/8-check-password-leak | 
    logger.go:42: 12:52:05 | haproxy/8-check-password-leak | + for p in '$pods'
    logger.go:42: 12:52:05 | haproxy/8-check-password-leak | ++ kubectl -n kuttl-test-live-kingfish get pod haproxy-mysql-2 -o 'jsonpath={.spec.containers[*].name}'
    logger.go:42: 12:52:06 | haproxy/8-check-password-leak | + containers='mysql xtrabackup pt-heartbeat'
    logger.go:42: 12:52:06 | haproxy/8-check-password-leak | + for c in '$containers'
    logger.go:42: 12:52:06 | haproxy/8-check-password-leak | + [[ mysql =~ pmm ]]
    logger.go:42: 12:52:06 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-live-kingfish logs haproxy-mysql-2 -c mysql
    logger.go:42: 12:52:07 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-mysql-2-mysql.txt
    logger.go:42: 12:52:07 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-mysql-2-mysql.txt
    logger.go:42: 12:52:07 | haproxy/8-check-password-leak | + for c in '$containers'
    logger.go:42: 12:52:07 | haproxy/8-check-password-leak | + [[ xtrabackup =~ pmm ]]
    logger.go:42: 12:52:07 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-live-kingfish logs haproxy-mysql-2 -c xtrabackup
    logger.go:42: 12:52:08 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-mysql-2-xtrabackup.txt
    logger.go:42: 12:52:08 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-mysql-2-xtrabackup.txt
    logger.go:42: 12:52:08 | haproxy/8-check-password-leak | + for c in '$containers'
    logger.go:42: 12:52:08 | haproxy/8-check-password-leak | + [[ pt-heartbeat =~ pmm ]]
    logger.go:42: 12:52:08 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-live-kingfish logs haproxy-mysql-2 -c pt-heartbeat
    logger.go:42: 12:52:09 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-mysql-2-pt-heartbeat.txt
    logger.go:42: 12:52:09 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-mysql-2-pt-heartbeat.txt
    logger.go:42: 12:52:09 | haproxy/8-check-password-leak | + echo
    logger.go:42: 12:52:09 | haproxy/8-check-password-leak | 
    logger.go:42: 12:52:09 | haproxy/8-check-password-leak | + for p in '$pods'
    logger.go:42: 12:52:09 | haproxy/8-check-password-leak | ++ kubectl -n kuttl-test-live-kingfish get pod haproxy-orc-0 -o 'jsonpath={.spec.containers[*].name}'
    logger.go:42: 12:52:10 | haproxy/8-check-password-leak | + containers='orc mysql-monit'
    logger.go:42: 12:52:10 | haproxy/8-check-password-leak | + for c in '$containers'
    logger.go:42: 12:52:10 | haproxy/8-check-password-leak | + [[ orc =~ pmm ]]
    logger.go:42: 12:52:10 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-live-kingfish logs haproxy-orc-0 -c orc
    logger.go:42: 12:52:11 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-orc-0-orc.txt
    logger.go:42: 12:52:11 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-orc-0-orc.txt
    logger.go:42: 12:52:11 | haproxy/8-check-password-leak | + for c in '$containers'
    logger.go:42: 12:52:11 | haproxy/8-check-password-leak | + [[ mysql-monit =~ pmm ]]
    logger.go:42: 12:52:11 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-live-kingfish logs haproxy-orc-0 -c mysql-monit
    logger.go:42: 12:52:12 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-orc-0-mysql-monit.txt
    logger.go:42: 12:52:12 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-orc-0-mysql-monit.txt
    logger.go:42: 12:52:12 | haproxy/8-check-password-leak | + echo
    logger.go:42: 12:52:12 | haproxy/8-check-password-leak | 
    logger.go:42: 12:52:12 | haproxy/8-check-password-leak | + for p in '$pods'
    logger.go:42: 12:52:12 | haproxy/8-check-password-leak | ++ kubectl -n kuttl-test-live-kingfish get pod haproxy-orc-1 -o 'jsonpath={.spec.containers[*].name}'
    logger.go:42: 12:52:13 | haproxy/8-check-password-leak | + containers='orc mysql-monit'
    logger.go:42: 12:52:13 | haproxy/8-check-password-leak | + for c in '$containers'
    logger.go:42: 12:52:13 | haproxy/8-check-password-leak | + [[ orc =~ pmm ]]
    logger.go:42: 12:52:13 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-live-kingfish logs haproxy-orc-1 -c orc
    logger.go:42: 12:52:14 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-orc-1-orc.txt
    logger.go:42: 12:52:14 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-orc-1-orc.txt
    logger.go:42: 12:52:14 | haproxy/8-check-password-leak | + for c in '$containers'
    logger.go:42: 12:52:14 | haproxy/8-check-password-leak | + [[ mysql-monit =~ pmm ]]
    logger.go:42: 12:52:14 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-live-kingfish logs haproxy-orc-1 -c mysql-monit
    logger.go:42: 12:52:15 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-orc-1-mysql-monit.txt
    logger.go:42: 12:52:15 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-orc-1-mysql-monit.txt
    logger.go:42: 12:52:15 | haproxy/8-check-password-leak | + echo
    logger.go:42: 12:52:15 | haproxy/8-check-password-leak | 
    logger.go:42: 12:52:15 | haproxy/8-check-password-leak | + for p in '$pods'
    logger.go:42: 12:52:15 | haproxy/8-check-password-leak | ++ kubectl -n kuttl-test-live-kingfish get pod haproxy-orc-2 -o 'jsonpath={.spec.containers[*].name}'
    logger.go:42: 12:52:16 | haproxy/8-check-password-leak | + containers='orc mysql-monit'
    logger.go:42: 12:52:16 | haproxy/8-check-password-leak | + for c in '$containers'
    logger.go:42: 12:52:16 | haproxy/8-check-password-leak | + [[ orc =~ pmm ]]
    logger.go:42: 12:52:16 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-live-kingfish logs haproxy-orc-2 -c orc
    logger.go:42: 12:52:17 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-orc-2-orc.txt
    logger.go:42: 12:52:17 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-orc-2-orc.txt
    logger.go:42: 12:52:17 | haproxy/8-check-password-leak | + for c in '$containers'
    logger.go:42: 12:52:17 | haproxy/8-check-password-leak | + [[ mysql-monit =~ pmm ]]
    logger.go:42: 12:52:17 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-live-kingfish logs haproxy-orc-2 -c mysql-monit
    logger.go:42: 12:52:18 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-orc-2-mysql-monit.txt
    logger.go:42: 12:52:18 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-haproxy-orc-2-mysql-monit.txt
    logger.go:42: 12:52:18 | haproxy/8-check-password-leak | + echo
    logger.go:42: 12:52:18 | haproxy/8-check-password-leak | 
    logger.go:42: 12:52:18 | haproxy/8-check-password-leak | + for p in '$pods'
    logger.go:42: 12:52:18 | haproxy/8-check-password-leak | ++ kubectl -n kuttl-test-live-kingfish get pod mysql-client -o 'jsonpath={.spec.containers[*].name}'
    logger.go:42: 12:52:19 | haproxy/8-check-password-leak | + containers=mysql-client
    logger.go:42: 12:52:19 | haproxy/8-check-password-leak | + for c in '$containers'
    logger.go:42: 12:52:19 | haproxy/8-check-password-leak | + [[ mysql-client =~ pmm ]]
    logger.go:42: 12:52:19 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-live-kingfish logs mysql-client -c mysql-client
    logger.go:42: 12:52:20 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-mysql-client-mysql-client.txt
    logger.go:42: 12:52:20 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-mysql-client-mysql-client.txt
    logger.go:42: 12:52:20 | haproxy/8-check-password-leak | + echo
    logger.go:42: 12:52:20 | haproxy/8-check-password-leak | 
    logger.go:42: 12:52:20 | haproxy/8-check-password-leak | + for p in '$pods'
    logger.go:42: 12:52:20 | haproxy/8-check-password-leak | ++ kubectl -n kuttl-test-live-kingfish get pod percona-server-mysql-operator-6dcd78d857-rs599 -o 'jsonpath={.spec.containers[*].name}'
    logger.go:42: 12:52:20 | haproxy/8-check-password-leak | + containers=manager
    logger.go:42: 12:52:20 | haproxy/8-check-password-leak | + for c in '$containers'
    logger.go:42: 12:52:20 | haproxy/8-check-password-leak | + [[ manager =~ pmm ]]
    logger.go:42: 12:52:20 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-live-kingfish logs percona-server-mysql-operator-6dcd78d857-rs599 -c manager
    logger.go:42: 12:52:21 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-percona-server-mysql-operator-6dcd78d857-rs599-manager.txt
    logger.go:42: 12:52:21 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.Eswj9CARVV/logs_output-percona-server-mysql-operator-6dcd78d857-rs599-manager.txt
    logger.go:42: 12:52:21 | haproxy/8-check-password-leak | + echo
    logger.go:42: 12:52:21 | haproxy/8-check-password-leak | 
    logger.go:42: 12:52:21 | haproxy/8-check-password-leak | + '[' -n ps-operator ']'
    logger.go:42: 12:52:21 | haproxy/8-check-password-leak | ++ kubectl -n ps-operator get pods -o name
    logger.go:42: 12:52:21 | haproxy/8-check-password-leak | ++ awk -F / '{print $2}'
    logger.go:42: 12:52:22 | haproxy/8-check-password-leak | + pods=
    logger.go:42: 12:52:22 | haproxy/8-check-password-leak | + collect_logs ps-operator
    logger.go:42: 12:52:22 | haproxy/8-check-password-leak | + NS=ps-operator
    logger.go:42: 12:52:23 | haproxy/8-check-password-leak | test step completed 8-check-password-leak
    logger.go:42: 12:52:23 | haproxy/9-disable-haproxy | starting test step 9-disable-haproxy
    logger.go:42: 12:52:24 | haproxy/9-disable-haproxy | PerconaServerMySQL:kuttl-test-live-kingfish/haproxy updated
    logger.go:42: 12:52:30 | haproxy/9-disable-haproxy | test step completed 9-disable-haproxy
    logger.go:42: 12:52:30 | haproxy/10-drop-finalizer | starting test step 10-drop-finalizer
    logger.go:42: 12:52:31 | haproxy/10-drop-finalizer | PerconaServerMySQL:kuttl-test-live-kingfish/haproxy updated
    logger.go:42: 12:52:31 | haproxy/10-drop-finalizer | test step completed 10-drop-finalizer
    logger.go:42: 12:52:32 | haproxy | haproxy events from ns kuttl-test-live-kingfish:
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:04 +0000 UTC	Normal	Pod percona-server-mysql-operator-6dcd78d857-rs599		Scheduled	Successfully assigned kuttl-test-live-kingfish/percona-server-mysql-operator-6dcd78d857-rs599 to gke-jen-ps-339-3a8483b-1-default-pool-092d3615-1n52		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:04 +0000 UTC	Normal	ReplicaSet.apps percona-server-mysql-operator-6dcd78d857		SuccessfulCreate	Created pod: percona-server-mysql-operator-6dcd78d857-rs599		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:04 +0000 UTC	Normal	Deployment.apps percona-server-mysql-operator		ScalingReplicaSet	Scaled up replica set percona-server-mysql-operator-6dcd78d857 to 1		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:07 +0000 UTC	Normal	Pod percona-server-mysql-operator-6dcd78d857-rs599.spec.containers{manager}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:07 +0000 UTC	Normal	Pod percona-server-mysql-operator-6dcd78d857-rs599.spec.containers{manager}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 456.264981ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:07 +0000 UTC	Normal	Pod percona-server-mysql-operator-6dcd78d857-rs599.spec.containers{manager}		Created	Created container manager		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:07 +0000 UTC	Normal	Pod percona-server-mysql-operator-6dcd78d857-rs599.spec.containers{manager}		Started	Started container manager		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:08 +0000 UTC	Normal	Lease.coordination.k8s.io 08db2feb.percona.com		LeaderElection	percona-server-mysql-operator-6dcd78d857-rs599_fef0b976-637f-4122-ba0c-41c34f377ba8 became leader		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:08 +0000 UTC	Normal	Pod mysql-client		Scheduled	Successfully assigned kuttl-test-live-kingfish/mysql-client to gke-jen-ps-339-3a8483b-1-default-pool-092d3615-1n52		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:09 +0000 UTC	Normal	Pod mysql-client.spec.containers{mysql-client}		Pulled	Container image "percona/percona-server:8.0.25" already present on machine		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:09 +0000 UTC	Normal	Pod mysql-client.spec.containers{mysql-client}		Created	Created container mysql-client		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:09 +0000 UTC	Normal	Pod mysql-client.spec.containers{mysql-client}		Started	Started container mysql-client		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:27 +0000 UTC	Normal	PersistentVolumeClaim datadir-haproxy-mysql-0		WaitForFirstConsumer	waiting for first consumer to be created before binding		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:27 +0000 UTC	Normal	PersistentVolumeClaim datadir-haproxy-mysql-0		Provisioning	External provisioner is provisioning volume for claim "kuttl-test-live-kingfish/datadir-haproxy-mysql-0"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:27 +0000 UTC	Normal	PersistentVolumeClaim datadir-haproxy-mysql-0		ExternalProvisioning	waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:27 +0000 UTC	Normal	StatefulSet.apps haproxy-mysql		SuccessfulCreate	create Claim datadir-haproxy-mysql-0 Pod haproxy-mysql-0 in StatefulSet haproxy-mysql success		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:27 +0000 UTC	Normal	StatefulSet.apps haproxy-mysql		SuccessfulCreate	create Pod haproxy-mysql-0 in StatefulSet haproxy-mysql successful		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:27 +0000 UTC	Normal	Pod haproxy-orc-0		Scheduled	Successfully assigned kuttl-test-live-kingfish/haproxy-orc-0 to gke-jen-ps-339-3a8483b-1-default-pool-092d3615-k278		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:27 +0000 UTC	Normal	StatefulSet.apps haproxy-orc		SuccessfulCreate	create Pod haproxy-orc-0 in StatefulSet haproxy-orc successful		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:28 +0000 UTC	Normal	Pod haproxy-orc-0.spec.initContainers{orc-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:29 +0000 UTC	Normal	Pod haproxy-orc-0.spec.initContainers{orc-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 668.853085ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:29 +0000 UTC	Normal	Pod haproxy-orc-0.spec.initContainers{orc-init}		Created	Created container orc-init		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:29 +0000 UTC	Normal	Pod haproxy-orc-0.spec.initContainers{orc-init}		Started	Started container orc-init		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:31 +0000 UTC	Normal	PersistentVolumeClaim datadir-haproxy-mysql-0		ProvisioningSucceeded	Successfully provisioned volume pvc-8840b96e-fab6-494c-a9d3-abc143a2014e		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:31 +0000 UTC	Normal	Pod haproxy-orc-0.spec.containers{orc}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:32 +0000 UTC	Normal	Pod haproxy-mysql-0		Scheduled	Successfully assigned kuttl-test-live-kingfish/haproxy-mysql-0 to gke-jen-ps-339-3a8483b-1-default-pool-092d3615-1n52		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:33 +0000 UTC	Normal	Pod haproxy-orc-0.spec.containers{orc}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 2.401963738s		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:33 +0000 UTC	Normal	Pod haproxy-orc-0.spec.containers{orc}		Created	Created container orc		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:34 +0000 UTC	Normal	Pod haproxy-orc-0.spec.containers{orc}		Started	Started container orc		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:34 +0000 UTC	Normal	Pod haproxy-orc-0.spec.containers{mysql-monit}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:34 +0000 UTC	Normal	Pod haproxy-orc-0.spec.containers{mysql-monit}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 448.560737ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:34 +0000 UTC	Normal	Pod haproxy-orc-0.spec.containers{mysql-monit}		Created	Created container mysql-monit		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:34 +0000 UTC	Normal	Pod haproxy-orc-0.spec.containers{mysql-monit}		Started	Started container mysql-monit		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:36 +0000 UTC	Normal	Pod haproxy-mysql-0		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-8840b96e-fab6-494c-a9d3-abc143a2014e" 		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:39 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.initContainers{mysql-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:39 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.initContainers{mysql-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 416.509924ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:39 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.initContainers{mysql-init}		Created	Created container mysql-init		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:39 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.initContainers{mysql-init}		Started	Started container mysql-init		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:41 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{mysql}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-psmysql"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:41 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 398.346135ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:41 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{mysql}		Created	Created container mysql		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:42 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{mysql}		Started	Started container mysql		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:42 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:42 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 400.67515ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:42 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{xtrabackup}		Created	Created container xtrabackup		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:42 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{xtrabackup}		Started	Started container xtrabackup		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:42 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{pt-heartbeat}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-toolkit"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:54 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{pt-heartbeat}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 11.75915848s		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:54 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{pt-heartbeat}		Created	Created container pt-heartbeat		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:46:54 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{pt-heartbeat}		Started	Started container pt-heartbeat		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:08 +0000 UTC	Normal	Pod haproxy-orc-1		Scheduled	Successfully assigned kuttl-test-live-kingfish/haproxy-orc-1 to gke-jen-ps-339-3a8483b-1-default-pool-092d3615-1n52		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:08 +0000 UTC	Normal	Pod haproxy-orc-1.spec.initContainers{orc-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:08 +0000 UTC	Normal	StatefulSet.apps haproxy-orc		SuccessfulCreate	create Pod haproxy-orc-1 in StatefulSet haproxy-orc successful		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:09 +0000 UTC	Normal	Pod haproxy-orc-1.spec.initContainers{orc-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 359.056028ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:09 +0000 UTC	Normal	Pod haproxy-orc-1.spec.initContainers{orc-init}		Created	Created container orc-init		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:09 +0000 UTC	Normal	Pod haproxy-orc-1.spec.initContainers{orc-init}		Started	Started container orc-init		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:11 +0000 UTC	Normal	Pod haproxy-orc-1.spec.containers{orc}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:13 +0000 UTC	Normal	PersistentVolumeClaim datadir-haproxy-mysql-1		WaitForFirstConsumer	waiting for first consumer to be created before binding		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:13 +0000 UTC	Normal	PersistentVolumeClaim datadir-haproxy-mysql-1		ExternalProvisioning	waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:13 +0000 UTC	Normal	PersistentVolumeClaim datadir-haproxy-mysql-1		Provisioning	External provisioner is provisioning volume for claim "kuttl-test-live-kingfish/datadir-haproxy-mysql-1"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:13 +0000 UTC	Normal	StatefulSet.apps haproxy-mysql		SuccessfulCreate	create Claim datadir-haproxy-mysql-1 Pod haproxy-mysql-1 in StatefulSet haproxy-mysql success		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:13 +0000 UTC	Normal	StatefulSet.apps haproxy-mysql		SuccessfulCreate	create Pod haproxy-mysql-1 in StatefulSet haproxy-mysql successful		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:14 +0000 UTC	Normal	Pod haproxy-orc-1.spec.containers{orc}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 2.35808979s		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:14 +0000 UTC	Normal	Pod haproxy-orc-1.spec.containers{orc}		Created	Created container orc		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:14 +0000 UTC	Normal	Pod haproxy-orc-1.spec.containers{orc}		Started	Started container orc		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:14 +0000 UTC	Normal	Pod haproxy-orc-1.spec.containers{mysql-monit}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:14 +0000 UTC	Normal	Pod haproxy-orc-1.spec.containers{mysql-monit}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 418.581395ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:14 +0000 UTC	Normal	Pod haproxy-orc-1.spec.containers{mysql-monit}		Created	Created container mysql-monit		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:14 +0000 UTC	Normal	Pod haproxy-orc-1.spec.containers{mysql-monit}		Started	Started container mysql-monit		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:17 +0000 UTC	Normal	PersistentVolumeClaim datadir-haproxy-mysql-1		ProvisioningSucceeded	Successfully provisioned volume pvc-aba93251-0e9f-43e7-aa87-4462f3bc26c5		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:18 +0000 UTC	Normal	Pod haproxy-haproxy-0		Scheduled	Successfully assigned kuttl-test-live-kingfish/haproxy-haproxy-0 to gke-jen-ps-339-3a8483b-1-default-pool-092d3615-1n52		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:18 +0000 UTC	Normal	Pod haproxy-haproxy-0.spec.initContainers{haproxy-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:18 +0000 UTC	Normal	StatefulSet.apps haproxy-haproxy		SuccessfulCreate	create Pod haproxy-haproxy-0 in StatefulSet haproxy-haproxy successful		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:18 +0000 UTC	Normal	Pod haproxy-mysql-1		Scheduled	Successfully assigned kuttl-test-live-kingfish/haproxy-mysql-1 to gke-jen-ps-339-3a8483b-1-default-pool-092d3615-k278		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:19 +0000 UTC	Normal	Pod haproxy-haproxy-0.spec.initContainers{haproxy-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 411.927129ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:19 +0000 UTC	Normal	Pod haproxy-haproxy-0.spec.initContainers{haproxy-init}		Created	Created container haproxy-init		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:19 +0000 UTC	Normal	Pod haproxy-haproxy-0.spec.initContainers{haproxy-init}		Started	Started container haproxy-init		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:21 +0000 UTC	Normal	Pod haproxy-haproxy-0.spec.containers{haproxy}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-haproxy"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:23 +0000 UTC	Normal	Pod haproxy-mysql-1		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-aba93251-0e9f-43e7-aa87-4462f3bc26c5" 		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:24 +0000 UTC	Normal	Pod haproxy-mysql-1.spec.initContainers{mysql-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:25 +0000 UTC	Normal	Pod haproxy-haproxy-0.spec.containers{haproxy}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 4.07452173s		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:25 +0000 UTC	Normal	Pod haproxy-haproxy-0.spec.containers{haproxy}		Created	Created container haproxy		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:25 +0000 UTC	Normal	Pod haproxy-haproxy-0.spec.containers{haproxy}		Started	Started container haproxy		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:25 +0000 UTC	Normal	Pod haproxy-haproxy-0.spec.containers{mysql-monit}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-haproxy"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:25 +0000 UTC	Normal	Pod haproxy-mysql-1.spec.initContainers{mysql-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 449.018608ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:25 +0000 UTC	Normal	Pod haproxy-mysql-1.spec.initContainers{mysql-init}		Created	Created container mysql-init		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:25 +0000 UTC	Normal	Pod haproxy-mysql-1.spec.initContainers{mysql-init}		Started	Started container mysql-init		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:26 +0000 UTC	Normal	Pod haproxy-haproxy-0.spec.containers{mysql-monit}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 416.893886ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:26 +0000 UTC	Normal	Pod haproxy-haproxy-0.spec.containers{mysql-monit}		Created	Created container mysql-monit		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:26 +0000 UTC	Normal	Pod haproxy-haproxy-0.spec.containers{mysql-monit}		Started	Started container mysql-monit		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:26 +0000 UTC	Normal	Pod haproxy-haproxy-1		Scheduled	Successfully assigned kuttl-test-live-kingfish/haproxy-haproxy-1 to gke-jen-ps-339-3a8483b-1-default-pool-092d3615-2s1q		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:26 +0000 UTC	Normal	StatefulSet.apps haproxy-haproxy		SuccessfulCreate	create Pod haproxy-haproxy-1 in StatefulSet haproxy-haproxy successful		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:27 +0000 UTC	Normal	Pod haproxy-mysql-1.spec.containers{mysql}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-psmysql"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:28 +0000 UTC	Normal	Pod haproxy-haproxy-1.spec.initContainers{haproxy-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:28 +0000 UTC	Normal	Pod haproxy-haproxy-1.spec.initContainers{haproxy-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 428.672133ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:28 +0000 UTC	Normal	Pod haproxy-haproxy-1.spec.initContainers{haproxy-init}		Created	Created container haproxy-init		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:28 +0000 UTC	Normal	Pod haproxy-mysql-1.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 367.866106ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:28 +0000 UTC	Normal	Pod haproxy-mysql-1.spec.containers{mysql}		Created	Created container mysql		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:28 +0000 UTC	Normal	Pod haproxy-mysql-1.spec.containers{mysql}		Started	Started container mysql		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:28 +0000 UTC	Normal	Pod haproxy-mysql-1.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:28 +0000 UTC	Normal	Pod haproxy-mysql-1.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 398.703942ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:28 +0000 UTC	Normal	Pod haproxy-mysql-1.spec.containers{xtrabackup}		Created	Created container xtrabackup		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:28 +0000 UTC	Normal	Pod haproxy-mysql-1.spec.containers{xtrabackup}		Started	Started container xtrabackup		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:28 +0000 UTC	Normal	Pod haproxy-mysql-1.spec.containers{pt-heartbeat}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-toolkit"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:29 +0000 UTC	Normal	Pod haproxy-haproxy-1.spec.initContainers{haproxy-init}		Started	Started container haproxy-init		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:31 +0000 UTC	Normal	Pod haproxy-haproxy-1.spec.containers{haproxy}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-haproxy"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:35 +0000 UTC	Normal	Pod haproxy-haproxy-1.spec.containers{haproxy}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 3.736340342s		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:35 +0000 UTC	Normal	Pod haproxy-haproxy-1.spec.containers{haproxy}		Created	Created container haproxy		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:35 +0000 UTC	Normal	Pod haproxy-haproxy-1.spec.containers{haproxy}		Started	Started container haproxy		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:35 +0000 UTC	Normal	Pod haproxy-haproxy-1.spec.containers{mysql-monit}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-haproxy"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:35 +0000 UTC	Normal	Pod haproxy-haproxy-1.spec.containers{mysql-monit}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 379.798676ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:35 +0000 UTC	Normal	Pod haproxy-haproxy-1.spec.containers{mysql-monit}		Created	Created container mysql-monit		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:35 +0000 UTC	Normal	Pod haproxy-haproxy-1.spec.containers{mysql-monit}		Started	Started container mysql-monit		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:36 +0000 UTC	Normal	Pod haproxy-haproxy-2		Scheduled	Successfully assigned kuttl-test-live-kingfish/haproxy-haproxy-2 to gke-jen-ps-339-3a8483b-1-default-pool-092d3615-k278		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:36 +0000 UTC	Normal	StatefulSet.apps haproxy-haproxy		SuccessfulCreate	create Pod haproxy-haproxy-2 in StatefulSet haproxy-haproxy successful		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:37 +0000 UTC	Normal	Pod haproxy-haproxy-2.spec.initContainers{haproxy-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:40 +0000 UTC	Normal	Pod haproxy-haproxy-2.spec.initContainers{haproxy-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 3.093845866s		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:40 +0000 UTC	Normal	Pod haproxy-haproxy-2.spec.initContainers{haproxy-init}		Created	Created container haproxy-init		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:40 +0000 UTC	Normal	Pod haproxy-haproxy-2.spec.initContainers{haproxy-init}		Started	Started container haproxy-init		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:40 +0000 UTC	Normal	Pod haproxy-mysql-1.spec.containers{pt-heartbeat}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 11.549758242s		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:40 +0000 UTC	Normal	Pod haproxy-mysql-1.spec.containers{pt-heartbeat}		Created	Created container pt-heartbeat		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:40 +0000 UTC	Normal	Pod haproxy-mysql-1.spec.containers{pt-heartbeat}		Started	Started container pt-heartbeat		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:46 +0000 UTC	Normal	Pod haproxy-haproxy-2.spec.containers{haproxy}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-haproxy"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:48 +0000 UTC	Normal	Pod haproxy-orc-2		Scheduled	Successfully assigned kuttl-test-live-kingfish/haproxy-orc-2 to gke-jen-ps-339-3a8483b-1-default-pool-092d3615-2s1q		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:48 +0000 UTC	Normal	StatefulSet.apps haproxy-orc		SuccessfulCreate	create Pod haproxy-orc-2 in StatefulSet haproxy-orc successful		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:49 +0000 UTC	Normal	Pod haproxy-orc-2.spec.initContainers{orc-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:49 +0000 UTC	Normal	Pod haproxy-orc-2.spec.initContainers{orc-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 389.587804ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:49 +0000 UTC	Normal	Pod haproxy-orc-2.spec.initContainers{orc-init}		Created	Created container orc-init		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:49 +0000 UTC	Normal	Pod haproxy-orc-2.spec.initContainers{orc-init}		Started	Started container orc-init		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:50 +0000 UTC	Warning	Pod haproxy-mysql-1.spec.containers{mysql}		Unhealthy	Startup probe failed: 		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:50 +0000 UTC	Normal	Pod haproxy-mysql-1.spec.containers{mysql}		Killing	Container mysql failed startup probe, will be restarted		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:51 +0000 UTC	Normal	Pod haproxy-haproxy-2.spec.containers{haproxy}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 4.544174747s		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:51 +0000 UTC	Normal	Pod haproxy-haproxy-2.spec.containers{haproxy}		Created	Created container haproxy		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:51 +0000 UTC	Normal	Pod haproxy-haproxy-2.spec.containers{haproxy}		Started	Started container haproxy		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:51 +0000 UTC	Normal	Pod haproxy-haproxy-2.spec.containers{mysql-monit}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-haproxy"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:51 +0000 UTC	Normal	Pod haproxy-haproxy-2.spec.containers{mysql-monit}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 382.901097ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:51 +0000 UTC	Normal	Pod haproxy-haproxy-2.spec.containers{mysql-monit}		Created	Created container mysql-monit		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:51 +0000 UTC	Normal	Pod haproxy-haproxy-2.spec.containers{mysql-monit}		Started	Started container mysql-monit		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:52 +0000 UTC	Normal	Pod haproxy-orc-2.spec.containers{orc}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:54 +0000 UTC	Normal	Pod haproxy-orc-2.spec.containers{orc}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 2.193097168s		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:54 +0000 UTC	Normal	Pod haproxy-orc-2.spec.containers{orc}		Created	Created container orc		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:54 +0000 UTC	Normal	Pod haproxy-orc-2.spec.containers{orc}		Started	Started container orc		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:54 +0000 UTC	Normal	Pod haproxy-orc-2.spec.containers{mysql-monit}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:55 +0000 UTC	Normal	Pod haproxy-mysql-1.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 418.632675ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:55 +0000 UTC	Normal	Pod haproxy-orc-2.spec.containers{mysql-monit}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 344.637216ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:55 +0000 UTC	Normal	Pod haproxy-orc-2.spec.containers{mysql-monit}		Created	Created container mysql-monit		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:47:55 +0000 UTC	Normal	Pod haproxy-orc-2.spec.containers{mysql-monit}		Started	Started container mysql-monit		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:48:29 +0000 UTC	Normal	PersistentVolumeClaim datadir-haproxy-mysql-2		WaitForFirstConsumer	waiting for first consumer to be created before binding		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:48:29 +0000 UTC	Normal	PersistentVolumeClaim datadir-haproxy-mysql-2		ExternalProvisioning	waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:48:29 +0000 UTC	Normal	PersistentVolumeClaim datadir-haproxy-mysql-2		Provisioning	External provisioner is provisioning volume for claim "kuttl-test-live-kingfish/datadir-haproxy-mysql-2"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:48:29 +0000 UTC	Normal	StatefulSet.apps haproxy-mysql		SuccessfulCreate	create Claim datadir-haproxy-mysql-2 Pod haproxy-mysql-2 in StatefulSet haproxy-mysql success		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:48:29 +0000 UTC	Normal	StatefulSet.apps haproxy-mysql		SuccessfulCreate	create Pod haproxy-mysql-2 in StatefulSet haproxy-mysql successful		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:48:33 +0000 UTC	Normal	PersistentVolumeClaim datadir-haproxy-mysql-2		ProvisioningSucceeded	Successfully provisioned volume pvc-3541d0c9-dc5e-4fe6-b503-6c72c6988e2a		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:48:33 +0000 UTC	Normal	Pod haproxy-mysql-2		Scheduled	Successfully assigned kuttl-test-live-kingfish/haproxy-mysql-2 to gke-jen-ps-339-3a8483b-1-default-pool-092d3615-2s1q		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:48:38 +0000 UTC	Normal	Pod haproxy-mysql-2		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-3541d0c9-dc5e-4fe6-b503-6c72c6988e2a" 		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:48:39 +0000 UTC	Normal	Pod haproxy-mysql-2.spec.initContainers{mysql-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:48:40 +0000 UTC	Normal	Pod haproxy-mysql-2.spec.initContainers{mysql-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 441.33013ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:48:40 +0000 UTC	Normal	Pod haproxy-mysql-2.spec.initContainers{mysql-init}		Created	Created container mysql-init		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:48:40 +0000 UTC	Normal	Pod haproxy-mysql-2.spec.initContainers{mysql-init}		Started	Started container mysql-init		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:48:42 +0000 UTC	Normal	Pod haproxy-mysql-2.spec.containers{mysql}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-psmysql"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:48:43 +0000 UTC	Normal	Pod haproxy-mysql-2.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 454.168514ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:48:43 +0000 UTC	Normal	Pod haproxy-mysql-2.spec.containers{mysql}		Created	Created container mysql		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:48:43 +0000 UTC	Normal	Pod haproxy-mysql-2.spec.containers{mysql}		Started	Started container mysql		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:48:43 +0000 UTC	Normal	Pod haproxy-mysql-2.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:48:43 +0000 UTC	Normal	Pod haproxy-mysql-2.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 409.522037ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:48:43 +0000 UTC	Normal	Pod haproxy-mysql-2.spec.containers{xtrabackup}		Created	Created container xtrabackup		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:48:43 +0000 UTC	Normal	Pod haproxy-mysql-2.spec.containers{xtrabackup}		Started	Started container xtrabackup		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:48:43 +0000 UTC	Normal	Pod haproxy-mysql-2.spec.containers{pt-heartbeat}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-toolkit"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:48:55 +0000 UTC	Normal	Pod haproxy-mysql-2.spec.containers{pt-heartbeat}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 11.884464132s		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:48:55 +0000 UTC	Normal	Pod haproxy-mysql-2.spec.containers{pt-heartbeat}		Created	Created container pt-heartbeat		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:48:55 +0000 UTC	Normal	Pod haproxy-mysql-2.spec.containers{pt-heartbeat}		Started	Started container pt-heartbeat		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:49:00 +0000 UTC	Warning	Pod haproxy-mysql-2.spec.containers{mysql}		Unhealthy	Startup probe failed: 		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:49:00 +0000 UTC	Normal	Pod haproxy-mysql-2.spec.containers{mysql}		Killing	Container mysql failed startup probe, will be restarted		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:49:06 +0000 UTC	Normal	Pod haproxy-mysql-2.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 357.950278ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:01 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{mysql}		Killing	Stopping container mysql		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:01 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{pt-heartbeat}		Killing	Stopping container pt-heartbeat		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:01 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{xtrabackup}		Killing	Stopping container xtrabackup		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:03 +0000 UTC	Warning	Pod haproxy-mysql-0.spec.containers{mysql}		Unhealthy	Readiness probe failed: 2023/08/09 12:51:03 readiness check failed: connect to db: ping database: dial tcp 10.117.162.25:33062: connect: connection refused
        		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:08 +0000 UTC	Warning	Pod haproxy-mysql-0.spec.containers{mysql}		Unhealthy	Readiness probe failed: 2023/08/09 12:51:08 readiness check failed: connect to db: ping database: dial tcp 10.117.162.25:33062: connect: connection refused
        		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:13 +0000 UTC	Warning	Pod haproxy-mysql-0.spec.containers{mysql}		Unhealthy	Readiness probe errored: rpc error: code = Unknown desc = failed to exec in container: container is in CONTAINER_EXITED state		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:21 +0000 UTC	Normal	Pod haproxy-mysql-0		Scheduled	Successfully assigned kuttl-test-live-kingfish/haproxy-mysql-0 to gke-jen-ps-339-3a8483b-1-default-pool-092d3615-1n52		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:22 +0000 UTC	Warning	Pod haproxy-mysql-0		FailedMount	MountVolume.SetUp failed for volume "config" : failed to sync configmap cache: timed out waiting for the condition		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:26 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.initContainers{mysql-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:26 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.initContainers{mysql-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 432.169748ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:26 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.initContainers{mysql-init}		Created	Created container mysql-init		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:26 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.initContainers{mysql-init}		Started	Started container mysql-init		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:28 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{mysql}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-psmysql"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:28 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 394.623849ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:28 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{mysql}		Created	Created container mysql		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:28 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{mysql}		Started	Started container mysql		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:28 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:29 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 394.383137ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:29 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{xtrabackup}		Created	Created container xtrabackup		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:29 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{xtrabackup}		Started	Started container xtrabackup		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:29 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{pt-heartbeat}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-toolkit"		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:29 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{pt-heartbeat}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 413.833358ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:29 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{pt-heartbeat}		Created	Created container pt-heartbeat		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:30 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{pt-heartbeat}		Started	Started container pt-heartbeat		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:46 +0000 UTC	Warning	Pod haproxy-mysql-0.spec.containers{mysql}		Unhealthy	Startup probe failed: 		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:46 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{mysql}		Killing	Container mysql failed startup probe, will be restarted		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:51:50 +0000 UTC	Normal	Pod haproxy-mysql-0.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 378.462536ms		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:52:29 +0000 UTC	Normal	Pod haproxy-haproxy-0.spec.containers{haproxy}		Killing	Stopping container haproxy		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:52:29 +0000 UTC	Normal	Pod haproxy-haproxy-0.spec.containers{mysql-monit}		Killing	Stopping container mysql-monit		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:52:29 +0000 UTC	Normal	Pod haproxy-haproxy-1.spec.containers{haproxy}		Killing	Stopping container haproxy		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:52:29 +0000 UTC	Normal	Pod haproxy-haproxy-1.spec.containers{mysql-monit}		Killing	Stopping container mysql-monit		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:52:29 +0000 UTC	Normal	Pod haproxy-haproxy-2.spec.containers{haproxy}		Killing	Stopping container haproxy		
    logger.go:42: 12:52:32 | haproxy | 2023-08-09 12:52:29 +0000 UTC	Normal	Pod haproxy-haproxy-2.spec.containers{mysql-monit}		Killing	Stopping container mysql-monit		
    logger.go:42: 12:52:32 | haproxy | Deleting namespace: kuttl-test-live-kingfish
=== CONT  kuttl
    harness.go:405: run tests finished
    harness.go:513: cleaning up
    harness.go:570: removing temp folder: ""
--- PASS: kuttl (443.50s)
    --- PASS: kuttl/harness (0.00s)
        --- PASS: kuttl/harness/haproxy (440.10s)
PASS