=== RUN   kuttl
    harness.go:464: starting setup
    harness.go:255: running tests using configured kubeconfig.
    harness.go:278: Successful connection to cluster at: https://34.132.18.191
    harness.go:363: running tests
    harness.go:75: going to run test suite with timeout of 180 seconds for each step
    harness.go:375: testsuite: e2e-tests/tests has 34 tests
=== RUN   kuttl/harness
=== RUN   kuttl/harness/gr-haproxy
=== PAUSE kuttl/harness/gr-haproxy
=== CONT  kuttl/harness/gr-haproxy
    logger.go:42: 02:18:32 | gr-haproxy | Creating namespace: kuttl-test-bright-bengal
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | starting test step 0-deploy-operator
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        init_temp_dir # do this only in the first TestStep
        
        deploy_operator
        deploy_non_tls_cluster_secrets
        deploy_tls_cluster_secrets
        deploy_client]
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | + source ../../functions
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ realpath ../../..
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-873
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | ++++ pwd
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/tests/gr-haproxy
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | ++ test_name=gr-haproxy
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/vars.sh
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-873
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-873
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/deploy
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/deploy
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/conf
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/conf
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-haproxy
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ TEMP_DIR=/tmp/kuttl/ps/gr-haproxy
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ export GIT_BRANCH=PR-873
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ GIT_BRANCH=PR-873
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ export VERSION=PR-873-af2827af
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ VERSION=PR-873-af2827af
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-873-af2827af
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-873-af2827af
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ export PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ export CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | ++++ which gdate
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-873/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | ++++ which date
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ date=/usr/bin/date
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ oc get projects
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ :
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ kubectl get nodes
    logger.go:42: 02:18:32 | gr-haproxy/0-deploy-operator | +++ grep '^minikube'
    logger.go:42: 02:18:33 | gr-haproxy/0-deploy-operator | + init_temp_dir
    logger.go:42: 02:18:33 | gr-haproxy/0-deploy-operator | + rm -rf /tmp/kuttl/ps/gr-haproxy
    logger.go:42: 02:18:33 | gr-haproxy/0-deploy-operator | + mkdir -p /tmp/kuttl/ps/gr-haproxy
    logger.go:42: 02:18:33 | gr-haproxy/0-deploy-operator | + deploy_operator
    logger.go:42: 02:18:33 | gr-haproxy/0-deploy-operator | + destroy_operator
    logger.go:42: 02:18:33 | gr-haproxy/0-deploy-operator | + kubectl -n ps-operator delete deployment percona-server-mysql-operator --force --grace-period=0
    logger.go:42: 02:18:33 | gr-haproxy/0-deploy-operator | Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.
    logger.go:42: 02:18:33 | gr-haproxy/0-deploy-operator | Error from server (NotFound): deployments.apps "percona-server-mysql-operator" not found
    logger.go:42: 02:18:33 | gr-haproxy/0-deploy-operator | + true
    logger.go:42: 02:18:33 | gr-haproxy/0-deploy-operator | + [[ -n ps-operator ]]
    logger.go:42: 02:18:33 | gr-haproxy/0-deploy-operator | + kubectl delete namespace ps-operator --force --grace-period=0
    logger.go:42: 02:18:33 | gr-haproxy/0-deploy-operator | Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.
    logger.go:42: 02:18:34 | gr-haproxy/0-deploy-operator | Error from server (NotFound): namespaces "ps-operator" not found
    logger.go:42: 02:18:34 | gr-haproxy/0-deploy-operator | + true
    logger.go:42: 02:18:34 | gr-haproxy/0-deploy-operator | + [[ -n ps-operator ]]
    logger.go:42: 02:18:34 | gr-haproxy/0-deploy-operator | + create_namespace ps-operator
    logger.go:42: 02:18:34 | gr-haproxy/0-deploy-operator | + local namespace=ps-operator
    logger.go:42: 02:18:34 | gr-haproxy/0-deploy-operator | + [[ -n '' ]]
    logger.go:42: 02:18:34 | gr-haproxy/0-deploy-operator | + kubectl delete namespace ps-operator --ignore-not-found
    logger.go:42: 02:18:34 | gr-haproxy/0-deploy-operator | + kubectl wait --for=delete namespace ps-operator
    logger.go:42: 02:18:35 | gr-haproxy/0-deploy-operator | + kubectl create namespace ps-operator
    logger.go:42: 02:18:35 | gr-haproxy/0-deploy-operator | namespace/ps-operator created
    logger.go:42: 02:18:35 | gr-haproxy/0-deploy-operator | + kubectl -n ps-operator apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-ps-operator_PR-873/deploy/crd.yaml
    logger.go:42: 02:18:36 | gr-haproxy/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlbackups.ps.percona.com serverside-applied
    logger.go:42: 02:18:36 | gr-haproxy/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlrestores.ps.percona.com serverside-applied
    logger.go:42: 02:18:37 | gr-haproxy/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqls.ps.percona.com serverside-applied
    logger.go:42: 02:18:37 | gr-haproxy/0-deploy-operator | + '[' -n ps-operator ']'
    logger.go:42: 02:18:37 | gr-haproxy/0-deploy-operator | + kubectl -n ps-operator apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-873/deploy/cw-rbac.yaml
    logger.go:42: 02:18:38 | gr-haproxy/0-deploy-operator | serviceaccount/percona-server-mysql-operator created
    logger.go:42: 02:18:38 | gr-haproxy/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created
    logger.go:42: 02:18:39 | gr-haproxy/0-deploy-operator | clusterrole.rbac.authorization.k8s.io/percona-server-mysql-operator unchanged
    logger.go:42: 02:18:39 | gr-haproxy/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created
    logger.go:42: 02:18:39 | gr-haproxy/0-deploy-operator | clusterrolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator unchanged
    logger.go:42: 02:18:39 | gr-haproxy/0-deploy-operator | + yq eval '(select(documentIndex==1).spec.template.spec.containers[] | select(.name=="manager").env[] | select(.name=="LOG_LEVEL").value) = "DEBUG"'
    logger.go:42: 02:18:39 | gr-haproxy/0-deploy-operator | + yq eval '(select(documentIndex==1).spec.template.spec.containers[] | select(.name=="manager").env[] | select(.name=="DISABLE_TELEMETRY").value) = "true"'
    logger.go:42: 02:18:39 | gr-haproxy/0-deploy-operator | + kubectl -n ps-operator apply -f -
    logger.go:42: 02:18:39 | gr-haproxy/0-deploy-operator | ++ printf 'select(documentIndex==1).spec.template.spec.containers[0].image="%s"' perconalab/percona-server-mysql-operator:PR-873-af2827af
    logger.go:42: 02:18:39 | gr-haproxy/0-deploy-operator | + yq eval 'select(documentIndex==1).spec.template.spec.containers[0].image="perconalab/percona-server-mysql-operator:PR-873-af2827af"' /mnt/jenkins/workspace/cloud-ps-operator_PR-873/deploy/cw-operator.yaml
    logger.go:42: 02:18:40 | gr-haproxy/0-deploy-operator | configmap/percona-server-mysql-operator-config created
    logger.go:42: 02:18:40 | gr-haproxy/0-deploy-operator | deployment.apps/percona-server-mysql-operator created
    logger.go:42: 02:18:40 | gr-haproxy/0-deploy-operator | + deploy_non_tls_cluster_secrets
    logger.go:42: 02:18:40 | gr-haproxy/0-deploy-operator | + kubectl -n kuttl-test-bright-bengal apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/conf/secrets.yaml
    logger.go:42: 02:18:41 | gr-haproxy/0-deploy-operator | secret/test-secrets created
    logger.go:42: 02:18:41 | gr-haproxy/0-deploy-operator | + deploy_tls_cluster_secrets
    logger.go:42: 02:18:41 | gr-haproxy/0-deploy-operator | + kubectl -n kuttl-test-bright-bengal apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/conf/ssl-secret.yaml
    logger.go:42: 02:18:42 | gr-haproxy/0-deploy-operator | secret/test-ssl created
    logger.go:42: 02:18:42 | gr-haproxy/0-deploy-operator | + deploy_client
    logger.go:42: 02:18:42 | gr-haproxy/0-deploy-operator | + kubectl -n kuttl-test-bright-bengal apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/conf/client.yaml
    logger.go:42: 02:18:43 | gr-haproxy/0-deploy-operator | pod/mysql-client created
    logger.go:42: 02:18:43 | gr-haproxy/0-deploy-operator | running command: [sh -c kubectl assert exist-enhanced deployment percona-server-mysql-operator -n ${OPERATOR_NS:-$NAMESPACE} --field-selector status.readyReplicas=1]
    logger.go:42: 02:18:43 | gr-haproxy/0-deploy-operator | ASSERT deployment percona-server-mysql-operator matching field criteria 'status.readyReplicas=1' should exist.
    logger.go:42: 02:18:44 | gr-haproxy/0-deploy-operator | ASSERT FAIL Resource(s) not found.
    logger.go:42: 02:18:45 | gr-haproxy/0-deploy-operator | running command: [sh -c kubectl assert exist-enhanced deployment percona-server-mysql-operator -n ${OPERATOR_NS:-$NAMESPACE} --field-selector status.readyReplicas=1]
    logger.go:42: 02:18:45 | gr-haproxy/0-deploy-operator | ASSERT deployment percona-server-mysql-operator matching field criteria 'status.readyReplicas=1' should exist.
    logger.go:42: 02:18:45 | gr-haproxy/0-deploy-operator | ASSERT FAIL Resource(s) not found.
    logger.go:42: 02:18:47 | gr-haproxy/0-deploy-operator | running command: [sh -c kubectl assert exist-enhanced deployment percona-server-mysql-operator -n ${OPERATOR_NS:-$NAMESPACE} --field-selector status.readyReplicas=1]
    logger.go:42: 02:18:47 | gr-haproxy/0-deploy-operator | ASSERT deployment percona-server-mysql-operator matching field criteria 'status.readyReplicas=1' should exist.
    logger.go:42: 02:18:47 | gr-haproxy/0-deploy-operator | ASSERT FAIL Resource(s) not found.
    logger.go:42: 02:18:48 | gr-haproxy/0-deploy-operator | running command: [sh -c kubectl assert exist-enhanced deployment percona-server-mysql-operator -n ${OPERATOR_NS:-$NAMESPACE} --field-selector status.readyReplicas=1]
    logger.go:42: 02:18:48 | gr-haproxy/0-deploy-operator | ASSERT deployment percona-server-mysql-operator matching field criteria 'status.readyReplicas=1' should exist.
    logger.go:42: 02:18:49 | gr-haproxy/0-deploy-operator | ASSERT FAIL Resource(s) not found.
    logger.go:42: 02:18:50 | gr-haproxy/0-deploy-operator | running command: [sh -c kubectl assert exist-enhanced deployment percona-server-mysql-operator -n ${OPERATOR_NS:-$NAMESPACE} --field-selector status.readyReplicas=1]
    logger.go:42: 02:18:50 | gr-haproxy/0-deploy-operator | ASSERT deployment percona-server-mysql-operator matching field criteria 'status.readyReplicas=1' should exist.
    logger.go:42: 02:18:50 | gr-haproxy/0-deploy-operator | ASSERT FAIL Resource(s) not found.
    logger.go:42: 02:18:52 | gr-haproxy/0-deploy-operator | running command: [sh -c kubectl assert exist-enhanced deployment percona-server-mysql-operator -n ${OPERATOR_NS:-$NAMESPACE} --field-selector status.readyReplicas=1]
    logger.go:42: 02:18:52 | gr-haproxy/0-deploy-operator | ASSERT deployment percona-server-mysql-operator matching field criteria 'status.readyReplicas=1' should exist.
    logger.go:42: 02:18:52 | gr-haproxy/0-deploy-operator | INFO   Found 1 resource(s).
    logger.go:42: 02:18:52 | gr-haproxy/0-deploy-operator | NAME                            NAMESPACE     COL0
    logger.go:42: 02:18:52 | gr-haproxy/0-deploy-operator | percona-server-mysql-operator   ps-operator   1
    logger.go:42: 02:18:52 | gr-haproxy/0-deploy-operator | ASSERT PASS
    logger.go:42: 02:18:52 | gr-haproxy/0-deploy-operator | test step completed 0-deploy-operator
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | starting test step 1-create-cluster
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        get_cr \
            | yq eval '.spec.mysql.clusterType="group-replication"' - \
            | yq eval '.spec.proxy.router.enabled=false' - \
            | yq eval '.spec.proxy.haproxy.enabled=true' - \
            | kubectl -n "${NAMESPACE}" apply -f -]
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | + source ../../functions
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ realpath ../../..
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-873
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | ++++ pwd
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/tests/gr-haproxy
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | ++ test_name=gr-haproxy
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/vars.sh
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-873
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-873
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/deploy
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/deploy
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/conf
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/conf
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-haproxy
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ TEMP_DIR=/tmp/kuttl/ps/gr-haproxy
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ export GIT_BRANCH=PR-873
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ GIT_BRANCH=PR-873
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ export VERSION=PR-873-af2827af
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ VERSION=PR-873-af2827af
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-873-af2827af
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-873-af2827af
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ export PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ export CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | ++++ which gdate
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-873/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | ++++ which date
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ date=/usr/bin/date
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ oc get projects
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ :
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ kubectl get nodes
    logger.go:42: 02:18:52 | gr-haproxy/1-create-cluster | +++ grep '^minikube'
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | + get_cr
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | + local name_suffix=
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | + yq eval '.spec.mysql.clusterType="group-replication"' -
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | + yq eval .spec.proxy.haproxy.enabled=true -
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | + yq eval .spec.proxy.router.enabled=false -
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | + yq eval '.spec.secretsName="test-secrets"' -
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | + yq eval '.spec.upgradeOptions.apply="disabled"' -
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | + yq eval '.spec.sslSecretName="test-ssl"' -
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | + yq eval '.spec.mysql.clusterType="async"' -
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | + kubectl -n kuttl-test-bright-bengal apply -f -
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | + '[' -n '' ']'
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | + yq eval -
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | ++ printf '.metadata.name="%s"' gr-haproxy
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | ++ printf '.spec.pmm.image="%s"' perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | + yq eval '.spec.pmm.image="perconalab/pmm-client:3-dev-latest"' -
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | ++ printf '.spec.backup.image="%s"' perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | + yq eval '.spec.backup.image="perconalab/percona-server-mysql-operator:main-backup"' -
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | + yq eval .spec.orchestrator.enabled=true -
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | ++ printf '.spec.initImage="%s"' perconalab/percona-server-mysql-operator:PR-873-af2827af
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | ++ printf '.spec.mysql.image="%s"' perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | + yq eval '.spec.initImage="perconalab/percona-server-mysql-operator:PR-873-af2827af"' -
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | + yq eval '.metadata.name="gr-haproxy"' /mnt/jenkins/workspace/cloud-ps-operator_PR-873/deploy/cr.yaml
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | ++ printf '.spec.proxy.router.image="%s"' perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | + yq eval '.spec.proxy.router.image="perconalab/percona-server-mysql-operator:main-router"' -
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | ++ printf '.spec.orchestrator.image="%s"' perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | + yq eval '.spec.orchestrator.image="perconalab/percona-server-mysql-operator:main-orchestrator"' -
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | ++ printf '.spec.proxy.haproxy.image="%s"' perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | + yq eval '.spec.proxy.haproxy.image="perconalab/percona-server-mysql-operator:main-haproxy"' -
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | ++ printf '.spec.toolkit.image="%s"' perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | + yq eval '.spec.toolkit.image="perconalab/percona-server-mysql-operator:main-toolkit"' -
    logger.go:42: 02:18:53 | gr-haproxy/1-create-cluster | + yq eval '.spec.mysql.image="perconalab/percona-server-mysql-operator:main-psmysql"' -
    logger.go:42: 02:18:54 | gr-haproxy/1-create-cluster | perconaservermysql.ps.percona.com/gr-haproxy created
    logger.go:42: 02:22:26 | gr-haproxy/1-create-cluster | test step completed 1-create-cluster
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | starting test step 2-write-data
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        run_mysql \
            "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" \
            "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password"
        
        run_mysql \
            "INSERT myDB.myTable (id) VALUES (100500)" \
            "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password"]
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | + source ../../functions
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ realpath ../../..
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-873
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | ++++ pwd
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/tests/gr-haproxy
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | ++ test_name=gr-haproxy
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/vars.sh
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-873
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-873
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/deploy
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/deploy
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/conf
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/conf
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-haproxy
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-haproxy
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ export GIT_BRANCH=PR-873
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ GIT_BRANCH=PR-873
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ export VERSION=PR-873-af2827af
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ VERSION=PR-873-af2827af
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-873-af2827af
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-873-af2827af
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ export PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ export CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | ++++ which gdate
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-873/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | ++++ which date
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ date=/usr/bin/date
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ oc get projects
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ :
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ kubectl get nodes
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ grep '^minikube'
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ get_cluster_name
    logger.go:42: 02:22:26 | gr-haproxy/2-write-data | +++ kubectl -n kuttl-test-bright-bengal get ps -o 'jsonpath={.items[0].metadata.name}'
    logger.go:42: 02:22:27 | gr-haproxy/2-write-data | ++ get_haproxy_svc gr-haproxy
    logger.go:42: 02:22:27 | gr-haproxy/2-write-data | ++ local cluster=gr-haproxy
    logger.go:42: 02:22:27 | gr-haproxy/2-write-data | ++ echo gr-haproxy-haproxy
    logger.go:42: 02:22:27 | gr-haproxy/2-write-data | + run_mysql 'CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' '-h gr-haproxy-haproxy -uroot -proot_password'
    logger.go:42: 02:22:27 | gr-haproxy/2-write-data | + local 'command=CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)'
    logger.go:42: 02:22:27 | gr-haproxy/2-write-data | + local 'uri=-h gr-haproxy-haproxy -uroot -proot_password'
    logger.go:42: 02:22:27 | gr-haproxy/2-write-data | + local pod=
    logger.go:42: 02:22:27 | gr-haproxy/2-write-data | ++ get_client_pod
    logger.go:42: 02:22:27 | gr-haproxy/2-write-data | ++ kubectl -n kuttl-test-bright-bengal get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:22:27 | gr-haproxy/2-write-data | + client_pod=mysql-client
    logger.go:42: 02:22:27 | gr-haproxy/2-write-data | + wait_pod mysql-client
    logger.go:42: 02:22:27 | gr-haproxy/2-write-data | + local pod=mysql-client
    logger.go:42: 02:22:27 | gr-haproxy/2-write-data | + set +o xtrace
    logger.go:42: 02:22:28 | gr-haproxy/2-write-data | mysql-clienttrue
    logger.go:42: 02:22:28 | gr-haproxy/2-write-data | + kubectl -n kuttl-test-bright-bengal exec mysql-client -- bash -c 'printf '\''%s\n'\'' "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" | mysql -sN -h gr-haproxy-haproxy -uroot -proot_password'
    logger.go:42: 02:22:28 | gr-haproxy/2-write-data | + sed -e 's/mysql: //'
    logger.go:42: 02:22:28 | gr-haproxy/2-write-data | + grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:22:29 | gr-haproxy/2-write-data | + :
    logger.go:42: 02:22:29 | gr-haproxy/2-write-data | +++ get_cluster_name
    logger.go:42: 02:22:29 | gr-haproxy/2-write-data | +++ kubectl -n kuttl-test-bright-bengal get ps -o 'jsonpath={.items[0].metadata.name}'
    logger.go:42: 02:22:30 | gr-haproxy/2-write-data | ++ get_haproxy_svc gr-haproxy
    logger.go:42: 02:22:30 | gr-haproxy/2-write-data | ++ local cluster=gr-haproxy
    logger.go:42: 02:22:30 | gr-haproxy/2-write-data | ++ echo gr-haproxy-haproxy
    logger.go:42: 02:22:30 | gr-haproxy/2-write-data | + run_mysql 'INSERT myDB.myTable (id) VALUES (100500)' '-h gr-haproxy-haproxy -uroot -proot_password'
    logger.go:42: 02:22:30 | gr-haproxy/2-write-data | + local 'command=INSERT myDB.myTable (id) VALUES (100500)'
    logger.go:42: 02:22:30 | gr-haproxy/2-write-data | + local 'uri=-h gr-haproxy-haproxy -uroot -proot_password'
    logger.go:42: 02:22:30 | gr-haproxy/2-write-data | + local pod=
    logger.go:42: 02:22:30 | gr-haproxy/2-write-data | ++ get_client_pod
    logger.go:42: 02:22:30 | gr-haproxy/2-write-data | ++ kubectl -n kuttl-test-bright-bengal get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:22:30 | gr-haproxy/2-write-data | + client_pod=mysql-client
    logger.go:42: 02:22:30 | gr-haproxy/2-write-data | + wait_pod mysql-client
    logger.go:42: 02:22:30 | gr-haproxy/2-write-data | + local pod=mysql-client
    logger.go:42: 02:22:30 | gr-haproxy/2-write-data | + set +o xtrace
    logger.go:42: 02:22:30 | gr-haproxy/2-write-data | mysql-clienttrue
    logger.go:42: 02:22:30 | gr-haproxy/2-write-data | + kubectl -n kuttl-test-bright-bengal exec mysql-client -- bash -c 'printf '\''%s\n'\'' "INSERT myDB.myTable (id) VALUES (100500)" | mysql -sN -h gr-haproxy-haproxy -uroot -proot_password'
    logger.go:42: 02:22:30 | gr-haproxy/2-write-data | + sed -e 's/mysql: //'
    logger.go:42: 02:22:30 | gr-haproxy/2-write-data | + grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:22:32 | gr-haproxy/2-write-data | + :
[controller-runtime] log.SetLogger(...) was never called; logs will not be displayed.
Detected at:
	>  goroutine 20 [running]:
	>  runtime/debug.Stack()
	>  	/nix/store/wkbckbd30nlhq4dxzg64q6y4vm1xx4fk-go-1.22.1/share/go/src/runtime/debug/stack.go:24 +0x5e
	>  sigs.k8s.io/controller-runtime/pkg/log.eventuallyFulfillRoot()
	>  	/home/mowsiany/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.3/pkg/log/log.go:60 +0xcd
	>  sigs.k8s.io/controller-runtime/pkg/log.(*delegatingLogSink).WithName(0xc0002a9c00, {0x184a055, 0x14})
	>  	/home/mowsiany/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.3/pkg/log/deleg.go:147 +0x3e
	>  github.com/go-logr/logr.Logger.WithName({{0x1acb7d8, 0xc0002a9c00}, 0x0}, {0x184a055?, 0xc0004eff80?})
	>  	/home/mowsiany/go/pkg/mod/github.com/go-logr/logr@v1.2.4/logr.go:336 +0x36
	>  sigs.k8s.io/controller-runtime/pkg/client.newClient(0x131ead3?, {0x0, 0xc0005b8850, {0x1accd90, 0xc00041ad00}, 0x0, {0x0, 0x0}, 0x0})
	>  	/home/mowsiany/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.3/pkg/client/client.go:122 +0xf1
	>  sigs.k8s.io/controller-runtime/pkg/client.New(0xc0004cf208?, {0x0, 0xc0005b8850, {0x1accd90, 0xc00041ad00}, 0x0, {0x0, 0x0}, 0x0})
	>  	/home/mowsiany/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.3/pkg/client/client.go:103 +0x7d
	>  github.com/kudobuilder/kuttl/pkg/test/utils.NewRetryClient(0xc0004cf208, {0x0, 0xc0005b8850, {0x1accd90, 0xc00041ad00}, 0x0, {0x0, 0x0}, 0x0})
	>  	/home/mowsiany/go/src/github.com/kudobuilder/kuttl/pkg/test/utils/kubernetes.go:177 +0x127
	>  github.com/kudobuilder/kuttl/pkg/test.(*Harness).Client(0xc0005c8008, 0x76?)
	>  	/home/mowsiany/go/src/github.com/kudobuilder/kuttl/pkg/test/harness.go:323 +0x18e
	>  github.com/kudobuilder/kuttl/pkg/test.(*Step).Create(0xc0003b1930, 0xc0005ca4e0, {0xc00047bf08, 0x18})
	>  	/home/mowsiany/go/src/github.com/kudobuilder/kuttl/pkg/test/step.go:177 +0x63
	>  github.com/kudobuilder/kuttl/pkg/test.(*Step).Run(0xc0003b1930, 0xc0005ca4e0, {0xc00047bf08, 0x18})
	>  	/home/mowsiany/go/src/github.com/kudobuilder/kuttl/pkg/test/step.go:457 +0x24a
	>  github.com/kudobuilder/kuttl/pkg/test.(*Case).Run(0xc0005c3900, 0xc0005ca4e0, 0xc000371b90)
	>  	/home/mowsiany/go/src/github.com/kudobuilder/kuttl/pkg/test/case.go:373 +0xaeb
	>  github.com/kudobuilder/kuttl/pkg/test.(*Harness).RunTests.func1.1(0xc0005ca4e0)
	>  	/home/mowsiany/go/src/github.com/kudobuilder/kuttl/pkg/test/harness.go:401 +0x12e
	>  testing.tRunner(0xc0005ca4e0, 0xc0005e2af8)
	>  	/nix/store/wkbckbd30nlhq4dxzg64q6y4vm1xx4fk-go-1.22.1/share/go/src/testing/testing.go:1689 +0xfb
	>  created by testing.(*T).Run in goroutine 19
	>  	/nix/store/wkbckbd30nlhq4dxzg64q6y4vm1xx4fk-go-1.22.1/share/go/src/testing/testing.go:1742 +0x390
    logger.go:42: 02:22:32 | gr-haproxy/2-write-data | test step completed 2-write-data
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | starting test step 3-check-connections
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        args=""
        
        primary=$(run_mysql "SELECT count(*) FROM myDB.myTable" "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password")
        args="${args} --from-literal=primary=${primary}"
        
        replica=$(run_mysql "SELECT count(*) FROM myDB.myTable" "-h $(get_haproxy_svc $(get_cluster_name)) -P 3307 -uroot -proot_password")
        args="${args} --from-literal=replica=${replica}"
        
        # uncomment after https://jira.percona.com/browse/K8SPS-284
        # proxy_protocol=$(run_mysql "SELECT count(*) FROM myDB.myTable" "-h $(get_haproxy_svc $(get_cluster_name)) -P 3309 -uroot -proot_password")
        # args="${args} --from-literal=proxy_protocol=${proxy_protocol}"
        
        mysqlx=$(run_mysqlsh "SELECT count(*) FROM myDB.myTable" "-h $(get_haproxy_svc $(get_cluster_name)) -P 33060 -uroot -proot_password")
        args="${args} --from-literal=mysqlx=${mysqlx}"
        
        mysql_admin=$(run_mysql "SELECT count(*) FROM myDB.myTable" "-h $(get_haproxy_svc $(get_cluster_name)) -P 33062 -uroot -proot_password")
        args="${args} --from-literal=mysql_admin=${mysql_admin}"
        
        kubectl create configmap -n "${NAMESPACE}" 03-check-connections $args]
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | + source ../../functions
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ realpath ../../..
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-873
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | ++++ pwd
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/tests/gr-haproxy
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | ++ test_name=gr-haproxy
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/vars.sh
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-873
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-873
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/deploy
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/deploy
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/conf
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/conf
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-haproxy
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ TEMP_DIR=/tmp/kuttl/ps/gr-haproxy
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ export GIT_BRANCH=PR-873
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ GIT_BRANCH=PR-873
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ export VERSION=PR-873-af2827af
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ VERSION=PR-873-af2827af
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-873-af2827af
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-873-af2827af
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ export PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ export CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | ++++ which gdate
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-873/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | ++++ which date
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ date=/usr/bin/date
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ oc get projects
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ :
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ kubectl get nodes
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | +++ grep '^minikube'
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | + args=
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | ++++ get_cluster_name
    logger.go:42: 02:22:32 | gr-haproxy/3-check-connections | ++++ kubectl -n kuttl-test-bright-bengal get ps -o 'jsonpath={.items[0].metadata.name}'
    logger.go:42: 02:22:33 | gr-haproxy/3-check-connections | +++ get_haproxy_svc gr-haproxy
    logger.go:42: 02:22:33 | gr-haproxy/3-check-connections | +++ local cluster=gr-haproxy
    logger.go:42: 02:22:33 | gr-haproxy/3-check-connections | +++ echo gr-haproxy-haproxy
    logger.go:42: 02:22:33 | gr-haproxy/3-check-connections | ++ run_mysql 'SELECT count(*) FROM myDB.myTable' '-h gr-haproxy-haproxy -uroot -proot_password'
    logger.go:42: 02:22:33 | gr-haproxy/3-check-connections | ++ local 'command=SELECT count(*) FROM myDB.myTable'
    logger.go:42: 02:22:33 | gr-haproxy/3-check-connections | ++ local 'uri=-h gr-haproxy-haproxy -uroot -proot_password'
    logger.go:42: 02:22:33 | gr-haproxy/3-check-connections | ++ local pod=
    logger.go:42: 02:22:33 | gr-haproxy/3-check-connections | +++ get_client_pod
    logger.go:42: 02:22:33 | gr-haproxy/3-check-connections | +++ kubectl -n kuttl-test-bright-bengal get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:22:33 | gr-haproxy/3-check-connections | ++ client_pod=mysql-client
    logger.go:42: 02:22:33 | gr-haproxy/3-check-connections | ++ wait_pod mysql-client
    logger.go:42: 02:22:33 | gr-haproxy/3-check-connections | ++ local pod=mysql-client
    logger.go:42: 02:22:33 | gr-haproxy/3-check-connections | ++ set +o xtrace
    logger.go:42: 02:22:34 | gr-haproxy/3-check-connections | mysql-clienttrue
    logger.go:42: 02:22:34 | gr-haproxy/3-check-connections | ++ kubectl -n kuttl-test-bright-bengal exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT count(*) FROM myDB.myTable" | mysql -sN -h gr-haproxy-haproxy -uroot -proot_password'
    logger.go:42: 02:22:34 | gr-haproxy/3-check-connections | ++ sed -e 's/mysql: //'
    logger.go:42: 02:22:34 | gr-haproxy/3-check-connections | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:22:35 | gr-haproxy/3-check-connections | + primary=1
    logger.go:42: 02:22:35 | gr-haproxy/3-check-connections | + args=' --from-literal=primary=1'
    logger.go:42: 02:22:35 | gr-haproxy/3-check-connections | ++++ get_cluster_name
    logger.go:42: 02:22:35 | gr-haproxy/3-check-connections | ++++ kubectl -n kuttl-test-bright-bengal get ps -o 'jsonpath={.items[0].metadata.name}'
    logger.go:42: 02:22:35 | gr-haproxy/3-check-connections | +++ get_haproxy_svc gr-haproxy
    logger.go:42: 02:22:35 | gr-haproxy/3-check-connections | +++ local cluster=gr-haproxy
    logger.go:42: 02:22:35 | gr-haproxy/3-check-connections | +++ echo gr-haproxy-haproxy
    logger.go:42: 02:22:35 | gr-haproxy/3-check-connections | ++ run_mysql 'SELECT count(*) FROM myDB.myTable' '-h gr-haproxy-haproxy -P 3307 -uroot -proot_password'
    logger.go:42: 02:22:35 | gr-haproxy/3-check-connections | ++ local 'command=SELECT count(*) FROM myDB.myTable'
    logger.go:42: 02:22:35 | gr-haproxy/3-check-connections | ++ local 'uri=-h gr-haproxy-haproxy -P 3307 -uroot -proot_password'
    logger.go:42: 02:22:35 | gr-haproxy/3-check-connections | ++ local pod=
    logger.go:42: 02:22:35 | gr-haproxy/3-check-connections | +++ get_client_pod
    logger.go:42: 02:22:35 | gr-haproxy/3-check-connections | +++ kubectl -n kuttl-test-bright-bengal get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:22:36 | gr-haproxy/3-check-connections | ++ client_pod=mysql-client
    logger.go:42: 02:22:36 | gr-haproxy/3-check-connections | ++ wait_pod mysql-client
    logger.go:42: 02:22:36 | gr-haproxy/3-check-connections | ++ local pod=mysql-client
    logger.go:42: 02:22:36 | gr-haproxy/3-check-connections | ++ set +o xtrace
    logger.go:42: 02:22:36 | gr-haproxy/3-check-connections | mysql-clienttrue
    logger.go:42: 02:22:36 | gr-haproxy/3-check-connections | ++ kubectl -n kuttl-test-bright-bengal exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT count(*) FROM myDB.myTable" | mysql -sN -h gr-haproxy-haproxy -P 3307 -uroot -proot_password'
    logger.go:42: 02:22:36 | gr-haproxy/3-check-connections | ++ sed -e 's/mysql: //'
    logger.go:42: 02:22:36 | gr-haproxy/3-check-connections | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:22:38 | gr-haproxy/3-check-connections | + replica=1
    logger.go:42: 02:22:38 | gr-haproxy/3-check-connections | + args=' --from-literal=primary=1 --from-literal=replica=1'
    logger.go:42: 02:22:38 | gr-haproxy/3-check-connections | ++++ get_cluster_name
    logger.go:42: 02:22:38 | gr-haproxy/3-check-connections | ++++ kubectl -n kuttl-test-bright-bengal get ps -o 'jsonpath={.items[0].metadata.name}'
    logger.go:42: 02:22:38 | gr-haproxy/3-check-connections | +++ get_haproxy_svc gr-haproxy
    logger.go:42: 02:22:38 | gr-haproxy/3-check-connections | +++ local cluster=gr-haproxy
    logger.go:42: 02:22:38 | gr-haproxy/3-check-connections | +++ echo gr-haproxy-haproxy
    logger.go:42: 02:22:38 | gr-haproxy/3-check-connections | ++ run_mysqlsh 'SELECT count(*) FROM myDB.myTable' '-h gr-haproxy-haproxy -P 33060 -uroot -proot_password'
    logger.go:42: 02:22:38 | gr-haproxy/3-check-connections | ++ local 'command=SELECT count(*) FROM myDB.myTable'
    logger.go:42: 02:22:38 | gr-haproxy/3-check-connections | ++ local 'uri=-h gr-haproxy-haproxy -P 33060 -uroot -proot_password'
    logger.go:42: 02:22:38 | gr-haproxy/3-check-connections | ++ local pod=
    logger.go:42: 02:22:38 | gr-haproxy/3-check-connections | +++ get_client_pod
    logger.go:42: 02:22:38 | gr-haproxy/3-check-connections | +++ kubectl -n kuttl-test-bright-bengal get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:22:39 | gr-haproxy/3-check-connections | ++ client_pod=mysql-client
    logger.go:42: 02:22:39 | gr-haproxy/3-check-connections | ++ wait_pod mysql-client
    logger.go:42: 02:22:39 | gr-haproxy/3-check-connections | ++ local pod=mysql-client
    logger.go:42: 02:22:39 | gr-haproxy/3-check-connections | ++ set +o xtrace
    logger.go:42: 02:22:39 | gr-haproxy/3-check-connections | mysql-clienttrue
    logger.go:42: 02:22:39 | gr-haproxy/3-check-connections | ++ tail -n +2
    logger.go:42: 02:22:39 | gr-haproxy/3-check-connections | ++ kubectl -n kuttl-test-bright-bengal exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT count(*) FROM myDB.myTable" | mysqlsh --sql --quiet-start=2 -h gr-haproxy-haproxy -P 33060 -uroot -proot_password'
    logger.go:42: 02:22:41 | gr-haproxy/3-check-connections | + mysqlx=1
    logger.go:42: 02:22:41 | gr-haproxy/3-check-connections | + args=' --from-literal=primary=1 --from-literal=replica=1 --from-literal=mysqlx=1'
    logger.go:42: 02:22:41 | gr-haproxy/3-check-connections | ++++ get_cluster_name
    logger.go:42: 02:22:41 | gr-haproxy/3-check-connections | ++++ kubectl -n kuttl-test-bright-bengal get ps -o 'jsonpath={.items[0].metadata.name}'
    logger.go:42: 02:22:41 | gr-haproxy/3-check-connections | +++ get_haproxy_svc gr-haproxy
    logger.go:42: 02:22:41 | gr-haproxy/3-check-connections | +++ local cluster=gr-haproxy
    logger.go:42: 02:22:41 | gr-haproxy/3-check-connections | +++ echo gr-haproxy-haproxy
    logger.go:42: 02:22:41 | gr-haproxy/3-check-connections | ++ run_mysql 'SELECT count(*) FROM myDB.myTable' '-h gr-haproxy-haproxy -P 33062 -uroot -proot_password'
    logger.go:42: 02:22:41 | gr-haproxy/3-check-connections | ++ local 'command=SELECT count(*) FROM myDB.myTable'
    logger.go:42: 02:22:41 | gr-haproxy/3-check-connections | ++ local 'uri=-h gr-haproxy-haproxy -P 33062 -uroot -proot_password'
    logger.go:42: 02:22:41 | gr-haproxy/3-check-connections | ++ local pod=
    logger.go:42: 02:22:41 | gr-haproxy/3-check-connections | +++ get_client_pod
    logger.go:42: 02:22:41 | gr-haproxy/3-check-connections | +++ kubectl -n kuttl-test-bright-bengal get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 02:22:42 | gr-haproxy/3-check-connections | ++ client_pod=mysql-client
    logger.go:42: 02:22:42 | gr-haproxy/3-check-connections | ++ wait_pod mysql-client
    logger.go:42: 02:22:42 | gr-haproxy/3-check-connections | ++ local pod=mysql-client
    logger.go:42: 02:22:42 | gr-haproxy/3-check-connections | ++ set +o xtrace
    logger.go:42: 02:22:42 | gr-haproxy/3-check-connections | mysql-clienttrue
    logger.go:42: 02:22:42 | gr-haproxy/3-check-connections | ++ kubectl -n kuttl-test-bright-bengal exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT count(*) FROM myDB.myTable" | mysql -sN -h gr-haproxy-haproxy -P 33062 -uroot -proot_password'
    logger.go:42: 02:22:42 | gr-haproxy/3-check-connections | ++ sed -e 's/mysql: //'
    logger.go:42: 02:22:42 | gr-haproxy/3-check-connections | ++ grep -v 'Using a password on the command line interface can be insecure.'
    logger.go:42: 02:22:44 | gr-haproxy/3-check-connections | + mysql_admin=1
    logger.go:42: 02:22:44 | gr-haproxy/3-check-connections | + args=' --from-literal=primary=1 --from-literal=replica=1 --from-literal=mysqlx=1 --from-literal=mysql_admin=1'
    logger.go:42: 02:22:44 | gr-haproxy/3-check-connections | + kubectl create configmap -n kuttl-test-bright-bengal 03-check-connections --from-literal=primary=1 --from-literal=replica=1 --from-literal=mysqlx=1 --from-literal=mysql_admin=1
    logger.go:42: 02:22:44 | gr-haproxy/3-check-connections | configmap/03-check-connections created
    logger.go:42: 02:22:44 | gr-haproxy/3-check-connections | test step completed 3-check-connections
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | starting test step 4-check-password-leak
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        check_passwords_leak]
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | + source ../../functions
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ realpath ../../..
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-873
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | ++++ pwd
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/tests/gr-haproxy
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | ++ test_name=gr-haproxy
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/vars.sh
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-873
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-873
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/deploy
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/deploy
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/conf
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/conf
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-haproxy
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ TEMP_DIR=/tmp/kuttl/ps/gr-haproxy
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ export GIT_BRANCH=PR-873
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ GIT_BRANCH=PR-873
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ export VERSION=PR-873-af2827af
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ VERSION=PR-873-af2827af
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-873-af2827af
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-873-af2827af
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ export PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ export CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | ++++ which gdate
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-873/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | ++++ which date
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ date=/usr/bin/date
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ oc get projects
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ :
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ kubectl get nodes
    logger.go:42: 02:22:44 | gr-haproxy/4-check-password-leak | +++ grep '^minikube'
    logger.go:42: 02:22:45 | gr-haproxy/4-check-password-leak | + check_passwords_leak
    logger.go:42: 02:22:45 | gr-haproxy/4-check-password-leak | + local secrets
    logger.go:42: 02:22:45 | gr-haproxy/4-check-password-leak | + local passwords
    logger.go:42: 02:22:45 | gr-haproxy/4-check-password-leak | + local pods
    logger.go:42: 02:22:45 | gr-haproxy/4-check-password-leak | ++ kubectl get secrets -o json
    logger.go:42: 02:22:45 | gr-haproxy/4-check-password-leak | ++ jq -r '.items[].data | to_entries | .[] | select(.key | (endswith(".crt") or endswith(".key") or endswith(".pub") or endswith(".pem") or endswith(".p12") or test("namespace")) | not) | .value'
    logger.go:42: 02:22:45 | gr-haproxy/4-check-password-leak | + secrets=
    logger.go:42: 02:22:45 | gr-haproxy/4-check-password-leak | + passwords=' '
    logger.go:42: 02:22:45 | gr-haproxy/4-check-password-leak | ++ kubectl -n kuttl-test-bright-bengal get pods -o name
    logger.go:42: 02:22:45 | gr-haproxy/4-check-password-leak | ++ awk -F / '{print $2}'
    logger.go:42: 02:22:46 | gr-haproxy/4-check-password-leak | + pods='gr-haproxy-haproxy-0
    logger.go:42: 02:22:46 | gr-haproxy/4-check-password-leak | gr-haproxy-haproxy-1
    logger.go:42: 02:22:46 | gr-haproxy/4-check-password-leak | gr-haproxy-haproxy-2
    logger.go:42: 02:22:46 | gr-haproxy/4-check-password-leak | gr-haproxy-mysql-0
    logger.go:42: 02:22:46 | gr-haproxy/4-check-password-leak | gr-haproxy-mysql-1
    logger.go:42: 02:22:46 | gr-haproxy/4-check-password-leak | gr-haproxy-mysql-2
    logger.go:42: 02:22:46 | gr-haproxy/4-check-password-leak | mysql-client'
    logger.go:42: 02:22:46 | gr-haproxy/4-check-password-leak | + collect_logs kuttl-test-bright-bengal
    logger.go:42: 02:22:46 | gr-haproxy/4-check-password-leak | + local containers
    logger.go:42: 02:22:46 | gr-haproxy/4-check-password-leak | + local count
    logger.go:42: 02:22:46 | gr-haproxy/4-check-password-leak | + NS=kuttl-test-bright-bengal
    logger.go:42: 02:22:46 | gr-haproxy/4-check-password-leak | + for p in '$pods'
    logger.go:42: 02:22:46 | gr-haproxy/4-check-password-leak | ++ kubectl -n kuttl-test-bright-bengal get pod gr-haproxy-haproxy-0 -o 'jsonpath={.spec.containers[*].name}'
    logger.go:42: 02:22:46 | gr-haproxy/4-check-password-leak | + containers='haproxy mysql-monit'
    logger.go:42: 02:22:46 | gr-haproxy/4-check-password-leak | + for c in '$containers'
    logger.go:42: 02:22:46 | gr-haproxy/4-check-password-leak | + kubectl -n kuttl-test-bright-bengal logs gr-haproxy-haproxy-0 -c haproxy
    logger.go:42: 02:22:47 | gr-haproxy/4-check-password-leak | + echo logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-gr-haproxy-haproxy-0-haproxy.txt
    logger.go:42: 02:22:47 | gr-haproxy/4-check-password-leak | logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-gr-haproxy-haproxy-0-haproxy.txt
    logger.go:42: 02:22:47 | gr-haproxy/4-check-password-leak | + for c in '$containers'
    logger.go:42: 02:22:47 | gr-haproxy/4-check-password-leak | + kubectl -n kuttl-test-bright-bengal logs gr-haproxy-haproxy-0 -c mysql-monit
    logger.go:42: 02:22:47 | gr-haproxy/4-check-password-leak | + echo logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-gr-haproxy-haproxy-0-mysql-monit.txt
    logger.go:42: 02:22:47 | gr-haproxy/4-check-password-leak | logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-gr-haproxy-haproxy-0-mysql-monit.txt
    logger.go:42: 02:22:47 | gr-haproxy/4-check-password-leak | + echo
    logger.go:42: 02:22:47 | gr-haproxy/4-check-password-leak | 
    logger.go:42: 02:22:47 | gr-haproxy/4-check-password-leak | + for p in '$pods'
    logger.go:42: 02:22:47 | gr-haproxy/4-check-password-leak | ++ kubectl -n kuttl-test-bright-bengal get pod gr-haproxy-haproxy-1 -o 'jsonpath={.spec.containers[*].name}'
    logger.go:42: 02:22:48 | gr-haproxy/4-check-password-leak | + containers='haproxy mysql-monit'
    logger.go:42: 02:22:48 | gr-haproxy/4-check-password-leak | + for c in '$containers'
    logger.go:42: 02:22:48 | gr-haproxy/4-check-password-leak | + kubectl -n kuttl-test-bright-bengal logs gr-haproxy-haproxy-1 -c haproxy
    logger.go:42: 02:22:49 | gr-haproxy/4-check-password-leak | + echo logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-gr-haproxy-haproxy-1-haproxy.txt
    logger.go:42: 02:22:49 | gr-haproxy/4-check-password-leak | logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-gr-haproxy-haproxy-1-haproxy.txt
    logger.go:42: 02:22:49 | gr-haproxy/4-check-password-leak | + for c in '$containers'
    logger.go:42: 02:22:49 | gr-haproxy/4-check-password-leak | + kubectl -n kuttl-test-bright-bengal logs gr-haproxy-haproxy-1 -c mysql-monit
    logger.go:42: 02:22:49 | gr-haproxy/4-check-password-leak | + echo logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-gr-haproxy-haproxy-1-mysql-monit.txt
    logger.go:42: 02:22:49 | gr-haproxy/4-check-password-leak | logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-gr-haproxy-haproxy-1-mysql-monit.txt
    logger.go:42: 02:22:49 | gr-haproxy/4-check-password-leak | + echo
    logger.go:42: 02:22:49 | gr-haproxy/4-check-password-leak | 
    logger.go:42: 02:22:49 | gr-haproxy/4-check-password-leak | + for p in '$pods'
    logger.go:42: 02:22:49 | gr-haproxy/4-check-password-leak | ++ kubectl -n kuttl-test-bright-bengal get pod gr-haproxy-haproxy-2 -o 'jsonpath={.spec.containers[*].name}'
    logger.go:42: 02:22:50 | gr-haproxy/4-check-password-leak | + containers='haproxy mysql-monit'
    logger.go:42: 02:22:50 | gr-haproxy/4-check-password-leak | + for c in '$containers'
    logger.go:42: 02:22:50 | gr-haproxy/4-check-password-leak | + kubectl -n kuttl-test-bright-bengal logs gr-haproxy-haproxy-2 -c haproxy
    logger.go:42: 02:22:50 | gr-haproxy/4-check-password-leak | + echo logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-gr-haproxy-haproxy-2-haproxy.txt
    logger.go:42: 02:22:50 | gr-haproxy/4-check-password-leak | logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-gr-haproxy-haproxy-2-haproxy.txt
    logger.go:42: 02:22:50 | gr-haproxy/4-check-password-leak | + for c in '$containers'
    logger.go:42: 02:22:50 | gr-haproxy/4-check-password-leak | + kubectl -n kuttl-test-bright-bengal logs gr-haproxy-haproxy-2 -c mysql-monit
    logger.go:42: 02:22:51 | gr-haproxy/4-check-password-leak | + echo logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-gr-haproxy-haproxy-2-mysql-monit.txt
    logger.go:42: 02:22:51 | gr-haproxy/4-check-password-leak | logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-gr-haproxy-haproxy-2-mysql-monit.txt
    logger.go:42: 02:22:51 | gr-haproxy/4-check-password-leak | + echo
    logger.go:42: 02:22:51 | gr-haproxy/4-check-password-leak | 
    logger.go:42: 02:22:51 | gr-haproxy/4-check-password-leak | + for p in '$pods'
    logger.go:42: 02:22:51 | gr-haproxy/4-check-password-leak | ++ kubectl -n kuttl-test-bright-bengal get pod gr-haproxy-mysql-0 -o 'jsonpath={.spec.containers[*].name}'
    logger.go:42: 02:22:52 | gr-haproxy/4-check-password-leak | + containers='mysql xtrabackup'
    logger.go:42: 02:22:52 | gr-haproxy/4-check-password-leak | + for c in '$containers'
    logger.go:42: 02:22:52 | gr-haproxy/4-check-password-leak | + kubectl -n kuttl-test-bright-bengal logs gr-haproxy-mysql-0 -c mysql
    logger.go:42: 02:22:52 | gr-haproxy/4-check-password-leak | + echo logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-gr-haproxy-mysql-0-mysql.txt
    logger.go:42: 02:22:52 | gr-haproxy/4-check-password-leak | logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-gr-haproxy-mysql-0-mysql.txt
    logger.go:42: 02:22:52 | gr-haproxy/4-check-password-leak | + for c in '$containers'
    logger.go:42: 02:22:52 | gr-haproxy/4-check-password-leak | + kubectl -n kuttl-test-bright-bengal logs gr-haproxy-mysql-0 -c xtrabackup
    logger.go:42: 02:22:53 | gr-haproxy/4-check-password-leak | + echo logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-gr-haproxy-mysql-0-xtrabackup.txt
    logger.go:42: 02:22:53 | gr-haproxy/4-check-password-leak | logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-gr-haproxy-mysql-0-xtrabackup.txt
    logger.go:42: 02:22:53 | gr-haproxy/4-check-password-leak | + echo
    logger.go:42: 02:22:53 | gr-haproxy/4-check-password-leak | 
    logger.go:42: 02:22:53 | gr-haproxy/4-check-password-leak | + for p in '$pods'
    logger.go:42: 02:22:53 | gr-haproxy/4-check-password-leak | ++ kubectl -n kuttl-test-bright-bengal get pod gr-haproxy-mysql-1 -o 'jsonpath={.spec.containers[*].name}'
    logger.go:42: 02:22:53 | gr-haproxy/4-check-password-leak | + containers='mysql xtrabackup'
    logger.go:42: 02:22:53 | gr-haproxy/4-check-password-leak | + for c in '$containers'
    logger.go:42: 02:22:53 | gr-haproxy/4-check-password-leak | + kubectl -n kuttl-test-bright-bengal logs gr-haproxy-mysql-1 -c mysql
    logger.go:42: 02:22:54 | gr-haproxy/4-check-password-leak | + echo logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-gr-haproxy-mysql-1-mysql.txt
    logger.go:42: 02:22:54 | gr-haproxy/4-check-password-leak | logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-gr-haproxy-mysql-1-mysql.txt
    logger.go:42: 02:22:54 | gr-haproxy/4-check-password-leak | + for c in '$containers'
    logger.go:42: 02:22:54 | gr-haproxy/4-check-password-leak | + kubectl -n kuttl-test-bright-bengal logs gr-haproxy-mysql-1 -c xtrabackup
    logger.go:42: 02:22:55 | gr-haproxy/4-check-password-leak | + echo logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-gr-haproxy-mysql-1-xtrabackup.txt
    logger.go:42: 02:22:55 | gr-haproxy/4-check-password-leak | logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-gr-haproxy-mysql-1-xtrabackup.txt
    logger.go:42: 02:22:55 | gr-haproxy/4-check-password-leak | + echo
    logger.go:42: 02:22:55 | gr-haproxy/4-check-password-leak | 
    logger.go:42: 02:22:55 | gr-haproxy/4-check-password-leak | + for p in '$pods'
    logger.go:42: 02:22:55 | gr-haproxy/4-check-password-leak | ++ kubectl -n kuttl-test-bright-bengal get pod gr-haproxy-mysql-2 -o 'jsonpath={.spec.containers[*].name}'
    logger.go:42: 02:22:55 | gr-haproxy/4-check-password-leak | + containers='mysql xtrabackup'
    logger.go:42: 02:22:55 | gr-haproxy/4-check-password-leak | + for c in '$containers'
    logger.go:42: 02:22:55 | gr-haproxy/4-check-password-leak | + kubectl -n kuttl-test-bright-bengal logs gr-haproxy-mysql-2 -c mysql
    logger.go:42: 02:22:56 | gr-haproxy/4-check-password-leak | + echo logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-gr-haproxy-mysql-2-mysql.txt
    logger.go:42: 02:22:56 | gr-haproxy/4-check-password-leak | logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-gr-haproxy-mysql-2-mysql.txt
    logger.go:42: 02:22:56 | gr-haproxy/4-check-password-leak | + for c in '$containers'
    logger.go:42: 02:22:56 | gr-haproxy/4-check-password-leak | + kubectl -n kuttl-test-bright-bengal logs gr-haproxy-mysql-2 -c xtrabackup
    logger.go:42: 02:22:56 | gr-haproxy/4-check-password-leak | + echo logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-gr-haproxy-mysql-2-xtrabackup.txt
    logger.go:42: 02:22:56 | gr-haproxy/4-check-password-leak | logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-gr-haproxy-mysql-2-xtrabackup.txt
    logger.go:42: 02:22:56 | gr-haproxy/4-check-password-leak | + echo
    logger.go:42: 02:22:56 | gr-haproxy/4-check-password-leak | 
    logger.go:42: 02:22:56 | gr-haproxy/4-check-password-leak | + for p in '$pods'
    logger.go:42: 02:22:56 | gr-haproxy/4-check-password-leak | ++ kubectl -n kuttl-test-bright-bengal get pod mysql-client -o 'jsonpath={.spec.containers[*].name}'
    logger.go:42: 02:22:57 | gr-haproxy/4-check-password-leak | + containers=mysql-client
    logger.go:42: 02:22:57 | gr-haproxy/4-check-password-leak | + for c in '$containers'
    logger.go:42: 02:22:57 | gr-haproxy/4-check-password-leak | + kubectl -n kuttl-test-bright-bengal logs mysql-client -c mysql-client
    logger.go:42: 02:22:57 | gr-haproxy/4-check-password-leak | + echo logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-mysql-client-mysql-client.txt
    logger.go:42: 02:22:57 | gr-haproxy/4-check-password-leak | logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-mysql-client-mysql-client.txt
    logger.go:42: 02:22:57 | gr-haproxy/4-check-password-leak | + echo
    logger.go:42: 02:22:57 | gr-haproxy/4-check-password-leak | 
    logger.go:42: 02:22:57 | gr-haproxy/4-check-password-leak | + '[' -n ps-operator ']'
    logger.go:42: 02:22:57 | gr-haproxy/4-check-password-leak | ++ kubectl -n ps-operator get pods -o name
    logger.go:42: 02:22:57 | gr-haproxy/4-check-password-leak | ++ awk -F / '{print $2}'
    logger.go:42: 02:22:58 | gr-haproxy/4-check-password-leak | + pods=percona-server-mysql-operator-5bfb78b44c-crztb
    logger.go:42: 02:22:58 | gr-haproxy/4-check-password-leak | + collect_logs ps-operator
    logger.go:42: 02:22:58 | gr-haproxy/4-check-password-leak | + local containers
    logger.go:42: 02:22:58 | gr-haproxy/4-check-password-leak | + local count
    logger.go:42: 02:22:58 | gr-haproxy/4-check-password-leak | + NS=ps-operator
    logger.go:42: 02:22:58 | gr-haproxy/4-check-password-leak | + for p in '$pods'
    logger.go:42: 02:22:58 | gr-haproxy/4-check-password-leak | ++ kubectl -n ps-operator get pod percona-server-mysql-operator-5bfb78b44c-crztb -o 'jsonpath={.spec.containers[*].name}'
    logger.go:42: 02:22:58 | gr-haproxy/4-check-password-leak | + containers=manager
    logger.go:42: 02:22:58 | gr-haproxy/4-check-password-leak | + for c in '$containers'
    logger.go:42: 02:22:58 | gr-haproxy/4-check-password-leak | + kubectl -n ps-operator logs percona-server-mysql-operator-5bfb78b44c-crztb -c manager
    logger.go:42: 02:22:59 | gr-haproxy/4-check-password-leak | + echo logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-percona-server-mysql-operator-5bfb78b44c-crztb-manager.txt
    logger.go:42: 02:22:59 | gr-haproxy/4-check-password-leak | logs saved in: /tmp/kuttl/ps/gr-haproxy/logs_output-percona-server-mysql-operator-5bfb78b44c-crztb-manager.txt
    logger.go:42: 02:22:59 | gr-haproxy/4-check-password-leak | + echo
    logger.go:42: 02:22:59 | gr-haproxy/4-check-password-leak | 
    logger.go:42: 02:22:59 | gr-haproxy/4-check-password-leak | test step completed 4-check-password-leak
    logger.go:42: 02:22:59 | gr-haproxy/98-drop-finalizer | starting test step 98-drop-finalizer
    logger.go:42: 02:22:59 | gr-haproxy/98-drop-finalizer | PerconaServerMySQL:kuttl-test-bright-bengal/gr-haproxy updated
    logger.go:42: 02:22:59 | gr-haproxy/98-drop-finalizer | test step completed 98-drop-finalizer
    logger.go:42: 02:22:59 | gr-haproxy/99-remove-cluster-gracefully | starting test step 99-remove-cluster-gracefully
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        destroy_operator]
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | + source ../../functions
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ realpath ../../..
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-873
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | ++++ pwd
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/tests/gr-haproxy
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | ++ test_name=gr-haproxy
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/vars.sh
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-873
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-873
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/deploy
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/deploy
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/conf
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-873/e2e-tests/conf
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-haproxy
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ TEMP_DIR=/tmp/kuttl/ps/gr-haproxy
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ export GIT_BRANCH=PR-873
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ GIT_BRANCH=PR-873
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ export VERSION=PR-873-af2827af
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ VERSION=PR-873-af2827af
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-873-af2827af
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-873-af2827af
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ export PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ PMM_SERVER_VERSION=1.4.0
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ export CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ CERT_MANAGER_VER=1.16.3
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | ++++ which gdate
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-873/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | ++++ which date
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ date=/usr/bin/date
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ oc get projects
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ :
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ kubectl get nodes
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | +++ grep '^minikube'
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | + destroy_operator
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | + kubectl -n ps-operator delete deployment percona-server-mysql-operator --force --grace-period=0
    logger.go:42: 02:23:00 | gr-haproxy/99-remove-cluster-gracefully | Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.
    logger.go:42: 02:23:01 | gr-haproxy/99-remove-cluster-gracefully | deployment.apps "percona-server-mysql-operator" force deleted
    logger.go:42: 02:23:01 | gr-haproxy/99-remove-cluster-gracefully | + [[ -n ps-operator ]]
    logger.go:42: 02:23:01 | gr-haproxy/99-remove-cluster-gracefully | + kubectl delete namespace ps-operator --force --grace-period=0
    logger.go:42: 02:23:01 | gr-haproxy/99-remove-cluster-gracefully | Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.
    logger.go:42: 02:23:01 | gr-haproxy/99-remove-cluster-gracefully | namespace "ps-operator" force deleted
    logger.go:42: 02:23:07 | gr-haproxy/99-remove-cluster-gracefully | test step completed 99-remove-cluster-gracefully
    logger.go:42: 02:23:07 | gr-haproxy | gr-haproxy events from ns kuttl-test-bright-bengal:
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:18:43 +0000 UTC	Normal	Pod mysql-client	Binding	Scheduled	Successfully assigned kuttl-test-bright-bengal/mysql-client to gke-jen-ps-873-af2827af--default-pool-f58c2b0d-wb8m	default-scheduler	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:18:43 +0000 UTC	Normal	Pod mysql-client.spec.containers{mysql-client}		Pulled	Container image "percona/percona-server:8.0.33" already present on machine	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:18:43 +0000 UTC	Normal	Pod mysql-client.spec.containers{mysql-client}		Created	Created container: mysql-client	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:18:43 +0000 UTC	Normal	Pod mysql-client.spec.containers{mysql-client}		Started	Started container mysql-client	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:18:55 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-haproxy-mysql-0		WaitForFirstConsumer	waiting for first consumer to be created before binding	persistentvolume-controller	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:18:55 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-haproxy-mysql-0		ExternalProvisioning	Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.	persistentvolume-controller	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:18:55 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-haproxy-mysql-0		Provisioning	External provisioner is provisioning volume for claim "kuttl-test-bright-bengal/datadir-gr-haproxy-mysql-0"	pd.csi.storage.gke.io_gke-b35679d75e564d2d98a0-bb78-fb92-vm_7b300296-10a9-4121-bf46-b146398f958e	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:18:55 +0000 UTC	Normal	StatefulSet.apps gr-haproxy-mysql		SuccessfulCreate	create Claim datadir-gr-haproxy-mysql-0 Pod gr-haproxy-mysql-0 in StatefulSet gr-haproxy-mysql success	statefulset-controller	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:18:55 +0000 UTC	Normal	StatefulSet.apps gr-haproxy-mysql		SuccessfulCreate	create Pod gr-haproxy-mysql-0 in StatefulSet gr-haproxy-mysql successful	statefulset-controller	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:18:58 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-haproxy-mysql-0		ProvisioningSucceeded	Successfully provisioned volume pvc-4590053d-a3d6-4037-95a0-8fed1065d4ab	pd.csi.storage.gke.io_gke-b35679d75e564d2d98a0-bb78-fb92-vm_7b300296-10a9-4121-bf46-b146398f958e	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:18:59 +0000 UTC	Normal	Pod gr-haproxy-mysql-0	Binding	Scheduled	Successfully assigned kuttl-test-bright-bengal/gr-haproxy-mysql-0 to gke-jen-ps-873-af2827af--default-pool-f58c2b0d-wb8m	default-scheduler	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:06 +0000 UTC	Normal	Pod gr-haproxy-mysql-0		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-4590053d-a3d6-4037-95a0-8fed1065d4ab" 	attachdetach-controller	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:08 +0000 UTC	Normal	Pod gr-haproxy-mysql-0.spec.initContainers{mysql-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-873-af2827af"	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:08 +0000 UTC	Normal	Pod gr-haproxy-mysql-0.spec.initContainers{mysql-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-873-af2827af" in 163ms (164ms including waiting)	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:08 +0000 UTC	Normal	Pod gr-haproxy-mysql-0.spec.initContainers{mysql-init}		Created	Created container: mysql-init	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:08 +0000 UTC	Normal	Pod gr-haproxy-mysql-0.spec.initContainers{mysql-init}		Started	Started container mysql-init	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:11 +0000 UTC	Normal	Pod gr-haproxy-mysql-0.spec.containers{mysql}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-psmysql"	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:11 +0000 UTC	Normal	Pod gr-haproxy-mysql-0.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 171ms (171ms including waiting)	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:11 +0000 UTC	Normal	Pod gr-haproxy-mysql-0.spec.containers{mysql}		Created	Created container: mysql	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:11 +0000 UTC	Normal	Pod gr-haproxy-mysql-0.spec.containers{mysql}		Started	Started container mysql	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:11 +0000 UTC	Normal	Pod gr-haproxy-mysql-0.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:11 +0000 UTC	Normal	Pod gr-haproxy-mysql-0.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 118ms (118ms including waiting)	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:11 +0000 UTC	Normal	Pod gr-haproxy-mysql-0.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:11 +0000 UTC	Normal	Pod gr-haproxy-mysql-0.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:43 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-haproxy-mysql-1		WaitForFirstConsumer	waiting for first consumer to be created before binding	persistentvolume-controller	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:43 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-haproxy-mysql-1		ExternalProvisioning	Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.	persistentvolume-controller	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:43 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-haproxy-mysql-1		Provisioning	External provisioner is provisioning volume for claim "kuttl-test-bright-bengal/datadir-gr-haproxy-mysql-1"	pd.csi.storage.gke.io_gke-b35679d75e564d2d98a0-bb78-fb92-vm_7b300296-10a9-4121-bf46-b146398f958e	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:43 +0000 UTC	Normal	StatefulSet.apps gr-haproxy-mysql		SuccessfulCreate	create Claim datadir-gr-haproxy-mysql-1 Pod gr-haproxy-mysql-1 in StatefulSet gr-haproxy-mysql success	statefulset-controller	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:43 +0000 UTC	Normal	StatefulSet.apps gr-haproxy-mysql		SuccessfulCreate	create Pod gr-haproxy-mysql-1 in StatefulSet gr-haproxy-mysql successful	statefulset-controller	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:45 +0000 UTC	Normal	StatefulSet.apps gr-haproxy-haproxy		SuccessfulCreate	create Pod gr-haproxy-haproxy-0 in StatefulSet gr-haproxy-haproxy successful	statefulset-controller	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:46 +0000 UTC	Normal	Pod gr-haproxy-haproxy-0	Binding	Scheduled	Successfully assigned kuttl-test-bright-bengal/gr-haproxy-haproxy-0 to gke-jen-ps-873-af2827af--default-pool-f58c2b0d-wb8m	default-scheduler	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:46 +0000 UTC	Normal	Pod gr-haproxy-haproxy-0.spec.initContainers{haproxy-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-873-af2827af"	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:46 +0000 UTC	Normal	Pod gr-haproxy-haproxy-0.spec.initContainers{haproxy-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-873-af2827af" in 177ms (177ms including waiting)	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:46 +0000 UTC	Normal	Pod gr-haproxy-haproxy-0.spec.initContainers{haproxy-init}		Created	Created container: haproxy-init	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:46 +0000 UTC	Normal	Pod gr-haproxy-haproxy-0.spec.initContainers{haproxy-init}		Started	Started container haproxy-init	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:47 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-haproxy-mysql-1		ProvisioningSucceeded	Successfully provisioned volume pvc-7a6fe709-2775-47a4-9e73-9071008d711d	pd.csi.storage.gke.io_gke-b35679d75e564d2d98a0-bb78-fb92-vm_7b300296-10a9-4121-bf46-b146398f958e	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:47 +0000 UTC	Normal	Pod gr-haproxy-mysql-1	Binding	Scheduled	Successfully assigned kuttl-test-bright-bengal/gr-haproxy-mysql-1 to gke-jen-ps-873-af2827af--default-pool-f58c2b0d-xcsn	default-scheduler	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:49 +0000 UTC	Normal	Pod gr-haproxy-haproxy-0.spec.containers{haproxy}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-haproxy"	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:49 +0000 UTC	Normal	Pod gr-haproxy-haproxy-0.spec.containers{haproxy}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 124ms (124ms including waiting)	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:49 +0000 UTC	Normal	Pod gr-haproxy-haproxy-0.spec.containers{haproxy}		Created	Created container: haproxy	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:49 +0000 UTC	Normal	Pod gr-haproxy-haproxy-0.spec.containers{haproxy}		Started	Started container haproxy	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:49 +0000 UTC	Normal	Pod gr-haproxy-haproxy-0.spec.containers{mysql-monit}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-haproxy"	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:49 +0000 UTC	Normal	Pod gr-haproxy-haproxy-0.spec.containers{mysql-monit}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 127ms (127ms including waiting)	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:49 +0000 UTC	Normal	Pod gr-haproxy-haproxy-0.spec.containers{mysql-monit}		Created	Created container: mysql-monit	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:49 +0000 UTC	Normal	Pod gr-haproxy-haproxy-0.spec.containers{mysql-monit}		Started	Started container mysql-monit	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:50 +0000 UTC	Normal	Pod gr-haproxy-haproxy-1	Binding	Scheduled	Successfully assigned kuttl-test-bright-bengal/gr-haproxy-haproxy-1 to gke-jen-ps-873-af2827af--default-pool-f58c2b0d-xcsn	default-scheduler	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:50 +0000 UTC	Normal	Pod gr-haproxy-haproxy-1.spec.initContainers{haproxy-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-873-af2827af"	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:50 +0000 UTC	Normal	StatefulSet.apps gr-haproxy-haproxy		SuccessfulCreate	create Pod gr-haproxy-haproxy-1 in StatefulSet gr-haproxy-haproxy successful	statefulset-controller	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:51 +0000 UTC	Normal	Pod gr-haproxy-haproxy-1.spec.initContainers{haproxy-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-873-af2827af" in 203ms (203ms including waiting)	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:51 +0000 UTC	Normal	Pod gr-haproxy-haproxy-1.spec.initContainers{haproxy-init}		Created	Created container: haproxy-init	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:51 +0000 UTC	Normal	Pod gr-haproxy-haproxy-1.spec.initContainers{haproxy-init}		Started	Started container haproxy-init	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:53 +0000 UTC	Normal	Pod gr-haproxy-haproxy-1.spec.containers{haproxy}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-haproxy"	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:53 +0000 UTC	Normal	Pod gr-haproxy-haproxy-1.spec.containers{haproxy}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 159ms (160ms including waiting)	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:53 +0000 UTC	Normal	Pod gr-haproxy-haproxy-1.spec.containers{haproxy}		Created	Created container: haproxy	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:53 +0000 UTC	Normal	Pod gr-haproxy-haproxy-1.spec.containers{haproxy}		Started	Started container haproxy	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:53 +0000 UTC	Normal	Pod gr-haproxy-haproxy-1.spec.containers{mysql-monit}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-haproxy"	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:53 +0000 UTC	Normal	Pod gr-haproxy-haproxy-1.spec.containers{mysql-monit}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 144ms (144ms including waiting)	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:53 +0000 UTC	Normal	Pod gr-haproxy-haproxy-1.spec.containers{mysql-monit}		Created	Created container: mysql-monit	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:53 +0000 UTC	Normal	Pod gr-haproxy-haproxy-1.spec.containers{mysql-monit}		Started	Started container mysql-monit	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:54 +0000 UTC	Normal	Pod gr-haproxy-haproxy-2	Binding	Scheduled	Successfully assigned kuttl-test-bright-bengal/gr-haproxy-haproxy-2 to gke-jen-ps-873-af2827af--default-pool-f58c2b0d-10lv	default-scheduler	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:54 +0000 UTC	Normal	Pod gr-haproxy-haproxy-2.spec.initContainers{haproxy-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-873-af2827af"	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:54 +0000 UTC	Normal	StatefulSet.apps gr-haproxy-haproxy		SuccessfulCreate	create Pod gr-haproxy-haproxy-2 in StatefulSet gr-haproxy-haproxy successful	statefulset-controller	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:55 +0000 UTC	Normal	Pod gr-haproxy-haproxy-2.spec.initContainers{haproxy-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-873-af2827af" in 186ms (186ms including waiting)	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:55 +0000 UTC	Normal	Pod gr-haproxy-haproxy-2.spec.initContainers{haproxy-init}		Created	Created container: haproxy-init	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:55 +0000 UTC	Normal	Pod gr-haproxy-haproxy-2.spec.initContainers{haproxy-init}		Started	Started container haproxy-init	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:55 +0000 UTC	Normal	Pod gr-haproxy-mysql-1		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-7a6fe709-2775-47a4-9e73-9071008d711d" 	attachdetach-controller	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:56 +0000 UTC	Normal	Pod gr-haproxy-haproxy-2.spec.containers{haproxy}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-haproxy"	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:56 +0000 UTC	Normal	Pod gr-haproxy-mysql-1.spec.initContainers{mysql-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-873-af2827af"	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:56 +0000 UTC	Normal	Pod gr-haproxy-mysql-1.spec.initContainers{mysql-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-873-af2827af" in 181ms (181ms including waiting)	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:56 +0000 UTC	Normal	Pod gr-haproxy-mysql-1.spec.initContainers{mysql-init}		Created	Created container: mysql-init	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:56 +0000 UTC	Normal	Pod gr-haproxy-mysql-1.spec.initContainers{mysql-init}		Started	Started container mysql-init	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:57 +0000 UTC	Normal	Pod gr-haproxy-haproxy-2.spec.containers{haproxy}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 135ms (135ms including waiting)	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:57 +0000 UTC	Normal	Pod gr-haproxy-haproxy-2.spec.containers{haproxy}		Created	Created container: haproxy	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:57 +0000 UTC	Normal	Pod gr-haproxy-haproxy-2.spec.containers{haproxy}		Started	Started container haproxy	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:57 +0000 UTC	Normal	Pod gr-haproxy-haproxy-2.spec.containers{mysql-monit}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-haproxy"	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:57 +0000 UTC	Normal	Pod gr-haproxy-haproxy-2.spec.containers{mysql-monit}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 121ms (121ms including waiting)	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:57 +0000 UTC	Normal	Pod gr-haproxy-haproxy-2.spec.containers{mysql-monit}		Created	Created container: mysql-monit	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:57 +0000 UTC	Normal	Pod gr-haproxy-haproxy-2.spec.containers{mysql-monit}		Started	Started container mysql-monit	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:59 +0000 UTC	Normal	Pod gr-haproxy-mysql-1.spec.containers{mysql}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-psmysql"	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:59 +0000 UTC	Normal	Pod gr-haproxy-mysql-1.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 145ms (145ms including waiting)	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:59 +0000 UTC	Normal	Pod gr-haproxy-mysql-1.spec.containers{mysql}		Created	Created container: mysql	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:59 +0000 UTC	Normal	Pod gr-haproxy-mysql-1.spec.containers{mysql}		Started	Started container mysql	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:59 +0000 UTC	Normal	Pod gr-haproxy-mysql-1.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:59 +0000 UTC	Normal	Pod gr-haproxy-mysql-1.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 120ms (120ms including waiting)	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:59 +0000 UTC	Normal	Pod gr-haproxy-mysql-1.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:19:59 +0000 UTC	Normal	Pod gr-haproxy-mysql-1.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:20:29 +0000 UTC	Warning	Pod gr-haproxy-mysql-1.spec.containers{mysql}		Unhealthy	Startup probe failed: 2025/03/26 02:20:16 Waiting for MySQL ready state
        2025/03/26 02:20:16 MySQL is ready
        2025/03/26 02:20:16 Bootstrap starting...
        2025/03/26 02:20:16 Running dba.configureLocalInstance('operator:*****@gr-haproxy-mysql-1.gr-haproxy-mysql.kuttl-test-bright-bengal', {'clearReadOnly': true})
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        WARNING: The clearReadOnly option is deprecated and will be removed in a future release.
        
        WARNING: This function is deprecated and will be removed in a future release of MySQL Shell, use dba.configureInstance() instead.
        Configuring local MySQL instance listening at port 3306 for use in an InnoDB cluster...
        
        This instance reports its own address as gr-haproxy-mysql-1.gr-haproxy-mysql.kuttl-test-bright-bengal:3306
        
        applierWorkerThreads will be set to the default value of 4.
        
        NOTE: Some configuration options need to be fixed:
        
        
        +----------------------------------------+---------------+----------------+----------------------------+
        
        +----------------------------------------+---------------+----------------+----------------------------+
        | Variable                               | Current Value | Required Value | Note                       |+----------------------------------------+---------------+----------------+----------------------------+
        | binlog_transaction_dependency_tracking | COMMIT_ORDER  | WRITESET       | Update the server variable |Disabled super_read_only on the instance 'gr-haproxy-mysql-1.gr-haproxy-mysql.kuttl-test-bright-bengal:3306'
        Enabling super_read_only on the instance 'gr-haproxy-mysql-1.gr-haproxy-mysql.kuttl-test-bright-bengal:3306'
        Configuring instance...
        
        WARNING:*****@binlog_transaction_dependency_tracking' is deprecated and will be removed in a future release. (Code 1287).
        The instance 'gr-haproxy-mysql-1.gr-haproxy-mysql.kuttl-test-bright-bengal:3306' was configured to be used in an InnoDB cluster.
        2025/03/26 02:20:16 Instance (gr-haproxy-mysql-1.gr-haproxy-mysql.kuttl-test-bright-bengal) configured to join to the InnoDB cluster
        2025/03/26 02:20:16 peers: [gr-haproxy-mysql-0.gr-haproxy-mysql.kuttl-test-bright-bengal gr-haproxy-mysql-1.gr-haproxy-mysql.kuttl-test-bright-bengal]
        2025/03/26 02:20:16 Running dba.getCluster('grhaproxy')
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        2025/03/26 02:20:17 Connected to peer gr-haproxy-mysql-0.gr-haproxy-mysql.kuttl-test-bright-bengal
        2025/03/26 02:20:17 Cluster status:
        
        ClusterName: grhaproxy
        Status: OK_NO_TOLERANCE
        StatusText: Cluster is NOT tolerant to any failures.
        SSL: REQUIRED
        Primary: gr-haproxy-mysql-0.gr-haproxy-mysql.kuttl-test-bright-bengal:3306
        Topology:
        	
        	Member 0
        	Address: gr-haproxy-mysql-0.gr-haproxy-mysql.kuttl-test-bright-bengal:3306
        	State: ONLINE
        	Errors: []
        
        		
        2025/03/26 02:20:17 Adding instance (gr-haproxy-mysql-1.gr-haproxy-mysql.kuttl-test-bright-bengal) to InnoDB cluster
        2025/03/26 02:20:17 Running dba.getCluster('grhaproxy').addInstance('operator:*****@gr-haproxy-mysql-1.gr-haproxy-mysql.kuttl-test-bright-bengal', {'recoveryMethod': 'clone', 'waitRecovery': 3})
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        WARNING: The waitRecovery option is deprecated. Please use the recoveryProgress option instead.
        
        
        NOTE: The target instance 'gr-haproxy-mysql-1.gr-haproxy-mysql.kuttl-test-bright-bengal:3306' has not been pre-provisioned (GTID set is empty). The Shell is unable to decide whether incremental state recovery can correctly provision it.
        
        Clone based recovery selected through the recoveryMethod option
        
        Validating instance configuration at gr-haproxy-mysql-1.gr-haproxy-mysql.kuttl-test-bright-bengal:3306...
        
        This instance reports its own address as gr-haproxy-mysql-1.gr-haproxy-mysql.kuttl-test-bright-bengal:3306
        
        Instance configuration is suitable.
        NOTE: Group Replication will communicate with other members using 'gr-haproxy-mysql-1.gr-haproxy-mysql.kuttl-test-bright-bengal:3306'. Use the localAddress option to override.
        
        * Checking connectivity and SSL configuration...
        A new instance will be added to the InnoDB Cluster. Depending on the amount of
        data on the cluster this might take from a few seconds to several hours.
        
        Adding instance to the cluster...
        
        Monitoring recovery process of the new cluster member. Press ^C to stop monitoring and let it continue in background.
        Clone based state recovery is now in progress.
        
        NOTE: A server restart is expected to happen as part of the clone process. If the
        server does not support the RESTART command or does not come back after a
        while, you may need to manually start it back.
        
        * Waiting for clone to finish...
        NOTE: gr-haproxy-mysql-1.gr-haproxy-mysql.kuttl-test-bright-bengal:3306 is being cloned from gr-haproxy-mysql-0.gr-haproxy-mysql.kuttl-test-bright-bengal:3306
        ** Stage DROP DATA: Completed
        ** Clone Transfer      FILE COPY      0%  In Progress    PAGE COPY      0%  Not Started    REDO COPY      0%  Not Started** Clone Transfer      FILE COPY    100%  Completed    PAGE COPY    100%  Completed    REDO COPY    100%  Completed
        NOTE: gr-haproxy-mysql-1.gr-haproxy-mysql.kuttl-test-bright-bengal:3306 is shutting down...
        
        * Waiting for server restart... \ 
* Waiting for server restart... | 
* Waiting for server restart... / 
* Waiting for server restart... - 
	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:20:29 +0000 UTC	Normal	Pod gr-haproxy-mysql-1.spec.containers{mysql}		Killing	Container mysql failed startup probe, will be restarted	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:20:29 +0000 UTC	Warning	Pod gr-haproxy-mysql-1.spec.containers{mysql}		FailedPreStopHook	PreStopHook failed	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:20:29 +0000 UTC	Normal	Pod gr-haproxy-mysql-1.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 142ms (142ms including waiting)	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:21:01 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-haproxy-mysql-2		WaitForFirstConsumer	waiting for first consumer to be created before binding	persistentvolume-controller	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:21:01 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-haproxy-mysql-2		ExternalProvisioning	Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.	persistentvolume-controller	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:21:01 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-haproxy-mysql-2		Provisioning	External provisioner is provisioning volume for claim "kuttl-test-bright-bengal/datadir-gr-haproxy-mysql-2"	pd.csi.storage.gke.io_gke-b35679d75e564d2d98a0-bb78-fb92-vm_7b300296-10a9-4121-bf46-b146398f958e	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:21:01 +0000 UTC	Normal	StatefulSet.apps gr-haproxy-mysql		SuccessfulCreate	create Claim datadir-gr-haproxy-mysql-2 Pod gr-haproxy-mysql-2 in StatefulSet gr-haproxy-mysql success	statefulset-controller	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:21:01 +0000 UTC	Normal	StatefulSet.apps gr-haproxy-mysql		SuccessfulCreate	create Pod gr-haproxy-mysql-2 in StatefulSet gr-haproxy-mysql successful	statefulset-controller	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:21:07 +0000 UTC	Normal	PersistentVolumeClaim datadir-gr-haproxy-mysql-2		ProvisioningSucceeded	Successfully provisioned volume pvc-73c4d4d1-b0e1-4336-8728-bb90fa7f4b23	pd.csi.storage.gke.io_gke-b35679d75e564d2d98a0-bb78-fb92-vm_7b300296-10a9-4121-bf46-b146398f958e	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:21:08 +0000 UTC	Normal	Pod gr-haproxy-mysql-2	Binding	Scheduled	Successfully assigned kuttl-test-bright-bengal/gr-haproxy-mysql-2 to gke-jen-ps-873-af2827af--default-pool-f58c2b0d-10lv	default-scheduler	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:21:16 +0000 UTC	Normal	Pod gr-haproxy-mysql-2		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-73c4d4d1-b0e1-4336-8728-bb90fa7f4b23" 	attachdetach-controller	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:21:17 +0000 UTC	Normal	Pod gr-haproxy-mysql-2.spec.initContainers{mysql-init}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:PR-873-af2827af"	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:21:17 +0000 UTC	Normal	Pod gr-haproxy-mysql-2.spec.initContainers{mysql-init}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:PR-873-af2827af" in 174ms (174ms including waiting)	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:21:17 +0000 UTC	Normal	Pod gr-haproxy-mysql-2.spec.initContainers{mysql-init}		Created	Created container: mysql-init	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:21:17 +0000 UTC	Normal	Pod gr-haproxy-mysql-2.spec.initContainers{mysql-init}		Started	Started container mysql-init	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:21:19 +0000 UTC	Normal	Pod gr-haproxy-mysql-2.spec.containers{mysql}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-psmysql"	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:21:19 +0000 UTC	Normal	Pod gr-haproxy-mysql-2.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 117ms (117ms including waiting)	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:21:19 +0000 UTC	Normal	Pod gr-haproxy-mysql-2.spec.containers{mysql}		Created	Created container: mysql	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:21:19 +0000 UTC	Normal	Pod gr-haproxy-mysql-2.spec.containers{mysql}		Started	Started container mysql	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:21:19 +0000 UTC	Normal	Pod gr-haproxy-mysql-2.spec.containers{xtrabackup}		Pulling	Pulling image "perconalab/percona-server-mysql-operator:main-backup"	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:21:19 +0000 UTC	Normal	Pod gr-haproxy-mysql-2.spec.containers{xtrabackup}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 133ms (133ms including waiting)	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:21:19 +0000 UTC	Normal	Pod gr-haproxy-mysql-2.spec.containers{xtrabackup}		Created	Created container: xtrabackup	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:21:19 +0000 UTC	Normal	Pod gr-haproxy-mysql-2.spec.containers{xtrabackup}		Started	Started container xtrabackup	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:21:50 +0000 UTC	Warning	Pod gr-haproxy-mysql-2.spec.containers{mysql}		Unhealthy	Startup probe failed: 2025/03/26 02:21:37 Waiting for MySQL ready state
        2025/03/26 02:21:37 MySQL is ready
        2025/03/26 02:21:37 Bootstrap starting...
        2025/03/26 02:21:37 Running dba.configureLocalInstance('operator:*****@gr-haproxy-mysql-2.gr-haproxy-mysql.kuttl-test-bright-bengal', {'clearReadOnly': true})
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        WARNING: The clearReadOnly option is deprecated and will be removed in a future release.
        
        WARNING: This function is deprecated and will be removed in a future release of MySQL Shell, use dba.configureInstance() instead.
        Configuring local MySQL instance listening at port 3306 for use in an InnoDB cluster...
        
        This instance reports its own address as gr-haproxy-mysql-2.gr-haproxy-mysql.kuttl-test-bright-bengal:3306
        
        applierWorkerThreads will be set to the default value of 4.
        
        NOTE: Some configuration options need to be fixed:
        
        
        +----------------------------------------+---------------+----------------+----------------------------+
        
        +----------------------------------------+---------------+----------------+----------------------------+
        | Variable                               | Current Value | Required Value | Note                       |+----------------------------------------+---------------+----------------+----------------------------+
        | binlog_transaction_dependency_tracking | COMMIT_ORDER  | WRITESET       | Update the server variable |Disabled super_read_only on the instance 'gr-haproxy-mysql-2.gr-haproxy-mysql.kuttl-test-bright-bengal:3306'
        Enabling super_read_only on the instance 'gr-haproxy-mysql-2.gr-haproxy-mysql.kuttl-test-bright-bengal:3306'
        Configuring instance...
        
        WARNING:*****@binlog_transaction_dependency_tracking' is deprecated and will be removed in a future release. (Code 1287).
        The instance 'gr-haproxy-mysql-2.gr-haproxy-mysql.kuttl-test-bright-bengal:3306' was configured to be used in an InnoDB cluster.
        2025/03/26 02:21:37 Instance (gr-haproxy-mysql-2.gr-haproxy-mysql.kuttl-test-bright-bengal) configured to join to the InnoDB cluster
        2025/03/26 02:21:37 peers: [gr-haproxy-mysql-0.gr-haproxy-mysql.kuttl-test-bright-bengal gr-haproxy-mysql-1.gr-haproxy-mysql.kuttl-test-bright-bengal gr-haproxy-mysql-2.gr-haproxy-mysql.kuttl-test-bright-bengal]
        2025/03/26 02:21:37 Running dba.getCluster('grhaproxy')
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        2025/03/26 02:21:38 Connected to peer gr-haproxy-mysql-0.gr-haproxy-mysql.kuttl-test-bright-bengal
        2025/03/26 02:21:38 Cluster status:
        
        ClusterName: grhaproxy
        Status: OK_NO_TOLERANCE
        StatusText: Cluster is NOT tolerant to any failures.
        SSL: REQUIRED
        Primary: gr-haproxy-mysql-0.gr-haproxy-mysql.kuttl-test-bright-bengal:3306
        Topology:
        	
        	Member 0
        	Address: gr-haproxy-mysql-0.gr-haproxy-mysql.kuttl-test-bright-bengal:3306
        	State: ONLINE
        	Errors: []
        
        		
        	Member 1
        	Address: gr-haproxy-mysql-1.gr-haproxy-mysql.kuttl-test-bright-bengal:3306
        	State: ONLINE
        	Errors: []
        
        		
        2025/03/26 02:21:38 Adding instance (gr-haproxy-mysql-2.gr-haproxy-mysql.kuttl-test-bright-bengal) to InnoDB cluster
        2025/03/26 02:21:38 Running dba.getCluster('grhaproxy').addInstance('operator:*****@gr-haproxy-mysql-2.gr-haproxy-mysql.kuttl-test-bright-bengal', {'recoveryMethod': 'clone', 'waitRecovery': 3})
        Cannot set LC_ALL to locale en_US.UTF-8: No such file or directory
        WARNING: Using a password on the command line interface can be insecure.
        WARNING: The waitRecovery option is deprecated. Please use the recoveryProgress option instead.
        
        
        NOTE: The target instance 'gr-haproxy-mysql-2.gr-haproxy-mysql.kuttl-test-bright-bengal:3306' has not been pre-provisioned (GTID set is empty). The Shell is unable to decide whether incremental state recovery can correctly provision it.
        
        Clone based recovery selected through the recoveryMethod option
        
        Validating instance configuration at gr-haproxy-mysql-2.gr-haproxy-mysql.kuttl-test-bright-bengal:3306...
        
        This instance reports its own address as gr-haproxy-mysql-2.gr-haproxy-mysql.kuttl-test-bright-bengal:3306
        
        Instance configuration is suitable.
        NOTE: Group Replication will communicate with other members using 'gr-haproxy-mysql-2.gr-haproxy-mysql.kuttl-test-bright-bengal:3306'. Use the localAddress option to override.
        
        * Checking connectivity and SSL configuration...
        A new instance will be added to the InnoDB Cluster. Depending on the amount of
        data on the cluster this might take from a few seconds to several hours.
        
        Adding instance to the cluster...
        
        Monitoring recovery process of the new cluster member. Press ^C to stop monitoring and let it continue in background.
        Clone based state recovery is now in progress.
        
        NOTE: A server restart is expected to happen as part of the clone process. If the
        server does not support the RESTART command or does not come back after a
        while, you may need to manually start it back.
        
        * Waiting for clone to finish...
        NOTE: gr-haproxy-mysql-2.gr-haproxy-mysql.kuttl-test-bright-bengal:3306 is being cloned from gr-haproxy-mysql-1.gr-haproxy-mysql.kuttl-test-bright-bengal:3306
        ** Stage DROP DATA: Completed
        ** Clone Transfer      FILE COPY    100%  Completed    PAGE COPY    100%  Completed    REDO COPY    100%  Completed
        NOTE: gr-haproxy-mysql-2.gr-haproxy-mysql.kuttl-test-bright-bengal:3306 is shutting down...
        
        * Waiting for server restart... \ 
* Waiting for server restart... | 
* Waiting for server restart... / 
* Waiting for server restart... - 
	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:21:50 +0000 UTC	Normal	Pod gr-haproxy-mysql-2.spec.containers{mysql}		Killing	Container mysql failed startup probe, will be restarted	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:21:50 +0000 UTC	Warning	Pod gr-haproxy-mysql-2.spec.containers{mysql}		FailedPreStopHook	PreStopHook failed	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:21:50 +0000 UTC	Normal	Pod gr-haproxy-mysql-2.spec.containers{mysql}		Pulled	Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 138ms (138ms including waiting)	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:23:00 +0000 UTC	Normal	Pod gr-haproxy-haproxy-0.spec.containers{haproxy}		Killing	Stopping container haproxy	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:23:00 +0000 UTC	Normal	Pod gr-haproxy-haproxy-0.spec.containers{mysql-monit}		Killing	Stopping container mysql-monit	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:23:00 +0000 UTC	Normal	Pod gr-haproxy-haproxy-1.spec.containers{haproxy}		Killing	Stopping container haproxy	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:23:00 +0000 UTC	Normal	Pod gr-haproxy-haproxy-1.spec.containers{mysql-monit}		Killing	Stopping container mysql-monit	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:23:00 +0000 UTC	Normal	Pod gr-haproxy-haproxy-2.spec.containers{haproxy}		Killing	Stopping container haproxy	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:23:00 +0000 UTC	Normal	Pod gr-haproxy-haproxy-2.spec.containers{mysql-monit}		Killing	Stopping container mysql-monit	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:23:00 +0000 UTC	Normal	Pod gr-haproxy-mysql-0.spec.containers{mysql}		Killing	Stopping container mysql	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:23:00 +0000 UTC	Normal	Pod gr-haproxy-mysql-0.spec.containers{xtrabackup}		Killing	Stopping container xtrabackup	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:23:00 +0000 UTC	Normal	Pod gr-haproxy-mysql-1.spec.containers{xtrabackup}		Killing	Stopping container xtrabackup	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:23:00 +0000 UTC	Normal	Pod gr-haproxy-mysql-1.spec.containers{mysql}		Killing	Stopping container mysql	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:23:00 +0000 UTC	Normal	Pod gr-haproxy-mysql-2.spec.containers{xtrabackup}		Killing	Stopping container xtrabackup	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:23:00 +0000 UTC	Normal	Pod gr-haproxy-mysql-2.spec.containers{mysql}		Killing	Stopping container mysql	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:23:01 +0000 UTC	Warning	Pod gr-haproxy-mysql-1.spec.containers{mysql}		Unhealthy	Readiness probe failed: 2025/03/26 02:23:01 MySQL state is not ready...
        	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:23:02 +0000 UTC	Warning	Pod gr-haproxy-mysql-2.spec.containers{mysql}		Unhealthy	Readiness probe failed: 2025/03/26 02:23:02 MySQL state is not ready...
        	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | 2025-03-26 02:23:06 +0000 UTC	Warning	Pod gr-haproxy-mysql-1.spec.containers{mysql}		Unhealthy	Readiness probe failed: 2025/03/26 02:23:06 MySQL state is not ready...
        	kubelet	
    logger.go:42: 02:23:07 | gr-haproxy | Deleting namespace: kuttl-test-bright-bengal
=== NAME  kuttl
    harness.go:407: run tests finished
    harness.go:515: cleaning up
    harness.go:572: removing temp folder: ""
--- PASS: kuttl (312.17s)
    --- PASS: kuttl/harness (0.00s)
        --- PASS: kuttl/harness/gr-haproxy (311.73s)
PASS