=== RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://35.232.42.60 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 180 seconds for each step harness.go:372: testsuite: e2e-tests/tests has 28 tests === RUN kuttl/harness === RUN kuttl/harness/haproxy === PAUSE kuttl/harness/haproxy === CONT kuttl/harness/haproxy logger.go:42: 14:47:29 | haproxy | Creating namespace: kuttl-test-viable-phoenix logger.go:42: 14:47:29 | haproxy/0-deploy-operator | starting test step 0-deploy-operator logger.go:42: 14:47:29 | haproxy/0-deploy-operator | running command: [sh -c set -o errexit set -o xtrace source ../../functions deploy_operator deploy_non_tls_cluster_secrets deploy_tls_cluster_secrets deploy_client] logger.go:42: 14:47:29 | haproxy/0-deploy-operator | + source ../../functions logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ realpath ../../.. logger.go:42: 14:47:29 | haproxy/0-deploy-operator | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:47:29 | haproxy/0-deploy-operator | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/vars.sh logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:47:29 | haproxy/0-deploy-operator | ++++ mktemp -d logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ export TEMP_DIR=/tmp/tmp.oVhMxqsBwp logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ TEMP_DIR=/tmp/tmp.oVhMxqsBwp logger.go:42: 14:47:29 | haproxy/0-deploy-operator | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ export GIT_BRANCH=PR-424 logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ GIT_BRANCH=PR-424 logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ export VERSION=PR-424-70568ae logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ VERSION=PR-424-70568ae logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:47:29 | haproxy/0-deploy-operator | ++++ which gdate logger.go:42: 14:47:29 | haproxy/0-deploy-operator | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-424/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:47:29 | haproxy/0-deploy-operator | ++++ which date logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ date=/usr/bin/date logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ command -v oc logger.go:42: 14:47:29 | haproxy/0-deploy-operator | +++ oc get projects logger.go:42: 14:47:34 | haproxy/0-deploy-operator | error: the server doesn't have a resource type "projects" logger.go:42: 14:47:34 | haproxy/0-deploy-operator | +++ grep '^minikube' logger.go:42: 14:47:34 | haproxy/0-deploy-operator | +++ kubectl get nodes logger.go:42: 14:47:35 | haproxy/0-deploy-operator | ++++ pwd logger.go:42: 14:47:35 | haproxy/0-deploy-operator | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/tests/haproxy logger.go:42: 14:47:35 | haproxy/0-deploy-operator | ++ test_name=haproxy logger.go:42: 14:47:35 | haproxy/0-deploy-operator | + deploy_operator logger.go:42: 14:47:35 | haproxy/0-deploy-operator | + kubectl -n kuttl-test-viable-phoenix apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy/crd.yaml logger.go:42: 14:47:37 | haproxy/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlbackups.ps.percona.com serverside-applied logger.go:42: 14:47:37 | haproxy/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlrestores.ps.percona.com serverside-applied logger.go:42: 14:47:38 | haproxy/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqls.ps.percona.com serverside-applied logger.go:42: 14:47:38 | haproxy/0-deploy-operator | + kubectl -n kuttl-test-viable-phoenix apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy/rbac.yaml logger.go:42: 14:47:39 | haproxy/0-deploy-operator | serviceaccount/percona-server-mysql-operator created logger.go:42: 14:47:40 | haproxy/0-deploy-operator | serviceaccount/percona-server-mysql-operator-orchestrator created logger.go:42: 14:47:40 | haproxy/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 14:47:40 | haproxy/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 14:47:41 | haproxy/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 14:47:41 | haproxy/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 14:47:41 | haproxy/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 14:47:42 | haproxy/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 14:47:42 | haproxy/0-deploy-operator | + yq eval '(select(documentIndex==1).spec.template.spec.containers[] | select(.name=="manager").env[] | select(.name=="DISABLE_TELEMETRY").value) = "true"' logger.go:42: 14:47:42 | haproxy/0-deploy-operator | + kubectl -n kuttl-test-viable-phoenix apply -f - logger.go:42: 14:47:42 | haproxy/0-deploy-operator | ++ printf 'select(documentIndex==1).spec.template.spec.containers[0].image="%s"' perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:47:42 | haproxy/0-deploy-operator | + yq eval 'select(documentIndex==1).spec.template.spec.containers[0].image="perconalab/percona-server-mysql-operator:PR-424-70568ae"' /mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy/operator.yaml logger.go:42: 14:47:43 | haproxy/0-deploy-operator | configmap/percona-server-mysql-operator-config created logger.go:42: 14:47:43 | haproxy/0-deploy-operator | deployment.apps/percona-server-mysql-operator created logger.go:42: 14:47:43 | haproxy/0-deploy-operator | + deploy_non_tls_cluster_secrets logger.go:42: 14:47:43 | haproxy/0-deploy-operator | + kubectl -n kuttl-test-viable-phoenix apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf/secrets.yaml logger.go:42: 14:47:45 | haproxy/0-deploy-operator | secret/test-secrets created logger.go:42: 14:47:45 | haproxy/0-deploy-operator | + deploy_tls_cluster_secrets logger.go:42: 14:47:45 | haproxy/0-deploy-operator | + kubectl -n kuttl-test-viable-phoenix apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf/ssl-secret.yaml logger.go:42: 14:47:46 | haproxy/0-deploy-operator | secret/test-ssl created logger.go:42: 14:47:46 | haproxy/0-deploy-operator | + deploy_client logger.go:42: 14:47:46 | haproxy/0-deploy-operator | + kubectl -n kuttl-test-viable-phoenix apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf/client.yaml logger.go:42: 14:47:47 | haproxy/0-deploy-operator | pod/mysql-client created logger.go:42: 14:47:55 | haproxy/0-deploy-operator | test step completed 0-deploy-operator logger.go:42: 14:47:55 | haproxy/1-create-cluster | starting test step 1-create-cluster logger.go:42: 14:47:55 | haproxy/1-create-cluster | running command: [sh -c set -o errexit set -o xtrace source ../../functions get_cr \ | yq eval '.spec.allowUnsafeConfigurations=false' - \ | yq eval '.spec.updateStrategy="RollingUpdate"' - \ | yq eval '.spec.mysql.clusterType="async"' - \ | yq eval '.spec.orchestrator.enabled=true' - \ | yq eval '.spec.proxy.haproxy.enabled=true' - \ | yq eval '.spec.proxy.haproxy.size=3' - \ | kubectl -n "${NAMESPACE}" apply -f -] logger.go:42: 14:47:55 | haproxy/1-create-cluster | + source ../../functions logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ realpath ../../.. logger.go:42: 14:47:55 | haproxy/1-create-cluster | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:47:55 | haproxy/1-create-cluster | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/vars.sh logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:47:55 | haproxy/1-create-cluster | ++++ mktemp -d logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ export TEMP_DIR=/tmp/tmp.TCqwikUkZT logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ TEMP_DIR=/tmp/tmp.TCqwikUkZT logger.go:42: 14:47:55 | haproxy/1-create-cluster | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ export GIT_BRANCH=PR-424 logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ GIT_BRANCH=PR-424 logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ export VERSION=PR-424-70568ae logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ VERSION=PR-424-70568ae logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:47:55 | haproxy/1-create-cluster | ++++ which gdate logger.go:42: 14:47:55 | haproxy/1-create-cluster | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-424/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:47:55 | haproxy/1-create-cluster | ++++ which date logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ date=/usr/bin/date logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ command -v oc logger.go:42: 14:47:55 | haproxy/1-create-cluster | +++ oc get projects logger.go:42: 14:48:00 | haproxy/1-create-cluster | error: the server doesn't have a resource type "projects" logger.go:42: 14:48:00 | haproxy/1-create-cluster | +++ grep '^minikube' logger.go:42: 14:48:00 | haproxy/1-create-cluster | +++ kubectl get nodes logger.go:42: 14:48:01 | haproxy/1-create-cluster | ++++ pwd logger.go:42: 14:48:01 | haproxy/1-create-cluster | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/tests/haproxy logger.go:42: 14:48:01 | haproxy/1-create-cluster | ++ test_name=haproxy logger.go:42: 14:48:01 | haproxy/1-create-cluster | + get_cr logger.go:42: 14:48:01 | haproxy/1-create-cluster | + local name_suffix= logger.go:42: 14:48:01 | haproxy/1-create-cluster | + yq eval .spec.allowUnsafeConfigurations=false - logger.go:42: 14:48:01 | haproxy/1-create-cluster | + yq eval '.spec.updateStrategy="RollingUpdate"' - logger.go:42: 14:48:01 | haproxy/1-create-cluster | + yq eval .spec.orchestrator.enabled=true - logger.go:42: 14:48:01 | haproxy/1-create-cluster | + yq eval .spec.proxy.haproxy.size=3 - logger.go:42: 14:48:01 | haproxy/1-create-cluster | + yq eval .spec.proxy.haproxy.enabled=true - logger.go:42: 14:48:01 | haproxy/1-create-cluster | + kubectl -n kuttl-test-viable-phoenix apply -f - logger.go:42: 14:48:01 | haproxy/1-create-cluster | + '[' -n '' ']' logger.go:42: 14:48:01 | haproxy/1-create-cluster | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 14:48:01 | haproxy/1-create-cluster | + yq eval - logger.go:42: 14:48:01 | haproxy/1-create-cluster | ++ printf '.spec.pmm.image="%s"' perconalab/pmm-client:dev-latest logger.go:42: 14:48:01 | haproxy/1-create-cluster | ++ printf '.spec.proxy.haproxy.image="%s"' perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:48:01 | haproxy/1-create-cluster | + yq eval '.spec.upgradeOptions.apply="disabled"' - logger.go:42: 14:48:01 | haproxy/1-create-cluster | + yq eval '.spec.pmm.image="perconalab/pmm-client:dev-latest"' - logger.go:42: 14:48:01 | haproxy/1-create-cluster | + yq eval '.spec.proxy.haproxy.image="perconalab/percona-server-mysql-operator:main-haproxy"' - logger.go:42: 14:48:01 | haproxy/1-create-cluster | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 14:48:01 | haproxy/1-create-cluster | ++ printf '.spec.proxy.router.image="%s"' perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:48:01 | haproxy/1-create-cluster | ++ printf '.spec.orchestrator.image="%s"' perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:48:01 | haproxy/1-create-cluster | + yq eval '.spec.orchestrator.image="perconalab/percona-server-mysql-operator:main-orchestrator"' - logger.go:42: 14:48:01 | haproxy/1-create-cluster | + yq eval '.spec.proxy.router.image="perconalab/percona-server-mysql-operator:main-router"' - logger.go:42: 14:48:01 | haproxy/1-create-cluster | ++ printf '.metadata.name="%s"' haproxy logger.go:42: 14:48:01 | haproxy/1-create-cluster | ++ printf '.spec.backup.image="%s"' perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:48:01 | haproxy/1-create-cluster | + yq eval '.spec.backup.image="perconalab/percona-server-mysql-operator:main-backup"' - logger.go:42: 14:48:01 | haproxy/1-create-cluster | + yq eval '.metadata.name="haproxy"' /mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy/cr.yaml logger.go:42: 14:48:01 | haproxy/1-create-cluster | ++ printf '.spec.initImage="%s"' perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:48:01 | haproxy/1-create-cluster | + yq eval '.spec.initImage="perconalab/percona-server-mysql-operator:PR-424-70568ae"' - logger.go:42: 14:48:01 | haproxy/1-create-cluster | ++ printf '.spec.mysql.image="%s"' perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:48:01 | haproxy/1-create-cluster | + yq eval '.spec.mysql.image="perconalab/percona-server-mysql-operator:main-psmysql"' - logger.go:42: 14:48:01 | haproxy/1-create-cluster | + yq eval '.spec.secretsName="test-secrets"' - logger.go:42: 14:48:01 | haproxy/1-create-cluster | ++ printf '.spec.toolkit.image="%s"' perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:48:01 | haproxy/1-create-cluster | + yq eval '.spec.sslSecretName="test-ssl"' - logger.go:42: 14:48:01 | haproxy/1-create-cluster | + yq eval '.spec.toolkit.image="perconalab/percona-server-mysql-operator:main-toolkit"' - logger.go:42: 14:48:03 | haproxy/1-create-cluster | perconaservermysql.ps.percona.com/haproxy created logger.go:42: 14:51:24 | haproxy/1-create-cluster | test step completed 1-create-cluster logger.go:42: 14:51:24 | haproxy/2-write-data | starting test step 2-write-data logger.go:42: 14:51:24 | haproxy/2-write-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password" run_mysql \ "INSERT myDB.myTable (id) VALUES (100500)" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password"] logger.go:42: 14:51:24 | haproxy/2-write-data | + source ../../functions logger.go:42: 14:51:24 | haproxy/2-write-data | +++ realpath ../../.. logger.go:42: 14:51:24 | haproxy/2-write-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:51:24 | haproxy/2-write-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/vars.sh logger.go:42: 14:51:24 | haproxy/2-write-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:51:24 | haproxy/2-write-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:51:24 | haproxy/2-write-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:51:24 | haproxy/2-write-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:51:24 | haproxy/2-write-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:51:24 | haproxy/2-write-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:51:24 | haproxy/2-write-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:51:24 | haproxy/2-write-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:51:24 | haproxy/2-write-data | ++++ mktemp -d logger.go:42: 14:51:24 | haproxy/2-write-data | +++ export TEMP_DIR=/tmp/tmp.tpMF95uXMw logger.go:42: 14:51:24 | haproxy/2-write-data | +++ TEMP_DIR=/tmp/tmp.tpMF95uXMw logger.go:42: 14:51:24 | haproxy/2-write-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:51:24 | haproxy/2-write-data | +++ export GIT_BRANCH=PR-424 logger.go:42: 14:51:24 | haproxy/2-write-data | +++ GIT_BRANCH=PR-424 logger.go:42: 14:51:24 | haproxy/2-write-data | +++ export VERSION=PR-424-70568ae logger.go:42: 14:51:24 | haproxy/2-write-data | +++ VERSION=PR-424-70568ae logger.go:42: 14:51:24 | haproxy/2-write-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:51:24 | haproxy/2-write-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:51:24 | haproxy/2-write-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:51:24 | haproxy/2-write-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:51:24 | haproxy/2-write-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:51:24 | haproxy/2-write-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:51:24 | haproxy/2-write-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:51:24 | haproxy/2-write-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:51:24 | haproxy/2-write-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:51:24 | haproxy/2-write-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:51:24 | haproxy/2-write-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:51:24 | haproxy/2-write-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:51:24 | haproxy/2-write-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:51:24 | haproxy/2-write-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:51:24 | haproxy/2-write-data | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:51:24 | haproxy/2-write-data | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:51:24 | haproxy/2-write-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:51:24 | haproxy/2-write-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:51:24 | haproxy/2-write-data | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:51:24 | haproxy/2-write-data | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:51:24 | haproxy/2-write-data | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:51:24 | haproxy/2-write-data | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:51:24 | haproxy/2-write-data | ++++ which gdate logger.go:42: 14:51:24 | haproxy/2-write-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-424/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:51:24 | haproxy/2-write-data | ++++ which date logger.go:42: 14:51:24 | haproxy/2-write-data | +++ date=/usr/bin/date logger.go:42: 14:51:24 | haproxy/2-write-data | +++ command -v oc logger.go:42: 14:51:24 | haproxy/2-write-data | +++ oc get projects logger.go:42: 14:51:30 | haproxy/2-write-data | error: the server doesn't have a resource type "projects" logger.go:42: 14:51:30 | haproxy/2-write-data | +++ kubectl get nodes logger.go:42: 14:51:30 | haproxy/2-write-data | +++ grep '^minikube' logger.go:42: 14:51:31 | haproxy/2-write-data | ++++ pwd logger.go:42: 14:51:31 | haproxy/2-write-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/tests/haproxy logger.go:42: 14:51:31 | haproxy/2-write-data | ++ test_name=haproxy logger.go:42: 14:51:31 | haproxy/2-write-data | +++ get_cluster_name logger.go:42: 14:51:31 | haproxy/2-write-data | +++ kubectl -n kuttl-test-viable-phoenix get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:51:32 | haproxy/2-write-data | ++ get_haproxy_svc haproxy logger.go:42: 14:51:32 | haproxy/2-write-data | ++ local cluster=haproxy logger.go:42: 14:51:32 | haproxy/2-write-data | ++ echo haproxy-haproxy logger.go:42: 14:51:32 | haproxy/2-write-data | + run_mysql 'CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' '-h haproxy-haproxy -uroot -proot_password' logger.go:42: 14:51:32 | haproxy/2-write-data | + local 'command=CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' logger.go:42: 14:51:32 | haproxy/2-write-data | + local 'uri=-h haproxy-haproxy -uroot -proot_password' logger.go:42: 14:51:32 | haproxy/2-write-data | + local pod= logger.go:42: 14:51:32 | haproxy/2-write-data | ++ get_client_pod logger.go:42: 14:51:32 | haproxy/2-write-data | ++ kubectl -n kuttl-test-viable-phoenix get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:51:32 | haproxy/2-write-data | + client_pod=mysql-client logger.go:42: 14:51:32 | haproxy/2-write-data | + wait_pod mysql-client logger.go:42: 14:51:32 | haproxy/2-write-data | + local pod=mysql-client logger.go:42: 14:51:32 | haproxy/2-write-data | + set +o xtrace logger.go:42: 14:51:33 | haproxy/2-write-data | mysql-clienttrue logger.go:42: 14:51:33 | haproxy/2-write-data | + kubectl -n kuttl-test-viable-phoenix exec mysql-client -- bash -c 'printf '\''%s\n'\'' "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" | mysql -sN -h haproxy-haproxy -uroot -proot_password' logger.go:42: 14:51:33 | haproxy/2-write-data | + sed -e 's/mysql: //' logger.go:42: 14:51:33 | haproxy/2-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:51:35 | haproxy/2-write-data | + : logger.go:42: 14:51:35 | haproxy/2-write-data | +++ get_cluster_name logger.go:42: 14:51:35 | haproxy/2-write-data | +++ kubectl -n kuttl-test-viable-phoenix get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:51:35 | haproxy/2-write-data | ++ get_haproxy_svc haproxy logger.go:42: 14:51:35 | haproxy/2-write-data | ++ local cluster=haproxy logger.go:42: 14:51:35 | haproxy/2-write-data | ++ echo haproxy-haproxy logger.go:42: 14:51:35 | haproxy/2-write-data | + run_mysql 'INSERT myDB.myTable (id) VALUES (100500)' '-h haproxy-haproxy -uroot -proot_password' logger.go:42: 14:51:35 | haproxy/2-write-data | + local 'command=INSERT myDB.myTable (id) VALUES (100500)' logger.go:42: 14:51:35 | haproxy/2-write-data | + local 'uri=-h haproxy-haproxy -uroot -proot_password' logger.go:42: 14:51:35 | haproxy/2-write-data | + local pod= logger.go:42: 14:51:35 | haproxy/2-write-data | ++ get_client_pod logger.go:42: 14:51:35 | haproxy/2-write-data | ++ kubectl -n kuttl-test-viable-phoenix get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:51:36 | haproxy/2-write-data | + client_pod=mysql-client logger.go:42: 14:51:36 | haproxy/2-write-data | + wait_pod mysql-client logger.go:42: 14:51:36 | haproxy/2-write-data | + local pod=mysql-client logger.go:42: 14:51:36 | haproxy/2-write-data | + set +o xtrace logger.go:42: 14:51:37 | haproxy/2-write-data | mysql-clienttrue logger.go:42: 14:51:37 | haproxy/2-write-data | + kubectl -n kuttl-test-viable-phoenix exec mysql-client -- bash -c 'printf '\''%s\n'\'' "INSERT myDB.myTable (id) VALUES (100500)" | mysql -sN -h haproxy-haproxy -uroot -proot_password' logger.go:42: 14:51:37 | haproxy/2-write-data | + sed -e 's/mysql: //' logger.go:42: 14:51:37 | haproxy/2-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:51:38 | haproxy/2-write-data | + : logger.go:42: 14:51:39 | haproxy/2-write-data | test step completed 2-write-data logger.go:42: 14:51:39 | haproxy/3-read-from-primary | starting test step 3-read-from-primary logger.go:42: 14:51:39 | haproxy/3-read-from-primary | running command: [sh -c set -o errexit set -o xtrace source ../../functions data=$(run_mysql "SELECT * FROM myDB.myTable" "-h $(get_haproxy_svc $(get_cluster_name)) -P3306 -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 03-read-from-primary --from-literal=data="${data}"] logger.go:42: 14:51:39 | haproxy/3-read-from-primary | + source ../../functions logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ realpath ../../.. logger.go:42: 14:51:39 | haproxy/3-read-from-primary | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:51:39 | haproxy/3-read-from-primary | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/vars.sh logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:51:39 | haproxy/3-read-from-primary | ++++ mktemp -d logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ export TEMP_DIR=/tmp/tmp.GTvdXWWEB8 logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ TEMP_DIR=/tmp/tmp.GTvdXWWEB8 logger.go:42: 14:51:39 | haproxy/3-read-from-primary | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ export GIT_BRANCH=PR-424 logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ GIT_BRANCH=PR-424 logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ export VERSION=PR-424-70568ae logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ VERSION=PR-424-70568ae logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:51:39 | haproxy/3-read-from-primary | ++++ which gdate logger.go:42: 14:51:39 | haproxy/3-read-from-primary | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-424/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:51:39 | haproxy/3-read-from-primary | ++++ which date logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ date=/usr/bin/date logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ command -v oc logger.go:42: 14:51:39 | haproxy/3-read-from-primary | +++ oc get projects logger.go:42: 14:51:45 | haproxy/3-read-from-primary | error: the server doesn't have a resource type "projects" logger.go:42: 14:51:45 | haproxy/3-read-from-primary | +++ kubectl get nodes logger.go:42: 14:51:45 | haproxy/3-read-from-primary | +++ grep '^minikube' logger.go:42: 14:51:46 | haproxy/3-read-from-primary | ++++ pwd logger.go:42: 14:51:46 | haproxy/3-read-from-primary | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/tests/haproxy logger.go:42: 14:51:46 | haproxy/3-read-from-primary | ++ test_name=haproxy logger.go:42: 14:51:46 | haproxy/3-read-from-primary | ++++ get_cluster_name logger.go:42: 14:51:46 | haproxy/3-read-from-primary | ++++ kubectl -n kuttl-test-viable-phoenix get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:51:47 | haproxy/3-read-from-primary | +++ get_haproxy_svc haproxy logger.go:42: 14:51:47 | haproxy/3-read-from-primary | +++ local cluster=haproxy logger.go:42: 14:51:47 | haproxy/3-read-from-primary | +++ echo haproxy-haproxy logger.go:42: 14:51:47 | haproxy/3-read-from-primary | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h haproxy-haproxy -P3306 -uroot -proot_password' logger.go:42: 14:51:47 | haproxy/3-read-from-primary | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:51:47 | haproxy/3-read-from-primary | ++ local 'uri=-h haproxy-haproxy -P3306 -uroot -proot_password' logger.go:42: 14:51:47 | haproxy/3-read-from-primary | ++ local pod= logger.go:42: 14:51:47 | haproxy/3-read-from-primary | +++ get_client_pod logger.go:42: 14:51:47 | haproxy/3-read-from-primary | +++ kubectl -n kuttl-test-viable-phoenix get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:51:47 | haproxy/3-read-from-primary | ++ client_pod=mysql-client logger.go:42: 14:51:47 | haproxy/3-read-from-primary | ++ wait_pod mysql-client logger.go:42: 14:51:47 | haproxy/3-read-from-primary | ++ local pod=mysql-client logger.go:42: 14:51:47 | haproxy/3-read-from-primary | ++ set +o xtrace logger.go:42: 14:51:48 | haproxy/3-read-from-primary | mysql-clienttrue logger.go:42: 14:51:48 | haproxy/3-read-from-primary | ++ kubectl -n kuttl-test-viable-phoenix exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h haproxy-haproxy -P3306 -uroot -proot_password' logger.go:42: 14:51:48 | haproxy/3-read-from-primary | ++ sed -e 's/mysql: //' logger.go:42: 14:51:48 | haproxy/3-read-from-primary | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:51:50 | haproxy/3-read-from-primary | + data=100500 logger.go:42: 14:51:50 | haproxy/3-read-from-primary | + kubectl create configmap -n kuttl-test-viable-phoenix 03-read-from-primary --from-literal=data=100500 logger.go:42: 14:51:50 | haproxy/3-read-from-primary | configmap/03-read-from-primary created logger.go:42: 14:51:52 | haproxy/3-read-from-primary | test step completed 3-read-from-primary logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | starting test step 4-read-from-replicas logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | running command: [sh -c set -o errexit set -o xtrace source ../../functions data=$(run_mysql "SELECT * FROM myDB.myTable" "-h $(get_haproxy_svc $(get_cluster_name)) -P3307 -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 04-read-from-replicas --from-literal=${test_name}-haproxy-replicas=${data}] logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | + source ../../functions logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ realpath ../../.. logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/vars.sh logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | ++++ mktemp -d logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ export TEMP_DIR=/tmp/tmp.WYHR4qGrOf logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ TEMP_DIR=/tmp/tmp.WYHR4qGrOf logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ export GIT_BRANCH=PR-424 logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ GIT_BRANCH=PR-424 logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ export VERSION=PR-424-70568ae logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ VERSION=PR-424-70568ae logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | ++++ which gdate logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-424/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | ++++ which date logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ date=/usr/bin/date logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ command -v oc logger.go:42: 14:51:52 | haproxy/4-read-from-replicas | +++ oc get projects logger.go:42: 14:51:57 | haproxy/4-read-from-replicas | error: the server doesn't have a resource type "projects" logger.go:42: 14:51:57 | haproxy/4-read-from-replicas | +++ kubectl get nodes logger.go:42: 14:51:57 | haproxy/4-read-from-replicas | +++ grep '^minikube' logger.go:42: 14:51:58 | haproxy/4-read-from-replicas | ++++ pwd logger.go:42: 14:51:58 | haproxy/4-read-from-replicas | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/tests/haproxy logger.go:42: 14:51:58 | haproxy/4-read-from-replicas | ++ test_name=haproxy logger.go:42: 14:51:58 | haproxy/4-read-from-replicas | ++++ get_cluster_name logger.go:42: 14:51:58 | haproxy/4-read-from-replicas | ++++ kubectl -n kuttl-test-viable-phoenix get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:51:59 | haproxy/4-read-from-replicas | +++ get_haproxy_svc haproxy logger.go:42: 14:51:59 | haproxy/4-read-from-replicas | +++ local cluster=haproxy logger.go:42: 14:51:59 | haproxy/4-read-from-replicas | +++ echo haproxy-haproxy logger.go:42: 14:51:59 | haproxy/4-read-from-replicas | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h haproxy-haproxy -P3307 -uroot -proot_password' logger.go:42: 14:51:59 | haproxy/4-read-from-replicas | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:51:59 | haproxy/4-read-from-replicas | ++ local 'uri=-h haproxy-haproxy -P3307 -uroot -proot_password' logger.go:42: 14:51:59 | haproxy/4-read-from-replicas | ++ local pod= logger.go:42: 14:51:59 | haproxy/4-read-from-replicas | +++ get_client_pod logger.go:42: 14:51:59 | haproxy/4-read-from-replicas | +++ kubectl -n kuttl-test-viable-phoenix get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:52:00 | haproxy/4-read-from-replicas | ++ client_pod=mysql-client logger.go:42: 14:52:00 | haproxy/4-read-from-replicas | ++ wait_pod mysql-client logger.go:42: 14:52:00 | haproxy/4-read-from-replicas | ++ local pod=mysql-client logger.go:42: 14:52:00 | haproxy/4-read-from-replicas | ++ set +o xtrace logger.go:42: 14:52:00 | haproxy/4-read-from-replicas | mysql-clienttrue logger.go:42: 14:52:00 | haproxy/4-read-from-replicas | ++ kubectl -n kuttl-test-viable-phoenix exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h haproxy-haproxy -P3307 -uroot -proot_password' logger.go:42: 14:52:00 | haproxy/4-read-from-replicas | ++ sed -e 's/mysql: //' logger.go:42: 14:52:00 | haproxy/4-read-from-replicas | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:52:02 | haproxy/4-read-from-replicas | + data=100500 logger.go:42: 14:52:02 | haproxy/4-read-from-replicas | + kubectl create configmap -n kuttl-test-viable-phoenix 04-read-from-replicas --from-literal=haproxy-haproxy-replicas=100500 logger.go:42: 14:52:03 | haproxy/4-read-from-replicas | configmap/04-read-from-replicas created logger.go:42: 14:52:04 | haproxy/4-read-from-replicas | test step completed 4-read-from-replicas logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | starting test step 5-check-pods-have-same-primary logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | running command: [sh -c set -o errexit set -o xtrace source ../../functions data=() for i in $(seq 0 2); do data+=("$(get_primary_from_haproxy ${test_name}-haproxy-$i)") done if [ "${data[0]}" != "${data[1]}" -o "${data[1]}" != "${data[2]}" ]; then echo "Not all haproxy pods point to same primary: 0: ${data[0]} 1: ${data[1]} 2: ${data[2]}" exit 1 fi] logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | + source ../../functions logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ realpath ../../.. logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/vars.sh logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | ++++ mktemp -d logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ export TEMP_DIR=/tmp/tmp.8MKXfNm6qo logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ TEMP_DIR=/tmp/tmp.8MKXfNm6qo logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ export GIT_BRANCH=PR-424 logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ GIT_BRANCH=PR-424 logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ export VERSION=PR-424-70568ae logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ VERSION=PR-424-70568ae logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | ++++ which gdate logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-424/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | ++++ which date logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ date=/usr/bin/date logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ command -v oc logger.go:42: 14:52:04 | haproxy/5-check-pods-have-same-primary | +++ oc get projects logger.go:42: 14:52:10 | haproxy/5-check-pods-have-same-primary | error: the server doesn't have a resource type "projects" logger.go:42: 14:52:10 | haproxy/5-check-pods-have-same-primary | +++ kubectl get nodes logger.go:42: 14:52:10 | haproxy/5-check-pods-have-same-primary | +++ grep '^minikube' logger.go:42: 14:52:10 | haproxy/5-check-pods-have-same-primary | ++++ pwd logger.go:42: 14:52:10 | haproxy/5-check-pods-have-same-primary | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/tests/haproxy logger.go:42: 14:52:10 | haproxy/5-check-pods-have-same-primary | ++ test_name=haproxy logger.go:42: 14:52:10 | haproxy/5-check-pods-have-same-primary | + data=() logger.go:42: 14:52:10 | haproxy/5-check-pods-have-same-primary | ++ seq 0 2 logger.go:42: 14:52:10 | haproxy/5-check-pods-have-same-primary | + for i in '$(seq 0 2)' logger.go:42: 14:52:10 | haproxy/5-check-pods-have-same-primary | + data+=("$(get_primary_from_haproxy ${test_name}-haproxy-$i)") logger.go:42: 14:52:10 | haproxy/5-check-pods-have-same-primary | ++ get_primary_from_haproxy haproxy-haproxy-0 logger.go:42: 14:52:10 | haproxy/5-check-pods-have-same-primary | ++ local haproxy_pod=haproxy-haproxy-0 logger.go:42: 14:52:10 | haproxy/5-check-pods-have-same-primary | +++ kubectl -n kuttl-test-viable-phoenix get pods haproxy-haproxy-0 -o 'jsonpath={.status.podIP}' logger.go:42: 14:52:11 | haproxy/5-check-pods-have-same-primary | ++ local haproxy_pod_ip=10.6.90.27 logger.go:42: 14:52:11 | haproxy/5-check-pods-have-same-primary | ++ run_mysql 'SHOW VARIABLES LIKE '\''%hostname%'\'';' '-h 10.6.90.27 -P3306 -uroot -proot_password' logger.go:42: 14:52:11 | haproxy/5-check-pods-have-same-primary | ++ local 'command=SHOW VARIABLES LIKE '\''%hostname%'\'';' logger.go:42: 14:52:11 | haproxy/5-check-pods-have-same-primary | ++ awk '{print $2}' logger.go:42: 14:52:11 | haproxy/5-check-pods-have-same-primary | ++ local 'uri=-h 10.6.90.27 -P3306 -uroot -proot_password' logger.go:42: 14:52:11 | haproxy/5-check-pods-have-same-primary | ++ local pod= logger.go:42: 14:52:11 | haproxy/5-check-pods-have-same-primary | +++ get_client_pod logger.go:42: 14:52:11 | haproxy/5-check-pods-have-same-primary | +++ kubectl -n kuttl-test-viable-phoenix get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:52:12 | haproxy/5-check-pods-have-same-primary | ++ client_pod=mysql-client logger.go:42: 14:52:12 | haproxy/5-check-pods-have-same-primary | ++ wait_pod mysql-client logger.go:42: 14:52:12 | haproxy/5-check-pods-have-same-primary | ++ local pod=mysql-client logger.go:42: 14:52:12 | haproxy/5-check-pods-have-same-primary | ++ set +o xtrace logger.go:42: 14:52:13 | haproxy/5-check-pods-have-same-primary | mysql-clienttrue logger.go:42: 14:52:13 | haproxy/5-check-pods-have-same-primary | ++ sed -e 's/mysql: //' logger.go:42: 14:52:13 | haproxy/5-check-pods-have-same-primary | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:52:13 | haproxy/5-check-pods-have-same-primary | ++ kubectl -n kuttl-test-viable-phoenix exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SHOW VARIABLES LIKE '\''%hostname%'\'';" | mysql -sN -h 10.6.90.27 -P3306 -uroot -proot_password' logger.go:42: 14:52:14 | haproxy/5-check-pods-have-same-primary | + for i in '$(seq 0 2)' logger.go:42: 14:52:14 | haproxy/5-check-pods-have-same-primary | + data+=("$(get_primary_from_haproxy ${test_name}-haproxy-$i)") logger.go:42: 14:52:14 | haproxy/5-check-pods-have-same-primary | ++ get_primary_from_haproxy haproxy-haproxy-1 logger.go:42: 14:52:14 | haproxy/5-check-pods-have-same-primary | ++ local haproxy_pod=haproxy-haproxy-1 logger.go:42: 14:52:14 | haproxy/5-check-pods-have-same-primary | +++ kubectl -n kuttl-test-viable-phoenix get pods haproxy-haproxy-1 -o 'jsonpath={.status.podIP}' logger.go:42: 14:52:15 | haproxy/5-check-pods-have-same-primary | ++ local haproxy_pod_ip=10.6.88.21 logger.go:42: 14:52:15 | haproxy/5-check-pods-have-same-primary | ++ run_mysql 'SHOW VARIABLES LIKE '\''%hostname%'\'';' '-h 10.6.88.21 -P3306 -uroot -proot_password' logger.go:42: 14:52:15 | haproxy/5-check-pods-have-same-primary | ++ awk '{print $2}' logger.go:42: 14:52:15 | haproxy/5-check-pods-have-same-primary | ++ local 'command=SHOW VARIABLES LIKE '\''%hostname%'\'';' logger.go:42: 14:52:15 | haproxy/5-check-pods-have-same-primary | ++ local 'uri=-h 10.6.88.21 -P3306 -uroot -proot_password' logger.go:42: 14:52:15 | haproxy/5-check-pods-have-same-primary | ++ local pod= logger.go:42: 14:52:15 | haproxy/5-check-pods-have-same-primary | +++ get_client_pod logger.go:42: 14:52:15 | haproxy/5-check-pods-have-same-primary | +++ kubectl -n kuttl-test-viable-phoenix get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:52:16 | haproxy/5-check-pods-have-same-primary | ++ client_pod=mysql-client logger.go:42: 14:52:16 | haproxy/5-check-pods-have-same-primary | ++ wait_pod mysql-client logger.go:42: 14:52:16 | haproxy/5-check-pods-have-same-primary | ++ local pod=mysql-client logger.go:42: 14:52:16 | haproxy/5-check-pods-have-same-primary | ++ set +o xtrace logger.go:42: 14:52:17 | haproxy/5-check-pods-have-same-primary | mysql-clienttrue logger.go:42: 14:52:17 | haproxy/5-check-pods-have-same-primary | ++ kubectl -n kuttl-test-viable-phoenix exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SHOW VARIABLES LIKE '\''%hostname%'\'';" | mysql -sN -h 10.6.88.21 -P3306 -uroot -proot_password' logger.go:42: 14:52:17 | haproxy/5-check-pods-have-same-primary | ++ sed -e 's/mysql: //' logger.go:42: 14:52:17 | haproxy/5-check-pods-have-same-primary | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:52:18 | haproxy/5-check-pods-have-same-primary | + for i in '$(seq 0 2)' logger.go:42: 14:52:18 | haproxy/5-check-pods-have-same-primary | + data+=("$(get_primary_from_haproxy ${test_name}-haproxy-$i)") logger.go:42: 14:52:18 | haproxy/5-check-pods-have-same-primary | ++ get_primary_from_haproxy haproxy-haproxy-2 logger.go:42: 14:52:18 | haproxy/5-check-pods-have-same-primary | ++ local haproxy_pod=haproxy-haproxy-2 logger.go:42: 14:52:18 | haproxy/5-check-pods-have-same-primary | +++ kubectl -n kuttl-test-viable-phoenix get pods haproxy-haproxy-2 -o 'jsonpath={.status.podIP}' logger.go:42: 14:52:19 | haproxy/5-check-pods-have-same-primary | ++ local haproxy_pod_ip=10.6.89.22 logger.go:42: 14:52:19 | haproxy/5-check-pods-have-same-primary | ++ run_mysql 'SHOW VARIABLES LIKE '\''%hostname%'\'';' '-h 10.6.89.22 -P3306 -uroot -proot_password' logger.go:42: 14:52:19 | haproxy/5-check-pods-have-same-primary | ++ awk '{print $2}' logger.go:42: 14:52:19 | haproxy/5-check-pods-have-same-primary | ++ local 'command=SHOW VARIABLES LIKE '\''%hostname%'\'';' logger.go:42: 14:52:19 | haproxy/5-check-pods-have-same-primary | ++ local 'uri=-h 10.6.89.22 -P3306 -uroot -proot_password' logger.go:42: 14:52:19 | haproxy/5-check-pods-have-same-primary | ++ local pod= logger.go:42: 14:52:19 | haproxy/5-check-pods-have-same-primary | +++ get_client_pod logger.go:42: 14:52:19 | haproxy/5-check-pods-have-same-primary | +++ kubectl -n kuttl-test-viable-phoenix get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:52:20 | haproxy/5-check-pods-have-same-primary | ++ client_pod=mysql-client logger.go:42: 14:52:20 | haproxy/5-check-pods-have-same-primary | ++ wait_pod mysql-client logger.go:42: 14:52:20 | haproxy/5-check-pods-have-same-primary | ++ local pod=mysql-client logger.go:42: 14:52:20 | haproxy/5-check-pods-have-same-primary | ++ set +o xtrace logger.go:42: 14:52:20 | haproxy/5-check-pods-have-same-primary | mysql-clienttrue logger.go:42: 14:52:20 | haproxy/5-check-pods-have-same-primary | ++ sed -e 's/mysql: //' logger.go:42: 14:52:20 | haproxy/5-check-pods-have-same-primary | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:52:20 | haproxy/5-check-pods-have-same-primary | ++ kubectl -n kuttl-test-viable-phoenix exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SHOW VARIABLES LIKE '\''%hostname%'\'';" | mysql -sN -h 10.6.89.22 -P3306 -uroot -proot_password' logger.go:42: 14:52:22 | haproxy/5-check-pods-have-same-primary | + '[' haproxy-mysql-0 '!=' haproxy-mysql-0 -o haproxy-mysql-0 '!=' haproxy-mysql-0 ']' logger.go:42: 14:52:23 | haproxy/5-check-pods-have-same-primary | test step completed 5-check-pods-have-same-primary logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | starting test step 6-check-label-haproxy-primary logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | running command: [sh -c set -o errexit set -o xtrace source ../../functions primary_pod_from_label="$(get_primary_from_label)" primary_pod_from_haproxy="$(get_primary_from_haproxy ${test_name}-haproxy-0)" if [ "${primary_pod_from_label}" != "${primary_pod_from_haproxy}" ]; then echo "Primary in k8s label (${primary_pod_from_label}) is not set to same pod as in haproxy (${primary_pod_from_haproxy})!" exit 1 fi] logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | + source ../../functions logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ realpath ../../.. logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/vars.sh logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | ++++ mktemp -d logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ export TEMP_DIR=/tmp/tmp.PIgLVfhhD3 logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ TEMP_DIR=/tmp/tmp.PIgLVfhhD3 logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ export GIT_BRANCH=PR-424 logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ GIT_BRANCH=PR-424 logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ export VERSION=PR-424-70568ae logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ VERSION=PR-424-70568ae logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | ++++ which gdate logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-424/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | ++++ which date logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ date=/usr/bin/date logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ command -v oc logger.go:42: 14:52:23 | haproxy/6-check-label-haproxy-primary | +++ oc get projects logger.go:42: 14:52:29 | haproxy/6-check-label-haproxy-primary | error: the server doesn't have a resource type "projects" logger.go:42: 14:52:29 | haproxy/6-check-label-haproxy-primary | +++ grep '^minikube' logger.go:42: 14:52:29 | haproxy/6-check-label-haproxy-primary | +++ kubectl get nodes logger.go:42: 14:52:29 | haproxy/6-check-label-haproxy-primary | ++++ pwd logger.go:42: 14:52:29 | haproxy/6-check-label-haproxy-primary | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/tests/haproxy logger.go:42: 14:52:29 | haproxy/6-check-label-haproxy-primary | ++ test_name=haproxy logger.go:42: 14:52:29 | haproxy/6-check-label-haproxy-primary | ++ get_primary_from_label logger.go:42: 14:52:29 | haproxy/6-check-label-haproxy-primary | ++ kubectl -n kuttl-test-viable-phoenix get pods -l mysql.percona.com/primary=true '-ojsonpath={.items[0].metadata.name}' logger.go:42: 14:52:30 | haproxy/6-check-label-haproxy-primary | + primary_pod_from_label=haproxy-mysql-0 logger.go:42: 14:52:30 | haproxy/6-check-label-haproxy-primary | ++ get_primary_from_haproxy haproxy-haproxy-0 logger.go:42: 14:52:30 | haproxy/6-check-label-haproxy-primary | ++ local haproxy_pod=haproxy-haproxy-0 logger.go:42: 14:52:30 | haproxy/6-check-label-haproxy-primary | +++ kubectl -n kuttl-test-viable-phoenix get pods haproxy-haproxy-0 -o 'jsonpath={.status.podIP}' logger.go:42: 14:52:31 | haproxy/6-check-label-haproxy-primary | ++ local haproxy_pod_ip=10.6.90.27 logger.go:42: 14:52:31 | haproxy/6-check-label-haproxy-primary | ++ run_mysql 'SHOW VARIABLES LIKE '\''%hostname%'\'';' '-h 10.6.90.27 -P3306 -uroot -proot_password' logger.go:42: 14:52:31 | haproxy/6-check-label-haproxy-primary | ++ awk '{print $2}' logger.go:42: 14:52:31 | haproxy/6-check-label-haproxy-primary | ++ local 'command=SHOW VARIABLES LIKE '\''%hostname%'\'';' logger.go:42: 14:52:31 | haproxy/6-check-label-haproxy-primary | ++ local 'uri=-h 10.6.90.27 -P3306 -uroot -proot_password' logger.go:42: 14:52:31 | haproxy/6-check-label-haproxy-primary | ++ local pod= logger.go:42: 14:52:31 | haproxy/6-check-label-haproxy-primary | +++ get_client_pod logger.go:42: 14:52:31 | haproxy/6-check-label-haproxy-primary | +++ kubectl -n kuttl-test-viable-phoenix get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:52:31 | haproxy/6-check-label-haproxy-primary | ++ client_pod=mysql-client logger.go:42: 14:52:31 | haproxy/6-check-label-haproxy-primary | ++ wait_pod mysql-client logger.go:42: 14:52:31 | haproxy/6-check-label-haproxy-primary | ++ local pod=mysql-client logger.go:42: 14:52:31 | haproxy/6-check-label-haproxy-primary | ++ set +o xtrace logger.go:42: 14:52:32 | haproxy/6-check-label-haproxy-primary | mysql-clienttrue logger.go:42: 14:52:32 | haproxy/6-check-label-haproxy-primary | ++ sed -e 's/mysql: //' logger.go:42: 14:52:32 | haproxy/6-check-label-haproxy-primary | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:52:32 | haproxy/6-check-label-haproxy-primary | ++ kubectl -n kuttl-test-viable-phoenix exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SHOW VARIABLES LIKE '\''%hostname%'\'';" | mysql -sN -h 10.6.90.27 -P3306 -uroot -proot_password' logger.go:42: 14:52:34 | haproxy/6-check-label-haproxy-primary | + primary_pod_from_haproxy=haproxy-mysql-0 logger.go:42: 14:52:34 | haproxy/6-check-label-haproxy-primary | + '[' haproxy-mysql-0 '!=' haproxy-mysql-0 ']' logger.go:42: 14:52:35 | haproxy/6-check-label-haproxy-primary | test step completed 6-check-label-haproxy-primary logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | starting test step 7-check-primary-failover logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | running command: [sh -c set -o errexit set -o xtrace source ../../functions primary_pod_from_label="$(get_primary_from_label)" kubectl -n "${NAMESPACE}" delete pod ${primary_pod_from_label} sleep 3 data=() for i in $(seq 0 2); do data+=("$(get_primary_from_haproxy ${test_name}-haproxy-$i)") done if [ "${data[0]}" != "${data[1]}" -o "${data[1]}" != "${data[2]}" ]; then echo "Not all haproxy pods point to same primary: 0: ${data[0]} 1: ${data[1]} 2: ${data[2]}" exit 1 fi primary_pod_from_label="$(get_primary_from_label)" primary_pod_from_haproxy="$(get_primary_from_haproxy ${test_name}-haproxy-0)" if [ "${primary_pod_from_label}" != "${primary_pod_from_haproxy}" ]; then echo "Primary in k8s label (${primary_pod_from_label}) is not set to same pod as in haproxy (${primary_pod_from_haproxy})!" exit 1 fi] logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | + source ../../functions logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ realpath ../../.. logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/vars.sh logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | ++++ mktemp -d logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ export TEMP_DIR=/tmp/tmp.oBuSSrc3YM logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ TEMP_DIR=/tmp/tmp.oBuSSrc3YM logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ export GIT_BRANCH=PR-424 logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ GIT_BRANCH=PR-424 logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ export VERSION=PR-424-70568ae logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ VERSION=PR-424-70568ae logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | ++++ which gdate logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-424/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | ++++ which date logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ date=/usr/bin/date logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ command -v oc logger.go:42: 14:52:35 | haproxy/7-check-primary-failover | +++ oc get projects logger.go:42: 14:52:41 | haproxy/7-check-primary-failover | error: the server doesn't have a resource type "projects" logger.go:42: 14:52:41 | haproxy/7-check-primary-failover | +++ grep '^minikube' logger.go:42: 14:52:41 | haproxy/7-check-primary-failover | +++ kubectl get nodes logger.go:42: 14:52:41 | haproxy/7-check-primary-failover | ++++ pwd logger.go:42: 14:52:41 | haproxy/7-check-primary-failover | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/tests/haproxy logger.go:42: 14:52:41 | haproxy/7-check-primary-failover | ++ test_name=haproxy logger.go:42: 14:52:41 | haproxy/7-check-primary-failover | ++ get_primary_from_label logger.go:42: 14:52:41 | haproxy/7-check-primary-failover | ++ kubectl -n kuttl-test-viable-phoenix get pods -l mysql.percona.com/primary=true '-ojsonpath={.items[0].metadata.name}' logger.go:42: 14:52:42 | haproxy/7-check-primary-failover | + primary_pod_from_label=haproxy-mysql-0 logger.go:42: 14:52:42 | haproxy/7-check-primary-failover | + kubectl -n kuttl-test-viable-phoenix delete pod haproxy-mysql-0 logger.go:42: 14:52:43 | haproxy/7-check-primary-failover | pod "haproxy-mysql-0" deleted logger.go:42: 14:53:03 | haproxy/7-check-primary-failover | + sleep 3 logger.go:42: 14:53:06 | haproxy/7-check-primary-failover | + data=() logger.go:42: 14:53:06 | haproxy/7-check-primary-failover | ++ seq 0 2 logger.go:42: 14:53:06 | haproxy/7-check-primary-failover | + for i in '$(seq 0 2)' logger.go:42: 14:53:06 | haproxy/7-check-primary-failover | + data+=("$(get_primary_from_haproxy ${test_name}-haproxy-$i)") logger.go:42: 14:53:06 | haproxy/7-check-primary-failover | ++ get_primary_from_haproxy haproxy-haproxy-0 logger.go:42: 14:53:06 | haproxy/7-check-primary-failover | ++ local haproxy_pod=haproxy-haproxy-0 logger.go:42: 14:53:06 | haproxy/7-check-primary-failover | +++ kubectl -n kuttl-test-viable-phoenix get pods haproxy-haproxy-0 -o 'jsonpath={.status.podIP}' logger.go:42: 14:53:07 | haproxy/7-check-primary-failover | ++ local haproxy_pod_ip=10.6.90.27 logger.go:42: 14:53:07 | haproxy/7-check-primary-failover | ++ run_mysql 'SHOW VARIABLES LIKE '\''%hostname%'\'';' '-h 10.6.90.27 -P3306 -uroot -proot_password' logger.go:42: 14:53:07 | haproxy/7-check-primary-failover | ++ awk '{print $2}' logger.go:42: 14:53:07 | haproxy/7-check-primary-failover | ++ local 'command=SHOW VARIABLES LIKE '\''%hostname%'\'';' logger.go:42: 14:53:07 | haproxy/7-check-primary-failover | ++ local 'uri=-h 10.6.90.27 -P3306 -uroot -proot_password' logger.go:42: 14:53:07 | haproxy/7-check-primary-failover | ++ local pod= logger.go:42: 14:53:07 | haproxy/7-check-primary-failover | +++ get_client_pod logger.go:42: 14:53:07 | haproxy/7-check-primary-failover | +++ kubectl -n kuttl-test-viable-phoenix get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:53:08 | haproxy/7-check-primary-failover | ++ client_pod=mysql-client logger.go:42: 14:53:08 | haproxy/7-check-primary-failover | ++ wait_pod mysql-client logger.go:42: 14:53:08 | haproxy/7-check-primary-failover | ++ local pod=mysql-client logger.go:42: 14:53:08 | haproxy/7-check-primary-failover | ++ set +o xtrace logger.go:42: 14:53:09 | haproxy/7-check-primary-failover | mysql-clienttrue logger.go:42: 14:53:09 | haproxy/7-check-primary-failover | ++ kubectl -n kuttl-test-viable-phoenix exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SHOW VARIABLES LIKE '\''%hostname%'\'';" | mysql -sN -h 10.6.90.27 -P3306 -uroot -proot_password' logger.go:42: 14:53:09 | haproxy/7-check-primary-failover | ++ sed -e 's/mysql: //' logger.go:42: 14:53:09 | haproxy/7-check-primary-failover | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:53:10 | haproxy/7-check-primary-failover | + for i in '$(seq 0 2)' logger.go:42: 14:53:10 | haproxy/7-check-primary-failover | + data+=("$(get_primary_from_haproxy ${test_name}-haproxy-$i)") logger.go:42: 14:53:10 | haproxy/7-check-primary-failover | ++ get_primary_from_haproxy haproxy-haproxy-1 logger.go:42: 14:53:10 | haproxy/7-check-primary-failover | ++ local haproxy_pod=haproxy-haproxy-1 logger.go:42: 14:53:10 | haproxy/7-check-primary-failover | +++ kubectl -n kuttl-test-viable-phoenix get pods haproxy-haproxy-1 -o 'jsonpath={.status.podIP}' logger.go:42: 14:53:11 | haproxy/7-check-primary-failover | ++ local haproxy_pod_ip=10.6.88.21 logger.go:42: 14:53:11 | haproxy/7-check-primary-failover | ++ run_mysql 'SHOW VARIABLES LIKE '\''%hostname%'\'';' '-h 10.6.88.21 -P3306 -uroot -proot_password' logger.go:42: 14:53:11 | haproxy/7-check-primary-failover | ++ awk '{print $2}' logger.go:42: 14:53:11 | haproxy/7-check-primary-failover | ++ local 'command=SHOW VARIABLES LIKE '\''%hostname%'\'';' logger.go:42: 14:53:11 | haproxy/7-check-primary-failover | ++ local 'uri=-h 10.6.88.21 -P3306 -uroot -proot_password' logger.go:42: 14:53:11 | haproxy/7-check-primary-failover | ++ local pod= logger.go:42: 14:53:11 | haproxy/7-check-primary-failover | +++ get_client_pod logger.go:42: 14:53:11 | haproxy/7-check-primary-failover | +++ kubectl -n kuttl-test-viable-phoenix get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:53:12 | haproxy/7-check-primary-failover | ++ client_pod=mysql-client logger.go:42: 14:53:12 | haproxy/7-check-primary-failover | ++ wait_pod mysql-client logger.go:42: 14:53:12 | haproxy/7-check-primary-failover | ++ local pod=mysql-client logger.go:42: 14:53:12 | haproxy/7-check-primary-failover | ++ set +o xtrace logger.go:42: 14:53:12 | haproxy/7-check-primary-failover | mysql-clienttrue logger.go:42: 14:53:12 | haproxy/7-check-primary-failover | ++ kubectl -n kuttl-test-viable-phoenix exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SHOW VARIABLES LIKE '\''%hostname%'\'';" | mysql -sN -h 10.6.88.21 -P3306 -uroot -proot_password' logger.go:42: 14:53:12 | haproxy/7-check-primary-failover | ++ sed -e 's/mysql: //' logger.go:42: 14:53:12 | haproxy/7-check-primary-failover | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:53:14 | haproxy/7-check-primary-failover | + for i in '$(seq 0 2)' logger.go:42: 14:53:14 | haproxy/7-check-primary-failover | + data+=("$(get_primary_from_haproxy ${test_name}-haproxy-$i)") logger.go:42: 14:53:14 | haproxy/7-check-primary-failover | ++ get_primary_from_haproxy haproxy-haproxy-2 logger.go:42: 14:53:14 | haproxy/7-check-primary-failover | ++ local haproxy_pod=haproxy-haproxy-2 logger.go:42: 14:53:14 | haproxy/7-check-primary-failover | +++ kubectl -n kuttl-test-viable-phoenix get pods haproxy-haproxy-2 -o 'jsonpath={.status.podIP}' logger.go:42: 14:53:15 | haproxy/7-check-primary-failover | ++ local haproxy_pod_ip=10.6.89.22 logger.go:42: 14:53:15 | haproxy/7-check-primary-failover | ++ run_mysql 'SHOW VARIABLES LIKE '\''%hostname%'\'';' '-h 10.6.89.22 -P3306 -uroot -proot_password' logger.go:42: 14:53:15 | haproxy/7-check-primary-failover | ++ awk '{print $2}' logger.go:42: 14:53:15 | haproxy/7-check-primary-failover | ++ local 'command=SHOW VARIABLES LIKE '\''%hostname%'\'';' logger.go:42: 14:53:15 | haproxy/7-check-primary-failover | ++ local 'uri=-h 10.6.89.22 -P3306 -uroot -proot_password' logger.go:42: 14:53:15 | haproxy/7-check-primary-failover | ++ local pod= logger.go:42: 14:53:15 | haproxy/7-check-primary-failover | +++ get_client_pod logger.go:42: 14:53:15 | haproxy/7-check-primary-failover | +++ kubectl -n kuttl-test-viable-phoenix get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:53:15 | haproxy/7-check-primary-failover | ++ client_pod=mysql-client logger.go:42: 14:53:15 | haproxy/7-check-primary-failover | ++ wait_pod mysql-client logger.go:42: 14:53:15 | haproxy/7-check-primary-failover | ++ local pod=mysql-client logger.go:42: 14:53:15 | haproxy/7-check-primary-failover | ++ set +o xtrace logger.go:42: 14:53:16 | haproxy/7-check-primary-failover | mysql-clienttrue logger.go:42: 14:53:16 | haproxy/7-check-primary-failover | ++ sed -e 's/mysql: //' logger.go:42: 14:53:16 | haproxy/7-check-primary-failover | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:53:16 | haproxy/7-check-primary-failover | ++ kubectl -n kuttl-test-viable-phoenix exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SHOW VARIABLES LIKE '\''%hostname%'\'';" | mysql -sN -h 10.6.89.22 -P3306 -uroot -proot_password' logger.go:42: 14:53:18 | haproxy/7-check-primary-failover | + '[' haproxy-mysql-1 '!=' haproxy-mysql-1 -o haproxy-mysql-1 '!=' haproxy-mysql-1 ']' logger.go:42: 14:53:18 | haproxy/7-check-primary-failover | ++ get_primary_from_label logger.go:42: 14:53:18 | haproxy/7-check-primary-failover | ++ kubectl -n kuttl-test-viable-phoenix get pods -l mysql.percona.com/primary=true '-ojsonpath={.items[0].metadata.name}' logger.go:42: 14:53:18 | haproxy/7-check-primary-failover | + primary_pod_from_label=haproxy-mysql-1 logger.go:42: 14:53:18 | haproxy/7-check-primary-failover | ++ get_primary_from_haproxy haproxy-haproxy-0 logger.go:42: 14:53:18 | haproxy/7-check-primary-failover | ++ local haproxy_pod=haproxy-haproxy-0 logger.go:42: 14:53:18 | haproxy/7-check-primary-failover | +++ kubectl -n kuttl-test-viable-phoenix get pods haproxy-haproxy-0 -o 'jsonpath={.status.podIP}' logger.go:42: 14:53:19 | haproxy/7-check-primary-failover | ++ local haproxy_pod_ip=10.6.90.27 logger.go:42: 14:53:19 | haproxy/7-check-primary-failover | ++ run_mysql 'SHOW VARIABLES LIKE '\''%hostname%'\'';' '-h 10.6.90.27 -P3306 -uroot -proot_password' logger.go:42: 14:53:19 | haproxy/7-check-primary-failover | ++ local 'command=SHOW VARIABLES LIKE '\''%hostname%'\'';' logger.go:42: 14:53:19 | haproxy/7-check-primary-failover | ++ local 'uri=-h 10.6.90.27 -P3306 -uroot -proot_password' logger.go:42: 14:53:19 | haproxy/7-check-primary-failover | ++ local pod= logger.go:42: 14:53:19 | haproxy/7-check-primary-failover | ++ awk '{print $2}' logger.go:42: 14:53:19 | haproxy/7-check-primary-failover | +++ get_client_pod logger.go:42: 14:53:19 | haproxy/7-check-primary-failover | +++ kubectl -n kuttl-test-viable-phoenix get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:53:20 | haproxy/7-check-primary-failover | ++ client_pod=mysql-client logger.go:42: 14:53:20 | haproxy/7-check-primary-failover | ++ wait_pod mysql-client logger.go:42: 14:53:20 | haproxy/7-check-primary-failover | ++ local pod=mysql-client logger.go:42: 14:53:20 | haproxy/7-check-primary-failover | ++ set +o xtrace logger.go:42: 14:53:20 | haproxy/7-check-primary-failover | mysql-clienttrue logger.go:42: 14:53:20 | haproxy/7-check-primary-failover | ++ kubectl -n kuttl-test-viable-phoenix exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SHOW VARIABLES LIKE '\''%hostname%'\'';" | mysql -sN -h 10.6.90.27 -P3306 -uroot -proot_password' logger.go:42: 14:53:20 | haproxy/7-check-primary-failover | ++ sed -e 's/mysql: //' logger.go:42: 14:53:20 | haproxy/7-check-primary-failover | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:53:22 | haproxy/7-check-primary-failover | + primary_pod_from_haproxy=haproxy-mysql-1 logger.go:42: 14:53:22 | haproxy/7-check-primary-failover | + '[' haproxy-mysql-1 '!=' haproxy-mysql-1 ']' logger.go:42: 14:53:23 | haproxy/7-check-primary-failover | test step completed 7-check-primary-failover logger.go:42: 14:53:23 | haproxy/8-check-password-leak | starting test step 8-check-password-leak logger.go:42: 14:53:23 | haproxy/8-check-password-leak | running command: [sh -c set -o errexit set -o xtrace source ../../functions check_passwords_leak] logger.go:42: 14:53:23 | haproxy/8-check-password-leak | + source ../../functions logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ realpath ../../.. logger.go:42: 14:53:23 | haproxy/8-check-password-leak | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:53:23 | haproxy/8-check-password-leak | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/vars.sh logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-424 logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/deploy logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/conf logger.go:42: 14:53:23 | haproxy/8-check-password-leak | ++++ mktemp -d logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ export TEMP_DIR=/tmp/tmp.cNnMs035kn logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ TEMP_DIR=/tmp/tmp.cNnMs035kn logger.go:42: 14:53:23 | haproxy/8-check-password-leak | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ export GIT_BRANCH=PR-424 logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ GIT_BRANCH=PR-424 logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ export VERSION=PR-424-70568ae logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ VERSION=PR-424-70568ae logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-424-70568ae logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 14:53:23 | haproxy/8-check-password-leak | ++++ which gdate logger.go:42: 14:53:23 | haproxy/8-check-password-leak | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-424/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:53:23 | haproxy/8-check-password-leak | ++++ which date logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ date=/usr/bin/date logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ command -v oc logger.go:42: 14:53:23 | haproxy/8-check-password-leak | +++ oc get projects logger.go:42: 14:53:29 | haproxy/8-check-password-leak | error: the server doesn't have a resource type "projects" logger.go:42: 14:53:29 | haproxy/8-check-password-leak | +++ grep '^minikube' logger.go:42: 14:53:29 | haproxy/8-check-password-leak | +++ kubectl get nodes logger.go:42: 14:53:30 | haproxy/8-check-password-leak | ++++ pwd logger.go:42: 14:53:30 | haproxy/8-check-password-leak | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-424/e2e-tests/tests/haproxy logger.go:42: 14:53:30 | haproxy/8-check-password-leak | ++ test_name=haproxy logger.go:42: 14:53:30 | haproxy/8-check-password-leak | + check_passwords_leak logger.go:42: 14:53:30 | haproxy/8-check-password-leak | ++ kubectl get secrets -o json logger.go:42: 14:53:30 | haproxy/8-check-password-leak | ++ jq -r '.items[].data | to_entries | .[] | select(.key | (endswith(".crt") or endswith(".key") or endswith(".pub") or endswith(".pem") or endswith(".p12")) | not) | .value' logger.go:42: 14:53:30 | haproxy/8-check-password-leak | + local secrets= logger.go:42: 14:53:30 | haproxy/8-check-password-leak | + local 'passwords= ' logger.go:42: 14:53:30 | haproxy/8-check-password-leak | ++ kubectl -n kuttl-test-viable-phoenix get pods -o name logger.go:42: 14:53:30 | haproxy/8-check-password-leak | ++ awk -F / '{print $2}' logger.go:42: 14:53:31 | haproxy/8-check-password-leak | + local 'pods=haproxy-haproxy-0 logger.go:42: 14:53:31 | haproxy/8-check-password-leak | haproxy-haproxy-1 logger.go:42: 14:53:31 | haproxy/8-check-password-leak | haproxy-haproxy-2 logger.go:42: 14:53:31 | haproxy/8-check-password-leak | haproxy-mysql-0 logger.go:42: 14:53:31 | haproxy/8-check-password-leak | haproxy-mysql-1 logger.go:42: 14:53:31 | haproxy/8-check-password-leak | haproxy-mysql-2 logger.go:42: 14:53:31 | haproxy/8-check-password-leak | haproxy-orc-0 logger.go:42: 14:53:31 | haproxy/8-check-password-leak | haproxy-orc-1 logger.go:42: 14:53:31 | haproxy/8-check-password-leak | haproxy-orc-2 logger.go:42: 14:53:31 | haproxy/8-check-password-leak | mysql-client logger.go:42: 14:53:31 | haproxy/8-check-password-leak | percona-server-mysql-operator-6b56d66f99-5q5ns' logger.go:42: 14:53:31 | haproxy/8-check-password-leak | + collect_logs kuttl-test-viable-phoenix logger.go:42: 14:53:31 | haproxy/8-check-password-leak | + NS=kuttl-test-viable-phoenix logger.go:42: 14:53:31 | haproxy/8-check-password-leak | + for p in '$pods' logger.go:42: 14:53:31 | haproxy/8-check-password-leak | ++ kubectl -n kuttl-test-viable-phoenix get pod haproxy-haproxy-0 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 14:53:32 | haproxy/8-check-password-leak | + local 'containers=haproxy mysql-monit' logger.go:42: 14:53:32 | haproxy/8-check-password-leak | + for c in '$containers' logger.go:42: 14:53:32 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-viable-phoenix logs haproxy-haproxy-0 -c haproxy logger.go:42: 14:53:33 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-haproxy-0-haproxy.txt logger.go:42: 14:53:33 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-haproxy-0-haproxy.txt logger.go:42: 14:53:33 | haproxy/8-check-password-leak | + for c in '$containers' logger.go:42: 14:53:33 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-viable-phoenix logs haproxy-haproxy-0 -c mysql-monit logger.go:42: 14:53:34 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-haproxy-0-mysql-monit.txt logger.go:42: 14:53:34 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-haproxy-0-mysql-monit.txt logger.go:42: 14:53:34 | haproxy/8-check-password-leak | + echo logger.go:42: 14:53:34 | haproxy/8-check-password-leak | logger.go:42: 14:53:34 | haproxy/8-check-password-leak | + for p in '$pods' logger.go:42: 14:53:34 | haproxy/8-check-password-leak | ++ kubectl -n kuttl-test-viable-phoenix get pod haproxy-haproxy-1 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 14:53:35 | haproxy/8-check-password-leak | + local 'containers=haproxy mysql-monit' logger.go:42: 14:53:35 | haproxy/8-check-password-leak | + for c in '$containers' logger.go:42: 14:53:35 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-viable-phoenix logs haproxy-haproxy-1 -c haproxy logger.go:42: 14:53:36 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-haproxy-1-haproxy.txt logger.go:42: 14:53:36 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-haproxy-1-haproxy.txt logger.go:42: 14:53:36 | haproxy/8-check-password-leak | + for c in '$containers' logger.go:42: 14:53:36 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-viable-phoenix logs haproxy-haproxy-1 -c mysql-monit logger.go:42: 14:53:37 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-haproxy-1-mysql-monit.txt logger.go:42: 14:53:37 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-haproxy-1-mysql-monit.txt logger.go:42: 14:53:37 | haproxy/8-check-password-leak | + echo logger.go:42: 14:53:37 | haproxy/8-check-password-leak | logger.go:42: 14:53:37 | haproxy/8-check-password-leak | + for p in '$pods' logger.go:42: 14:53:37 | haproxy/8-check-password-leak | ++ kubectl -n kuttl-test-viable-phoenix get pod haproxy-haproxy-2 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 14:53:38 | haproxy/8-check-password-leak | + local 'containers=haproxy mysql-monit' logger.go:42: 14:53:38 | haproxy/8-check-password-leak | + for c in '$containers' logger.go:42: 14:53:38 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-viable-phoenix logs haproxy-haproxy-2 -c haproxy logger.go:42: 14:53:39 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-haproxy-2-haproxy.txt logger.go:42: 14:53:39 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-haproxy-2-haproxy.txt logger.go:42: 14:53:39 | haproxy/8-check-password-leak | + for c in '$containers' logger.go:42: 14:53:39 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-viable-phoenix logs haproxy-haproxy-2 -c mysql-monit logger.go:42: 14:53:40 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-haproxy-2-mysql-monit.txt logger.go:42: 14:53:40 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-haproxy-2-mysql-monit.txt logger.go:42: 14:53:40 | haproxy/8-check-password-leak | + echo logger.go:42: 14:53:40 | haproxy/8-check-password-leak | logger.go:42: 14:53:40 | haproxy/8-check-password-leak | + for p in '$pods' logger.go:42: 14:53:40 | haproxy/8-check-password-leak | ++ kubectl -n kuttl-test-viable-phoenix get pod haproxy-mysql-0 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 14:53:41 | haproxy/8-check-password-leak | + local 'containers=mysql xtrabackup pt-heartbeat' logger.go:42: 14:53:41 | haproxy/8-check-password-leak | + for c in '$containers' logger.go:42: 14:53:41 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-viable-phoenix logs haproxy-mysql-0 -c mysql logger.go:42: 14:53:42 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-mysql-0-mysql.txt logger.go:42: 14:53:42 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-mysql-0-mysql.txt logger.go:42: 14:53:42 | haproxy/8-check-password-leak | + for c in '$containers' logger.go:42: 14:53:42 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-viable-phoenix logs haproxy-mysql-0 -c xtrabackup logger.go:42: 14:53:43 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-mysql-0-xtrabackup.txt logger.go:42: 14:53:43 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-mysql-0-xtrabackup.txt logger.go:42: 14:53:43 | haproxy/8-check-password-leak | + for c in '$containers' logger.go:42: 14:53:43 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-viable-phoenix logs haproxy-mysql-0 -c pt-heartbeat logger.go:42: 14:53:44 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-mysql-0-pt-heartbeat.txt logger.go:42: 14:53:44 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-mysql-0-pt-heartbeat.txt logger.go:42: 14:53:44 | haproxy/8-check-password-leak | + echo logger.go:42: 14:53:44 | haproxy/8-check-password-leak | logger.go:42: 14:53:44 | haproxy/8-check-password-leak | + for p in '$pods' logger.go:42: 14:53:44 | haproxy/8-check-password-leak | ++ kubectl -n kuttl-test-viable-phoenix get pod haproxy-mysql-1 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 14:53:45 | haproxy/8-check-password-leak | + local 'containers=mysql xtrabackup pt-heartbeat' logger.go:42: 14:53:45 | haproxy/8-check-password-leak | + for c in '$containers' logger.go:42: 14:53:45 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-viable-phoenix logs haproxy-mysql-1 -c mysql logger.go:42: 14:53:46 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-mysql-1-mysql.txt logger.go:42: 14:53:46 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-mysql-1-mysql.txt logger.go:42: 14:53:46 | haproxy/8-check-password-leak | + for c in '$containers' logger.go:42: 14:53:46 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-viable-phoenix logs haproxy-mysql-1 -c xtrabackup logger.go:42: 14:53:47 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-mysql-1-xtrabackup.txt logger.go:42: 14:53:47 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-mysql-1-xtrabackup.txt logger.go:42: 14:53:47 | haproxy/8-check-password-leak | + for c in '$containers' logger.go:42: 14:53:47 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-viable-phoenix logs haproxy-mysql-1 -c pt-heartbeat logger.go:42: 14:53:47 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-mysql-1-pt-heartbeat.txt logger.go:42: 14:53:47 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-mysql-1-pt-heartbeat.txt logger.go:42: 14:53:47 | haproxy/8-check-password-leak | + echo logger.go:42: 14:53:47 | haproxy/8-check-password-leak | logger.go:42: 14:53:47 | haproxy/8-check-password-leak | + for p in '$pods' logger.go:42: 14:53:47 | haproxy/8-check-password-leak | ++ kubectl -n kuttl-test-viable-phoenix get pod haproxy-mysql-2 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 14:53:48 | haproxy/8-check-password-leak | + local 'containers=mysql xtrabackup pt-heartbeat' logger.go:42: 14:53:48 | haproxy/8-check-password-leak | + for c in '$containers' logger.go:42: 14:53:48 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-viable-phoenix logs haproxy-mysql-2 -c mysql logger.go:42: 14:53:49 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-mysql-2-mysql.txt logger.go:42: 14:53:49 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-mysql-2-mysql.txt logger.go:42: 14:53:49 | haproxy/8-check-password-leak | + for c in '$containers' logger.go:42: 14:53:49 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-viable-phoenix logs haproxy-mysql-2 -c xtrabackup logger.go:42: 14:53:50 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-mysql-2-xtrabackup.txt logger.go:42: 14:53:50 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-mysql-2-xtrabackup.txt logger.go:42: 14:53:50 | haproxy/8-check-password-leak | + for c in '$containers' logger.go:42: 14:53:50 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-viable-phoenix logs haproxy-mysql-2 -c pt-heartbeat logger.go:42: 14:53:51 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-mysql-2-pt-heartbeat.txt logger.go:42: 14:53:51 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-mysql-2-pt-heartbeat.txt logger.go:42: 14:53:51 | haproxy/8-check-password-leak | + echo logger.go:42: 14:53:51 | haproxy/8-check-password-leak | logger.go:42: 14:53:51 | haproxy/8-check-password-leak | + for p in '$pods' logger.go:42: 14:53:51 | haproxy/8-check-password-leak | ++ kubectl -n kuttl-test-viable-phoenix get pod haproxy-orc-0 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 14:53:52 | haproxy/8-check-password-leak | + local 'containers=orc mysql-monit' logger.go:42: 14:53:52 | haproxy/8-check-password-leak | + for c in '$containers' logger.go:42: 14:53:52 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-viable-phoenix logs haproxy-orc-0 -c orc logger.go:42: 14:53:53 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-orc-0-orc.txt logger.go:42: 14:53:53 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-orc-0-orc.txt logger.go:42: 14:53:53 | haproxy/8-check-password-leak | + for c in '$containers' logger.go:42: 14:53:53 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-viable-phoenix logs haproxy-orc-0 -c mysql-monit logger.go:42: 14:53:54 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-orc-0-mysql-monit.txt logger.go:42: 14:53:54 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-orc-0-mysql-monit.txt logger.go:42: 14:53:54 | haproxy/8-check-password-leak | + echo logger.go:42: 14:53:54 | haproxy/8-check-password-leak | logger.go:42: 14:53:54 | haproxy/8-check-password-leak | + for p in '$pods' logger.go:42: 14:53:54 | haproxy/8-check-password-leak | ++ kubectl -n kuttl-test-viable-phoenix get pod haproxy-orc-1 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 14:53:55 | haproxy/8-check-password-leak | + local 'containers=orc mysql-monit' logger.go:42: 14:53:55 | haproxy/8-check-password-leak | + for c in '$containers' logger.go:42: 14:53:55 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-viable-phoenix logs haproxy-orc-1 -c orc logger.go:42: 14:53:56 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-orc-1-orc.txt logger.go:42: 14:53:56 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-orc-1-orc.txt logger.go:42: 14:53:56 | haproxy/8-check-password-leak | + for c in '$containers' logger.go:42: 14:53:56 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-viable-phoenix logs haproxy-orc-1 -c mysql-monit logger.go:42: 14:53:57 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-orc-1-mysql-monit.txt logger.go:42: 14:53:57 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-orc-1-mysql-monit.txt logger.go:42: 14:53:57 | haproxy/8-check-password-leak | + echo logger.go:42: 14:53:57 | haproxy/8-check-password-leak | logger.go:42: 14:53:57 | haproxy/8-check-password-leak | + for p in '$pods' logger.go:42: 14:53:57 | haproxy/8-check-password-leak | ++ kubectl -n kuttl-test-viable-phoenix get pod haproxy-orc-2 -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 14:53:58 | haproxy/8-check-password-leak | + local 'containers=orc mysql-monit' logger.go:42: 14:53:58 | haproxy/8-check-password-leak | + for c in '$containers' logger.go:42: 14:53:58 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-viable-phoenix logs haproxy-orc-2 -c orc logger.go:42: 14:53:59 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-orc-2-orc.txt logger.go:42: 14:53:59 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-orc-2-orc.txt logger.go:42: 14:53:59 | haproxy/8-check-password-leak | + for c in '$containers' logger.go:42: 14:53:59 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-viable-phoenix logs haproxy-orc-2 -c mysql-monit logger.go:42: 14:54:00 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-orc-2-mysql-monit.txt logger.go:42: 14:54:00 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.cNnMs035kn/logs_output-haproxy-orc-2-mysql-monit.txt logger.go:42: 14:54:00 | haproxy/8-check-password-leak | + echo logger.go:42: 14:54:00 | haproxy/8-check-password-leak | logger.go:42: 14:54:00 | haproxy/8-check-password-leak | + for p in '$pods' logger.go:42: 14:54:00 | haproxy/8-check-password-leak | ++ kubectl -n kuttl-test-viable-phoenix get pod mysql-client -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 14:54:01 | haproxy/8-check-password-leak | + local containers=mysql-client logger.go:42: 14:54:01 | haproxy/8-check-password-leak | + for c in '$containers' logger.go:42: 14:54:01 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-viable-phoenix logs mysql-client -c mysql-client logger.go:42: 14:54:02 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.cNnMs035kn/logs_output-mysql-client-mysql-client.txt logger.go:42: 14:54:02 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.cNnMs035kn/logs_output-mysql-client-mysql-client.txt logger.go:42: 14:54:02 | haproxy/8-check-password-leak | + echo logger.go:42: 14:54:02 | haproxy/8-check-password-leak | logger.go:42: 14:54:02 | haproxy/8-check-password-leak | + for p in '$pods' logger.go:42: 14:54:02 | haproxy/8-check-password-leak | ++ kubectl -n kuttl-test-viable-phoenix get pod percona-server-mysql-operator-6b56d66f99-5q5ns -o 'jsonpath={.spec.containers[*].name}' logger.go:42: 14:54:02 | haproxy/8-check-password-leak | + local containers=manager logger.go:42: 14:54:02 | haproxy/8-check-password-leak | + for c in '$containers' logger.go:42: 14:54:02 | haproxy/8-check-password-leak | + kubectl -n kuttl-test-viable-phoenix logs percona-server-mysql-operator-6b56d66f99-5q5ns -c manager logger.go:42: 14:54:03 | haproxy/8-check-password-leak | + echo logs saved in: /tmp/tmp.cNnMs035kn/logs_output-percona-server-mysql-operator-6b56d66f99-5q5ns-manager.txt logger.go:42: 14:54:03 | haproxy/8-check-password-leak | logs saved in: /tmp/tmp.cNnMs035kn/logs_output-percona-server-mysql-operator-6b56d66f99-5q5ns-manager.txt logger.go:42: 14:54:03 | haproxy/8-check-password-leak | + echo logger.go:42: 14:54:03 | haproxy/8-check-password-leak | logger.go:42: 14:54:03 | haproxy/8-check-password-leak | + '[' -n '' ']' logger.go:42: 14:54:04 | haproxy/8-check-password-leak | test step completed 8-check-password-leak logger.go:42: 14:54:04 | haproxy/9-disable-haproxy | starting test step 9-disable-haproxy logger.go:42: 14:54:06 | haproxy/9-disable-haproxy | PerconaServerMySQL:kuttl-test-viable-phoenix/haproxy updated logger.go:42: 14:54:13 | haproxy/9-disable-haproxy | test step completed 9-disable-haproxy logger.go:42: 14:54:13 | haproxy/10-drop-finalizer | starting test step 10-drop-finalizer logger.go:42: 14:54:15 | haproxy/10-drop-finalizer | PerconaServerMySQL:kuttl-test-viable-phoenix/haproxy updated logger.go:42: 14:54:15 | haproxy/10-drop-finalizer | test step completed 10-drop-finalizer logger.go:42: 14:54:15 | haproxy | haproxy events from ns kuttl-test-viable-phoenix: logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:47:43 +0000 UTC Normal ReplicaSet.apps percona-server-mysql-operator-6b56d66f99 SuccessfulCreate Created pod: percona-server-mysql-operator-6b56d66f99-5q5ns logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:47:43 +0000 UTC Normal Deployment.apps percona-server-mysql-operator ScalingReplicaSet Scaled up replica set percona-server-mysql-operator-6b56d66f99 to 1 logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:47:44 +0000 UTC Normal Pod percona-server-mysql-operator-6b56d66f99-5q5ns Scheduled Successfully assigned kuttl-test-viable-phoenix/percona-server-mysql-operator-6b56d66f99-5q5ns to gke-jen-ps-424-70568ae-7-default-pool-aeac7083-dd2p logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:47:46 +0000 UTC Normal Lease.coordination.k8s.io 08db2feb.percona.com LeaderElection percona-server-mysql-operator-6b56d66f99-5q5ns_d3a9a440-492b-4588-b3b1-ba264b1f2d8f became leader logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:47:46 +0000 UTC Normal Pod percona-server-mysql-operator-6b56d66f99-5q5ns.spec.containers{manager} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-424-70568ae" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:47:46 +0000 UTC Normal Pod percona-server-mysql-operator-6b56d66f99-5q5ns.spec.containers{manager} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-424-70568ae" in 422.887871ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:47:46 +0000 UTC Normal Pod percona-server-mysql-operator-6b56d66f99-5q5ns.spec.containers{manager} Created Created container manager logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:47:46 +0000 UTC Normal Pod percona-server-mysql-operator-6b56d66f99-5q5ns.spec.containers{manager} Started Started container manager logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:47:47 +0000 UTC Normal Pod mysql-client Scheduled Successfully assigned kuttl-test-viable-phoenix/mysql-client to gke-jen-ps-424-70568ae-7-default-pool-aeac7083-sbvs logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:47:48 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Pulled Container image "percona/percona-server:8.0.25" already present on machine logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:47:48 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Created Created container mysql-client logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:47:48 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Started Started container mysql-client logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:03 +0000 UTC Normal Certificate.cert-manager.io haproxy-ca-cert Issuing Issuing certificate as Secret does not exist logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:04 +0000 UTC Normal CertificateRequest.cert-manager.io haproxy-ca-cert-pvfn8 WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:04 +0000 UTC Normal CertificateRequest.cert-manager.io haproxy-ca-cert-pvfn8 WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:04 +0000 UTC Normal CertificateRequest.cert-manager.io haproxy-ca-cert-pvfn8 WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:04 +0000 UTC Normal CertificateRequest.cert-manager.io haproxy-ca-cert-pvfn8 WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:04 +0000 UTC Normal CertificateRequest.cert-manager.io haproxy-ca-cert-pvfn8 WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:04 +0000 UTC Normal CertificateRequest.cert-manager.io haproxy-ca-cert-pvfn8 cert-manager.io Certificate request has been approved by cert-manager.io logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:04 +0000 UTC Normal CertificateRequest.cert-manager.io haproxy-ca-cert-pvfn8 CertificateIssued Certificate fetched from issuer successfully logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:04 +0000 UTC Normal Certificate.cert-manager.io haproxy-ca-cert Generated Stored new private key in temporary Secret resource "haproxy-ca-cert-m9c26" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:04 +0000 UTC Normal Certificate.cert-manager.io haproxy-ca-cert Requested Created new CertificateRequest resource "haproxy-ca-cert-pvfn8" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:04 +0000 UTC Normal Certificate.cert-manager.io haproxy-ca-cert Issuing The certificate has been successfully issued logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:06 +0000 UTC Normal Issuer.cert-manager.io haproxy-pso-issuer KeyPairVerified Signing CA verified logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:06 +0000 UTC Normal CertificateRequest.cert-manager.io haproxy-ssl-jtcmn WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:06 +0000 UTC Normal CertificateRequest.cert-manager.io haproxy-ssl-jtcmn WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:06 +0000 UTC Normal CertificateRequest.cert-manager.io haproxy-ssl-jtcmn WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:06 +0000 UTC Normal CertificateRequest.cert-manager.io haproxy-ssl-jtcmn WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:06 +0000 UTC Normal CertificateRequest.cert-manager.io haproxy-ssl-jtcmn WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:06 +0000 UTC Normal CertificateRequest.cert-manager.io haproxy-ssl-jtcmn cert-manager.io Certificate request has been approved by cert-manager.io logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:06 +0000 UTC Normal CertificateRequest.cert-manager.io haproxy-ssl-jtcmn CertificateIssued Certificate fetched from issuer successfully logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:06 +0000 UTC Normal Certificate.cert-manager.io haproxy-ssl Issuing Issuing certificate as Secret was previously issued by Issuer.cert-manager.io/ logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:06 +0000 UTC Normal Certificate.cert-manager.io haproxy-ssl Reused Reusing private key stored in existing Secret resource "test-ssl" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:06 +0000 UTC Normal Certificate.cert-manager.io haproxy-ssl Requested Created new CertificateRequest resource "haproxy-ssl-jtcmn" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:06 +0000 UTC Normal Certificate.cert-manager.io haproxy-ssl Issuing The certificate has been successfully issued logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:10 +0000 UTC Normal PersistentVolumeClaim datadir-haproxy-mysql-0 WaitForFirstConsumer waiting for first consumer to be created before binding logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:10 +0000 UTC Normal PersistentVolumeClaim datadir-haproxy-mysql-0 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:10 +0000 UTC Normal PersistentVolumeClaim datadir-haproxy-mysql-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-viable-phoenix/datadir-haproxy-mysql-0" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:10 +0000 UTC Normal StatefulSet.apps haproxy-mysql SuccessfulCreate create Claim datadir-haproxy-mysql-0 Pod haproxy-mysql-0 in StatefulSet haproxy-mysql success logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:10 +0000 UTC Normal StatefulSet.apps haproxy-mysql SuccessfulCreate create Pod haproxy-mysql-0 in StatefulSet haproxy-mysql successful logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:10 +0000 UTC Normal Pod haproxy-orc-0 Scheduled Successfully assigned kuttl-test-viable-phoenix/haproxy-orc-0 to gke-jen-ps-424-70568ae-7-default-pool-aeac7083-sbvs logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:10 +0000 UTC Normal StatefulSet.apps haproxy-orc SuccessfulCreate create Pod haproxy-orc-0 in StatefulSet haproxy-orc successful logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:11 +0000 UTC Normal Pod haproxy-orc-0.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-424-70568ae" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:11 +0000 UTC Normal Pod haproxy-orc-0.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-424-70568ae" in 376.685916ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:11 +0000 UTC Normal Pod haproxy-orc-0.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:11 +0000 UTC Normal Pod haproxy-orc-0.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:13 +0000 UTC Normal Pod haproxy-orc-0.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:14 +0000 UTC Normal PersistentVolumeClaim datadir-haproxy-mysql-0 ProvisioningSucceeded Successfully provisioned volume pvc-ff09a1a7-d652-476c-a701-6e32b2a316dd logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:14 +0000 UTC Normal Pod haproxy-orc-0.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 421.435946ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:14 +0000 UTC Normal Pod haproxy-orc-0.spec.containers{orc} Created Created container orc logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:14 +0000 UTC Normal Pod haproxy-orc-0.spec.containers{orc} Started Started container orc logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:14 +0000 UTC Normal Pod haproxy-orc-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:14 +0000 UTC Normal Pod haproxy-orc-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 373.581753ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:14 +0000 UTC Normal Pod haproxy-orc-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:14 +0000 UTC Normal Pod haproxy-orc-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:15 +0000 UTC Normal Pod haproxy-mysql-0 Scheduled Successfully assigned kuttl-test-viable-phoenix/haproxy-mysql-0 to gke-jen-ps-424-70568ae-7-default-pool-aeac7083-dd2p logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:24 +0000 UTC Normal Pod haproxy-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-ff09a1a7-d652-476c-a701-6e32b2a316dd" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:25 +0000 UTC Normal Pod haproxy-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-424-70568ae" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:25 +0000 UTC Normal Pod haproxy-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-424-70568ae" in 405.282091ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:26 +0000 UTC Normal Pod haproxy-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:26 +0000 UTC Normal Pod haproxy-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:27 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:28 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 514.698977ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:28 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{mysql} Created Created container mysql logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:28 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{mysql} Started Started container mysql logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:28 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:28 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 416.844138ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:28 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:28 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:29 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:29 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 774.443218ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:29 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:29 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:46 +0000 UTC Normal Pod haproxy-orc-1 Scheduled Successfully assigned kuttl-test-viable-phoenix/haproxy-orc-1 to gke-jen-ps-424-70568ae-7-default-pool-aeac7083-zvbd logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:46 +0000 UTC Normal StatefulSet.apps haproxy-orc SuccessfulCreate create Pod haproxy-orc-1 in StatefulSet haproxy-orc successful logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:47 +0000 UTC Normal Pod haproxy-orc-1.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-424-70568ae" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:47 +0000 UTC Normal Pod haproxy-orc-1.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-424-70568ae" in 388.568281ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:47 +0000 UTC Normal Pod haproxy-orc-1.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:47 +0000 UTC Normal Pod haproxy-orc-1.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:49 +0000 UTC Normal Pod haproxy-orc-1.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:49 +0000 UTC Normal Pod haproxy-orc-1.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 502.73677ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:49 +0000 UTC Normal Pod haproxy-orc-1.spec.containers{orc} Created Created container orc logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:49 +0000 UTC Normal Pod haproxy-orc-1.spec.containers{orc} Started Started container orc logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:49 +0000 UTC Normal Pod haproxy-orc-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:50 +0000 UTC Normal Pod haproxy-orc-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 365.975143ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:50 +0000 UTC Normal Pod haproxy-orc-1.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:48:50 +0000 UTC Normal Pod haproxy-orc-1.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:00 +0000 UTC Normal PersistentVolumeClaim datadir-haproxy-mysql-1 WaitForFirstConsumer waiting for first consumer to be created before binding logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:00 +0000 UTC Normal PersistentVolumeClaim datadir-haproxy-mysql-1 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:00 +0000 UTC Normal PersistentVolumeClaim datadir-haproxy-mysql-1 Provisioning External provisioner is provisioning volume for claim "kuttl-test-viable-phoenix/datadir-haproxy-mysql-1" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:00 +0000 UTC Normal StatefulSet.apps haproxy-mysql SuccessfulCreate create Claim datadir-haproxy-mysql-1 Pod haproxy-mysql-1 in StatefulSet haproxy-mysql success logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:00 +0000 UTC Normal StatefulSet.apps haproxy-mysql SuccessfulCreate create Pod haproxy-mysql-1 in StatefulSet haproxy-mysql successful logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:04 +0000 UTC Normal PersistentVolumeClaim datadir-haproxy-mysql-1 ProvisioningSucceeded Successfully provisioned volume pvc-243c78c4-46b3-4e96-bbb3-7d21d770140d logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:04 +0000 UTC Normal Pod haproxy-mysql-1 Scheduled Successfully assigned kuttl-test-viable-phoenix/haproxy-mysql-1 to gke-jen-ps-424-70568ae-7-default-pool-aeac7083-zvbd logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:05 +0000 UTC Normal Pod haproxy-haproxy-0 Scheduled Successfully assigned kuttl-test-viable-phoenix/haproxy-haproxy-0 to gke-jen-ps-424-70568ae-7-default-pool-aeac7083-sbvs logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:05 +0000 UTC Normal StatefulSet.apps haproxy-haproxy SuccessfulCreate create Pod haproxy-haproxy-0 in StatefulSet haproxy-haproxy successful logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:06 +0000 UTC Normal Pod haproxy-haproxy-0.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-424-70568ae" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:06 +0000 UTC Normal Pod haproxy-haproxy-0.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-424-70568ae" in 360.979709ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:06 +0000 UTC Normal Pod haproxy-haproxy-0.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:06 +0000 UTC Normal Pod haproxy-haproxy-0.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:08 +0000 UTC Normal Pod haproxy-haproxy-0.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:08 +0000 UTC Normal Pod haproxy-haproxy-0.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 416.01098ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:08 +0000 UTC Normal Pod haproxy-haproxy-0.spec.containers{haproxy} Created Created container haproxy logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:08 +0000 UTC Normal Pod haproxy-haproxy-0.spec.containers{haproxy} Started Started container haproxy logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:08 +0000 UTC Normal Pod haproxy-haproxy-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:09 +0000 UTC Normal Pod haproxy-haproxy-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 388.609055ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:09 +0000 UTC Normal Pod haproxy-haproxy-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:09 +0000 UTC Normal Pod haproxy-haproxy-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:09 +0000 UTC Normal Pod haproxy-haproxy-1 Scheduled Successfully assigned kuttl-test-viable-phoenix/haproxy-haproxy-1 to gke-jen-ps-424-70568ae-7-default-pool-aeac7083-dd2p logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:09 +0000 UTC Normal StatefulSet.apps haproxy-haproxy SuccessfulCreate create Pod haproxy-haproxy-1 in StatefulSet haproxy-haproxy successful logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:10 +0000 UTC Normal Pod haproxy-haproxy-1.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-424-70568ae" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:10 +0000 UTC Normal Pod haproxy-haproxy-1.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-424-70568ae" in 428.202668ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:10 +0000 UTC Normal Pod haproxy-haproxy-1.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:10 +0000 UTC Normal Pod haproxy-haproxy-1.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:11 +0000 UTC Normal Pod haproxy-haproxy-1.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:11 +0000 UTC Normal Pod haproxy-mysql-1 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-243c78c4-46b3-4e96-bbb3-7d21d770140d" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:12 +0000 UTC Normal Pod haproxy-haproxy-1.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 378.652504ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:12 +0000 UTC Normal Pod haproxy-haproxy-1.spec.containers{haproxy} Created Created container haproxy logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:12 +0000 UTC Normal Pod haproxy-haproxy-1.spec.containers{haproxy} Started Started container haproxy logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:12 +0000 UTC Normal Pod haproxy-haproxy-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:12 +0000 UTC Normal Pod haproxy-haproxy-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 367.448113ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:12 +0000 UTC Normal Pod haproxy-haproxy-1.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:12 +0000 UTC Normal Pod haproxy-haproxy-1.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:13 +0000 UTC Normal Pod haproxy-haproxy-2 Scheduled Successfully assigned kuttl-test-viable-phoenix/haproxy-haproxy-2 to gke-jen-ps-424-70568ae-7-default-pool-aeac7083-zvbd logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:13 +0000 UTC Normal StatefulSet.apps haproxy-haproxy SuccessfulCreate create Pod haproxy-haproxy-2 in StatefulSet haproxy-haproxy successful logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:13 +0000 UTC Normal Pod haproxy-mysql-1.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-424-70568ae" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:13 +0000 UTC Normal Pod haproxy-mysql-1.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-424-70568ae" in 397.379932ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:13 +0000 UTC Normal Pod haproxy-mysql-1.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:13 +0000 UTC Normal Pod haproxy-mysql-1.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:14 +0000 UTC Normal Pod haproxy-haproxy-2.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-424-70568ae" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:15 +0000 UTC Normal Pod haproxy-haproxy-2.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-424-70568ae" in 1.015272327s logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:15 +0000 UTC Normal Pod haproxy-haproxy-2.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:15 +0000 UTC Normal Pod haproxy-haproxy-2.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:15 +0000 UTC Normal Pod haproxy-mysql-1.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:16 +0000 UTC Normal Pod haproxy-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 1.02031251s logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:16 +0000 UTC Normal Pod haproxy-mysql-1.spec.containers{mysql} Created Created container mysql logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:16 +0000 UTC Normal Pod haproxy-mysql-1.spec.containers{mysql} Started Started container mysql logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:16 +0000 UTC Normal Pod haproxy-mysql-1.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:16 +0000 UTC Normal Pod haproxy-mysql-1.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 377.7431ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:16 +0000 UTC Normal Pod haproxy-mysql-1.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:16 +0000 UTC Normal Pod haproxy-mysql-1.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:16 +0000 UTC Normal Pod haproxy-mysql-1.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:17 +0000 UTC Normal Pod haproxy-haproxy-2.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:17 +0000 UTC Normal Pod haproxy-haproxy-2.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 469.113617ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:17 +0000 UTC Normal Pod haproxy-haproxy-2.spec.containers{haproxy} Created Created container haproxy logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:17 +0000 UTC Normal Pod haproxy-haproxy-2.spec.containers{haproxy} Started Started container haproxy logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:17 +0000 UTC Normal Pod haproxy-haproxy-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:17 +0000 UTC Normal Pod haproxy-mysql-1.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 383.811888ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:17 +0000 UTC Normal Pod haproxy-mysql-1.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:17 +0000 UTC Normal Pod haproxy-mysql-1.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:18 +0000 UTC Normal Pod haproxy-haproxy-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 372.818ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:18 +0000 UTC Normal Pod haproxy-haproxy-2.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:18 +0000 UTC Normal Pod haproxy-haproxy-2.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:21 +0000 UTC Normal Pod haproxy-orc-2 Scheduled Successfully assigned kuttl-test-viable-phoenix/haproxy-orc-2 to gke-jen-ps-424-70568ae-7-default-pool-aeac7083-dd2p logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:21 +0000 UTC Normal StatefulSet.apps haproxy-orc SuccessfulCreate create Pod haproxy-orc-2 in StatefulSet haproxy-orc successful logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:22 +0000 UTC Normal Pod haproxy-orc-2.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-424-70568ae" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:23 +0000 UTC Normal Pod haproxy-orc-2.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-424-70568ae" in 379.668944ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:23 +0000 UTC Normal Pod haproxy-orc-2.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:23 +0000 UTC Normal Pod haproxy-orc-2.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:25 +0000 UTC Normal Pod haproxy-orc-2.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:25 +0000 UTC Normal Pod haproxy-orc-2.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 352.52639ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:25 +0000 UTC Normal Pod haproxy-orc-2.spec.containers{orc} Created Created container orc logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:25 +0000 UTC Normal Pod haproxy-orc-2.spec.containers{orc} Started Started container orc logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:25 +0000 UTC Normal Pod haproxy-orc-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:25 +0000 UTC Normal Pod haproxy-orc-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 506.430376ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:25 +0000 UTC Normal Pod haproxy-orc-2.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:26 +0000 UTC Normal Pod haproxy-orc-2.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:34 +0000 UTC Warning Pod haproxy-mysql-1.spec.containers{mysql} Unhealthy Startup probe failed: logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:34 +0000 UTC Normal Pod haproxy-mysql-1.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:49:37 +0000 UTC Normal Pod haproxy-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 415.215223ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:13 +0000 UTC Normal PersistentVolumeClaim datadir-haproxy-mysql-2 WaitForFirstConsumer waiting for first consumer to be created before binding logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:13 +0000 UTC Normal PersistentVolumeClaim datadir-haproxy-mysql-2 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:13 +0000 UTC Normal PersistentVolumeClaim datadir-haproxy-mysql-2 Provisioning External provisioner is provisioning volume for claim "kuttl-test-viable-phoenix/datadir-haproxy-mysql-2" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:13 +0000 UTC Normal StatefulSet.apps haproxy-mysql SuccessfulCreate create Claim datadir-haproxy-mysql-2 Pod haproxy-mysql-2 in StatefulSet haproxy-mysql success logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:13 +0000 UTC Normal StatefulSet.apps haproxy-mysql SuccessfulCreate create Pod haproxy-mysql-2 in StatefulSet haproxy-mysql successful logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:17 +0000 UTC Normal PersistentVolumeClaim datadir-haproxy-mysql-2 ProvisioningSucceeded Successfully provisioned volume pvc-7997fc25-aa55-4da5-97ff-b4645bc1fe1c logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:17 +0000 UTC Normal Pod haproxy-mysql-2 Scheduled Successfully assigned kuttl-test-viable-phoenix/haproxy-mysql-2 to gke-jen-ps-424-70568ae-7-default-pool-aeac7083-sbvs logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:21 +0000 UTC Normal Pod haproxy-mysql-2 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-7997fc25-aa55-4da5-97ff-b4645bc1fe1c" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:26 +0000 UTC Normal Pod haproxy-mysql-2.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-424-70568ae" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:26 +0000 UTC Normal Pod haproxy-mysql-2.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-424-70568ae" in 393.549957ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:26 +0000 UTC Normal Pod haproxy-mysql-2.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:27 +0000 UTC Normal Pod haproxy-mysql-2.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:28 +0000 UTC Normal Pod haproxy-mysql-2.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:28 +0000 UTC Normal Pod haproxy-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 414.535215ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:28 +0000 UTC Normal Pod haproxy-mysql-2.spec.containers{mysql} Created Created container mysql logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:28 +0000 UTC Normal Pod haproxy-mysql-2.spec.containers{mysql} Started Started container mysql logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:28 +0000 UTC Normal Pod haproxy-mysql-2.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:29 +0000 UTC Normal Pod haproxy-mysql-2.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 473.028324ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:29 +0000 UTC Normal Pod haproxy-mysql-2.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:29 +0000 UTC Normal Pod haproxy-mysql-2.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:29 +0000 UTC Normal Pod haproxy-mysql-2.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:29 +0000 UTC Normal Pod haproxy-mysql-2.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 360.439931ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:29 +0000 UTC Normal Pod haproxy-mysql-2.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:29 +0000 UTC Normal Pod haproxy-mysql-2.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:47 +0000 UTC Warning Pod haproxy-mysql-2.spec.containers{mysql} Unhealthy Startup probe failed: logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:47 +0000 UTC Normal Pod haproxy-mysql-2.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:50:50 +0000 UTC Normal Pod haproxy-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 363.51619ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:52:43 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{mysql} Killing Stopping container mysql logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:52:43 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{pt-heartbeat} Killing Stopping container pt-heartbeat logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:52:43 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{xtrabackup} Killing Stopping container xtrabackup logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:52:45 +0000 UTC Warning Pod haproxy-mysql-0.spec.containers{mysql} Unhealthy Readiness probe failed: 2023/08/17 14:52:45 readiness check failed: connect to db: ping database: dial tcp 10.6.88.20:33062: connect: connection refused logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:52:50 +0000 UTC Warning Pod haproxy-mysql-0.spec.containers{mysql} Unhealthy Readiness probe failed: 2023/08/17 14:52:50 readiness check failed: connect to db: ping database: dial tcp 10.6.88.20:33062: connect: connection refused logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:52:55 +0000 UTC Warning Pod haproxy-mysql-0.spec.containers{mysql} Unhealthy Readiness probe errored: rpc error: code = Unknown desc = failed to exec in container: container is in CONTAINER_EXITED state logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:53:03 +0000 UTC Normal Pod haproxy-mysql-0 Scheduled Successfully assigned kuttl-test-viable-phoenix/haproxy-mysql-0 to gke-jen-ps-424-70568ae-7-default-pool-aeac7083-dd2p logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:53:09 +0000 UTC Normal Pod haproxy-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-424-70568ae" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:53:09 +0000 UTC Normal Pod haproxy-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-424-70568ae" in 406.598688ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:53:09 +0000 UTC Normal Pod haproxy-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:53:09 +0000 UTC Normal Pod haproxy-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:53:11 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:53:12 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 406.608185ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:53:12 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{mysql} Created Created container mysql logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:53:12 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{mysql} Started Started container mysql logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:53:12 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:53:12 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 439.851227ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:53:12 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:53:12 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:53:12 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:53:13 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 358.737025ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:53:13 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:53:13 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:53:30 +0000 UTC Warning Pod haproxy-mysql-0.spec.containers{mysql} Unhealthy Startup probe failed: logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:53:30 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:53:33 +0000 UTC Normal Pod haproxy-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 438.124543ms logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:54:13 +0000 UTC Normal Pod haproxy-haproxy-0.spec.containers{haproxy} Killing Stopping container haproxy logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:54:13 +0000 UTC Normal Pod haproxy-haproxy-0.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:54:13 +0000 UTC Normal Pod haproxy-haproxy-1.spec.containers{haproxy} Killing Stopping container haproxy logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:54:13 +0000 UTC Normal Pod haproxy-haproxy-1.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:54:13 +0000 UTC Normal Pod haproxy-haproxy-2.spec.containers{haproxy} Killing Stopping container haproxy logger.go:42: 14:54:15 | haproxy | 2023-08-17 14:54:13 +0000 UTC Normal Pod haproxy-haproxy-2.spec.containers{mysql-monit} Killing Stopping container mysql-monit logger.go:42: 14:54:15 | haproxy | Deleting namespace: kuttl-test-viable-phoenix === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (451.56s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/haproxy (447.86s) PASS