=== RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://35.184.89.132 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 180 seconds for each step harness.go:372: testsuite: e2e-tests/tests has 30 tests === RUN kuttl/harness === RUN kuttl/harness/recreate === PAUSE kuttl/harness/recreate === CONT kuttl/harness/recreate logger.go:42: 14:45:09 | recreate | Creating namespace: kuttl-test-wired-lacewing logger.go:42: 14:45:09 | recreate/0-deploy-operator | starting test step 0-deploy-operator logger.go:42: 14:45:09 | recreate/0-deploy-operator | running command: [sh -c set -o errexit set -o xtrace source ../../functions init_temp_dir # do this only in the first TestStep deploy_operator deploy_non_tls_cluster_secrets deploy_tls_cluster_secrets deploy_client] logger.go:42: 14:45:09 | recreate/0-deploy-operator | + source ../../functions logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ realpath ../../.. logger.go:42: 14:45:09 | recreate/0-deploy-operator | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:45:09 | recreate/0-deploy-operator | ++++ pwd logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/recreate logger.go:42: 14:45:09 | recreate/0-deploy-operator | ++ test_name=recreate logger.go:42: 14:45:09 | recreate/0-deploy-operator | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ export TEMP_DIR=/tmp/kuttl/ps/recreate logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ TEMP_DIR=/tmp/kuttl/ps/recreate logger.go:42: 14:45:09 | recreate/0-deploy-operator | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ export GIT_BRANCH=PR-523 logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ GIT_BRANCH=PR-523 logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ export VERSION=PR-523-f00253e logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ VERSION=PR-523-f00253e logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:45:09 | recreate/0-deploy-operator | ++++ which gdate logger.go:42: 14:45:09 | recreate/0-deploy-operator | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:45:09 | recreate/0-deploy-operator | ++++ which date logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ date=/usr/bin/date logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ command -v oc logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ kubectl get nodes logger.go:42: 14:45:09 | recreate/0-deploy-operator | +++ grep '^minikube' logger.go:42: 14:45:10 | recreate/0-deploy-operator | + init_temp_dir logger.go:42: 14:45:10 | recreate/0-deploy-operator | + rm -rf /tmp/kuttl/ps/recreate logger.go:42: 14:45:10 | recreate/0-deploy-operator | + mkdir -p /tmp/kuttl/ps/recreate logger.go:42: 14:45:10 | recreate/0-deploy-operator | + deploy_operator logger.go:42: 14:45:10 | recreate/0-deploy-operator | + kubectl -n kuttl-test-wired-lacewing apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy/crd.yaml logger.go:42: 14:45:10 | recreate/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlbackups.ps.percona.com serverside-applied logger.go:42: 14:45:11 | recreate/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlrestores.ps.percona.com serverside-applied logger.go:42: 14:45:12 | recreate/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqls.ps.percona.com serverside-applied logger.go:42: 14:45:12 | recreate/0-deploy-operator | + kubectl -n kuttl-test-wired-lacewing apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy/rbac.yaml logger.go:42: 14:45:13 | recreate/0-deploy-operator | serviceaccount/percona-server-mysql-operator created logger.go:42: 14:45:13 | recreate/0-deploy-operator | serviceaccount/percona-server-mysql-operator-orchestrator created logger.go:42: 14:45:13 | recreate/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 14:45:14 | recreate/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 14:45:14 | recreate/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 14:45:14 | recreate/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 14:45:15 | recreate/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 14:45:15 | recreate/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 14:45:15 | recreate/0-deploy-operator | + yq eval '(select(documentIndex==1).spec.template.spec.containers[] | select(.name=="manager").env[] | select(.name=="LOG_LEVEL").value) = "DEBUG"' logger.go:42: 14:45:15 | recreate/0-deploy-operator | + yq eval '(select(documentIndex==1).spec.template.spec.containers[] | select(.name=="manager").env[] | select(.name=="DISABLE_TELEMETRY").value) = "true"' logger.go:42: 14:45:15 | recreate/0-deploy-operator | + kubectl -n kuttl-test-wired-lacewing apply -f - logger.go:42: 14:45:15 | recreate/0-deploy-operator | ++ printf 'select(documentIndex==1).spec.template.spec.containers[0].image="%s"' perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:45:15 | recreate/0-deploy-operator | + yq eval 'select(documentIndex==1).spec.template.spec.containers[0].image="perconalab/percona-server-mysql-operator:PR-523-f00253e"' /mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy/operator.yaml logger.go:42: 14:45:16 | recreate/0-deploy-operator | configmap/percona-server-mysql-operator-config created logger.go:42: 14:45:16 | recreate/0-deploy-operator | deployment.apps/percona-server-mysql-operator created logger.go:42: 14:45:16 | recreate/0-deploy-operator | + deploy_non_tls_cluster_secrets logger.go:42: 14:45:16 | recreate/0-deploy-operator | + kubectl -n kuttl-test-wired-lacewing apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf/secrets.yaml logger.go:42: 14:45:17 | recreate/0-deploy-operator | secret/test-secrets created logger.go:42: 14:45:17 | recreate/0-deploy-operator | + deploy_tls_cluster_secrets logger.go:42: 14:45:17 | recreate/0-deploy-operator | + kubectl -n kuttl-test-wired-lacewing apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf/ssl-secret.yaml logger.go:42: 14:45:18 | recreate/0-deploy-operator | secret/test-ssl created logger.go:42: 14:45:18 | recreate/0-deploy-operator | + deploy_client logger.go:42: 14:45:18 | recreate/0-deploy-operator | + kubectl -n kuttl-test-wired-lacewing apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf/client.yaml logger.go:42: 14:45:19 | recreate/0-deploy-operator | pod/mysql-client created logger.go:42: 14:45:28 | recreate/0-deploy-operator | test step completed 0-deploy-operator logger.go:42: 14:45:28 | recreate/1-create-cluster | starting test step 1-create-cluster logger.go:42: 14:45:28 | recreate/1-create-cluster | running command: [sh -c set -o errexit set -o xtrace source ../../functions get_cr \ | yq eval '.spec.mysql.clusterType="async"' - \ | yq eval '.spec.mysql.size=3' - \ | yq eval '.spec.proxy.haproxy.enabled=true' - \ | yq eval '.spec.proxy.haproxy.size=3' - \ | yq eval '.spec.orchestrator.enabled=true' - \ | yq eval '.spec.orchestrator.size=3' - \ | kubectl -n "${NAMESPACE}" apply -f -] logger.go:42: 14:45:28 | recreate/1-create-cluster | + source ../../functions logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ realpath ../../.. logger.go:42: 14:45:28 | recreate/1-create-cluster | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:45:28 | recreate/1-create-cluster | ++++ pwd logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/recreate logger.go:42: 14:45:28 | recreate/1-create-cluster | ++ test_name=recreate logger.go:42: 14:45:28 | recreate/1-create-cluster | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ export TEMP_DIR=/tmp/kuttl/ps/recreate logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ TEMP_DIR=/tmp/kuttl/ps/recreate logger.go:42: 14:45:28 | recreate/1-create-cluster | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ export GIT_BRANCH=PR-523 logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ GIT_BRANCH=PR-523 logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ export VERSION=PR-523-f00253e logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ VERSION=PR-523-f00253e logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:45:28 | recreate/1-create-cluster | ++++ which gdate logger.go:42: 14:45:28 | recreate/1-create-cluster | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:45:28 | recreate/1-create-cluster | ++++ which date logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ date=/usr/bin/date logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ command -v oc logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ kubectl get nodes logger.go:42: 14:45:28 | recreate/1-create-cluster | +++ grep '^minikube' logger.go:42: 14:45:29 | recreate/1-create-cluster | + get_cr logger.go:42: 14:45:29 | recreate/1-create-cluster | + local name_suffix= logger.go:42: 14:45:29 | recreate/1-create-cluster | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 14:45:29 | recreate/1-create-cluster | + yq eval .spec.mysql.size=3 - logger.go:42: 14:45:29 | recreate/1-create-cluster | ++ printf '.metadata.name="%s"' recreate logger.go:42: 14:45:29 | recreate/1-create-cluster | + kubectl -n kuttl-test-wired-lacewing apply -f - logger.go:42: 14:45:29 | recreate/1-create-cluster | + yq eval '.metadata.name="recreate"' /mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy/cr.yaml logger.go:42: 14:45:29 | recreate/1-create-cluster | + yq eval .spec.orchestrator.enabled=true - logger.go:42: 14:45:29 | recreate/1-create-cluster | + yq eval .spec.proxy.haproxy.size=3 - logger.go:42: 14:45:29 | recreate/1-create-cluster | + yq eval .spec.proxy.haproxy.enabled=true - logger.go:42: 14:45:29 | recreate/1-create-cluster | + yq eval .spec.orchestrator.size=3 - logger.go:42: 14:45:29 | recreate/1-create-cluster | + yq eval '.spec.secretsName="test-secrets"' - logger.go:42: 14:45:29 | recreate/1-create-cluster | + yq eval '.spec.sslSecretName="test-ssl"' - logger.go:42: 14:45:29 | recreate/1-create-cluster | + yq eval '.spec.upgradeOptions.apply="disabled"' - logger.go:42: 14:45:29 | recreate/1-create-cluster | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 14:45:29 | recreate/1-create-cluster | ++ printf '.spec.mysql.image="%s"' perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:45:29 | recreate/1-create-cluster | + yq eval '.spec.mysql.image="perconalab/percona-server-mysql-operator:main-psmysql"' - logger.go:42: 14:45:29 | recreate/1-create-cluster | ++ printf '.spec.orchestrator.image="%s"' perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:45:29 | recreate/1-create-cluster | + yq eval '.spec.orchestrator.image="perconalab/percona-server-mysql-operator:main-orchestrator"' - logger.go:42: 14:45:29 | recreate/1-create-cluster | ++ printf '.spec.toolkit.image="%s"' perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:45:29 | recreate/1-create-cluster | + yq eval '.spec.toolkit.image="perconalab/percona-server-mysql-operator:main-toolkit"' - logger.go:42: 14:45:29 | recreate/1-create-cluster | ++ printf '.spec.initImage="%s"' perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:45:29 | recreate/1-create-cluster | + yq eval '.spec.initImage="perconalab/percona-server-mysql-operator:PR-523-f00253e"' - logger.go:42: 14:45:29 | recreate/1-create-cluster | ++ printf '.spec.proxy.router.image="%s"' perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:45:29 | recreate/1-create-cluster | ++ printf '.spec.backup.image="%s"' perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:45:29 | recreate/1-create-cluster | + yq eval '.spec.backup.image="perconalab/percona-server-mysql-operator:main-backup"' - logger.go:42: 14:45:29 | recreate/1-create-cluster | + yq eval '.spec.proxy.router.image="perconalab/percona-server-mysql-operator:main-router"' - logger.go:42: 14:45:29 | recreate/1-create-cluster | ++ printf '.spec.proxy.haproxy.image="%s"' perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:45:29 | recreate/1-create-cluster | + yq eval '.spec.proxy.haproxy.image="perconalab/percona-server-mysql-operator:main-haproxy"' - logger.go:42: 14:45:29 | recreate/1-create-cluster | + '[' -n '' ']' logger.go:42: 14:45:29 | recreate/1-create-cluster | + yq eval - logger.go:42: 14:45:29 | recreate/1-create-cluster | ++ printf '.spec.pmm.image="%s"' perconalab/pmm-client:dev-latest logger.go:42: 14:45:29 | recreate/1-create-cluster | + yq eval '.spec.pmm.image="perconalab/pmm-client:dev-latest"' - logger.go:42: 14:45:30 | recreate/1-create-cluster | perconaservermysql.ps.percona.com/recreate created logger.go:42: 14:48:44 | recreate/1-create-cluster | test step completed 1-create-cluster logger.go:42: 14:48:44 | recreate/2-write-data | starting test step 2-write-data logger.go:42: 14:48:44 | recreate/2-write-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password" run_mysql \ "INSERT myDB.myTable (id) VALUES (100500)" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password" for i in 0 1 2; do host=$(get_mysql_headless_fqdn $(get_cluster_name) $i) data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${host} -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 02-write-data-${i} --from-literal=data="${data}" done] logger.go:42: 14:48:44 | recreate/2-write-data | + source ../../functions logger.go:42: 14:48:44 | recreate/2-write-data | +++ realpath ../../.. logger.go:42: 14:48:44 | recreate/2-write-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:48:44 | recreate/2-write-data | ++++ pwd logger.go:42: 14:48:44 | recreate/2-write-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/recreate logger.go:42: 14:48:44 | recreate/2-write-data | ++ test_name=recreate logger.go:42: 14:48:44 | recreate/2-write-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 14:48:44 | recreate/2-write-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:48:44 | recreate/2-write-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:48:44 | recreate/2-write-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:48:44 | recreate/2-write-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:48:44 | recreate/2-write-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:48:44 | recreate/2-write-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:48:44 | recreate/2-write-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:48:44 | recreate/2-write-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:48:44 | recreate/2-write-data | +++ export TEMP_DIR=/tmp/kuttl/ps/recreate logger.go:42: 14:48:44 | recreate/2-write-data | +++ TEMP_DIR=/tmp/kuttl/ps/recreate logger.go:42: 14:48:44 | recreate/2-write-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:48:44 | recreate/2-write-data | +++ export GIT_BRANCH=PR-523 logger.go:42: 14:48:44 | recreate/2-write-data | +++ GIT_BRANCH=PR-523 logger.go:42: 14:48:44 | recreate/2-write-data | +++ export VERSION=PR-523-f00253e logger.go:42: 14:48:44 | recreate/2-write-data | +++ VERSION=PR-523-f00253e logger.go:42: 14:48:44 | recreate/2-write-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:48:44 | recreate/2-write-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:48:44 | recreate/2-write-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:48:44 | recreate/2-write-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:48:44 | recreate/2-write-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:48:44 | recreate/2-write-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:48:44 | recreate/2-write-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:48:44 | recreate/2-write-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:48:44 | recreate/2-write-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:48:44 | recreate/2-write-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:48:44 | recreate/2-write-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:48:44 | recreate/2-write-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:48:44 | recreate/2-write-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:48:44 | recreate/2-write-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:48:44 | recreate/2-write-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:48:44 | recreate/2-write-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:48:44 | recreate/2-write-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:48:44 | recreate/2-write-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:48:44 | recreate/2-write-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:48:44 | recreate/2-write-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:48:44 | recreate/2-write-data | ++++ which gdate logger.go:42: 14:48:44 | recreate/2-write-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:48:44 | recreate/2-write-data | ++++ which date logger.go:42: 14:48:44 | recreate/2-write-data | +++ date=/usr/bin/date logger.go:42: 14:48:44 | recreate/2-write-data | +++ command -v oc logger.go:42: 14:48:44 | recreate/2-write-data | +++ kubectl get nodes logger.go:42: 14:48:44 | recreate/2-write-data | +++ grep '^minikube' logger.go:42: 14:48:45 | recreate/2-write-data | +++ get_cluster_name logger.go:42: 14:48:45 | recreate/2-write-data | +++ kubectl -n kuttl-test-wired-lacewing get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:48:45 | recreate/2-write-data | ++ get_haproxy_svc recreate logger.go:42: 14:48:45 | recreate/2-write-data | ++ local cluster=recreate logger.go:42: 14:48:45 | recreate/2-write-data | ++ echo recreate-haproxy logger.go:42: 14:48:45 | recreate/2-write-data | + run_mysql 'CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' '-h recreate-haproxy -uroot -proot_password' logger.go:42: 14:48:45 | recreate/2-write-data | + local 'command=CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' logger.go:42: 14:48:45 | recreate/2-write-data | + local 'uri=-h recreate-haproxy -uroot -proot_password' logger.go:42: 14:48:45 | recreate/2-write-data | + local pod= logger.go:42: 14:48:45 | recreate/2-write-data | ++ get_client_pod logger.go:42: 14:48:45 | recreate/2-write-data | ++ kubectl -n kuttl-test-wired-lacewing get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:48:46 | recreate/2-write-data | + client_pod=mysql-client logger.go:42: 14:48:46 | recreate/2-write-data | + wait_pod mysql-client logger.go:42: 14:48:46 | recreate/2-write-data | + local pod=mysql-client logger.go:42: 14:48:46 | recreate/2-write-data | + set +o xtrace logger.go:42: 14:48:46 | recreate/2-write-data | mysql-clienttrue logger.go:42: 14:48:46 | recreate/2-write-data | + sed -e 's/mysql: //' logger.go:42: 14:48:46 | recreate/2-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:48:46 | recreate/2-write-data | + kubectl -n kuttl-test-wired-lacewing exec mysql-client -- bash -c 'printf '\''%s\n'\'' "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" | mysql -sN -h recreate-haproxy -uroot -proot_password' logger.go:42: 14:48:47 | recreate/2-write-data | + : logger.go:42: 14:48:47 | recreate/2-write-data | +++ get_cluster_name logger.go:42: 14:48:47 | recreate/2-write-data | +++ kubectl -n kuttl-test-wired-lacewing get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:48:48 | recreate/2-write-data | ++ get_haproxy_svc recreate logger.go:42: 14:48:48 | recreate/2-write-data | ++ local cluster=recreate logger.go:42: 14:48:48 | recreate/2-write-data | ++ echo recreate-haproxy logger.go:42: 14:48:48 | recreate/2-write-data | + run_mysql 'INSERT myDB.myTable (id) VALUES (100500)' '-h recreate-haproxy -uroot -proot_password' logger.go:42: 14:48:48 | recreate/2-write-data | + local 'command=INSERT myDB.myTable (id) VALUES (100500)' logger.go:42: 14:48:48 | recreate/2-write-data | + local 'uri=-h recreate-haproxy -uroot -proot_password' logger.go:42: 14:48:48 | recreate/2-write-data | + local pod= logger.go:42: 14:48:48 | recreate/2-write-data | ++ get_client_pod logger.go:42: 14:48:48 | recreate/2-write-data | ++ kubectl -n kuttl-test-wired-lacewing get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:48:48 | recreate/2-write-data | + client_pod=mysql-client logger.go:42: 14:48:48 | recreate/2-write-data | + wait_pod mysql-client logger.go:42: 14:48:48 | recreate/2-write-data | + local pod=mysql-client logger.go:42: 14:48:48 | recreate/2-write-data | + set +o xtrace logger.go:42: 14:48:48 | recreate/2-write-data | mysql-clienttrue logger.go:42: 14:48:48 | recreate/2-write-data | + kubectl -n kuttl-test-wired-lacewing exec mysql-client -- bash -c 'printf '\''%s\n'\'' "INSERT myDB.myTable (id) VALUES (100500)" | mysql -sN -h recreate-haproxy -uroot -proot_password' logger.go:42: 14:48:48 | recreate/2-write-data | + sed -e 's/mysql: //' logger.go:42: 14:48:48 | recreate/2-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:48:50 | recreate/2-write-data | + : logger.go:42: 14:48:50 | recreate/2-write-data | + for i in 0 1 2 logger.go:42: 14:48:50 | recreate/2-write-data | +++ get_cluster_name logger.go:42: 14:48:50 | recreate/2-write-data | +++ kubectl -n kuttl-test-wired-lacewing get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:48:50 | recreate/2-write-data | ++ get_mysql_headless_fqdn recreate 0 logger.go:42: 14:48:50 | recreate/2-write-data | ++ local cluster=recreate logger.go:42: 14:48:50 | recreate/2-write-data | ++ local index=0 logger.go:42: 14:48:50 | recreate/2-write-data | ++ echo recreate-mysql-0.recreate-mysql logger.go:42: 14:48:50 | recreate/2-write-data | + host=recreate-mysql-0.recreate-mysql logger.go:42: 14:48:50 | recreate/2-write-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h recreate-mysql-0.recreate-mysql -uroot -proot_password' logger.go:42: 14:48:50 | recreate/2-write-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:48:50 | recreate/2-write-data | ++ local 'uri=-h recreate-mysql-0.recreate-mysql -uroot -proot_password' logger.go:42: 14:48:50 | recreate/2-write-data | ++ local pod= logger.go:42: 14:48:50 | recreate/2-write-data | +++ get_client_pod logger.go:42: 14:48:50 | recreate/2-write-data | +++ kubectl -n kuttl-test-wired-lacewing get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:48:50 | recreate/2-write-data | ++ client_pod=mysql-client logger.go:42: 14:48:50 | recreate/2-write-data | ++ wait_pod mysql-client logger.go:42: 14:48:50 | recreate/2-write-data | ++ local pod=mysql-client logger.go:42: 14:48:50 | recreate/2-write-data | ++ set +o xtrace logger.go:42: 14:48:51 | recreate/2-write-data | mysql-clienttrue logger.go:42: 14:48:51 | recreate/2-write-data | ++ kubectl -n kuttl-test-wired-lacewing exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h recreate-mysql-0.recreate-mysql -uroot -proot_password' logger.go:42: 14:48:51 | recreate/2-write-data | ++ sed -e 's/mysql: //' logger.go:42: 14:48:51 | recreate/2-write-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:48:52 | recreate/2-write-data | + data=100500 logger.go:42: 14:48:52 | recreate/2-write-data | + kubectl create configmap -n kuttl-test-wired-lacewing 02-write-data-0 --from-literal=data=100500 logger.go:42: 14:48:52 | recreate/2-write-data | configmap/02-write-data-0 created logger.go:42: 14:48:52 | recreate/2-write-data | + for i in 0 1 2 logger.go:42: 14:48:52 | recreate/2-write-data | +++ get_cluster_name logger.go:42: 14:48:52 | recreate/2-write-data | +++ kubectl -n kuttl-test-wired-lacewing get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:48:53 | recreate/2-write-data | ++ get_mysql_headless_fqdn recreate 1 logger.go:42: 14:48:53 | recreate/2-write-data | ++ local cluster=recreate logger.go:42: 14:48:53 | recreate/2-write-data | ++ local index=1 logger.go:42: 14:48:53 | recreate/2-write-data | ++ echo recreate-mysql-1.recreate-mysql logger.go:42: 14:48:53 | recreate/2-write-data | + host=recreate-mysql-1.recreate-mysql logger.go:42: 14:48:53 | recreate/2-write-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h recreate-mysql-1.recreate-mysql -uroot -proot_password' logger.go:42: 14:48:53 | recreate/2-write-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:48:53 | recreate/2-write-data | ++ local 'uri=-h recreate-mysql-1.recreate-mysql -uroot -proot_password' logger.go:42: 14:48:53 | recreate/2-write-data | ++ local pod= logger.go:42: 14:48:53 | recreate/2-write-data | +++ get_client_pod logger.go:42: 14:48:53 | recreate/2-write-data | +++ kubectl -n kuttl-test-wired-lacewing get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:48:53 | recreate/2-write-data | ++ client_pod=mysql-client logger.go:42: 14:48:53 | recreate/2-write-data | ++ wait_pod mysql-client logger.go:42: 14:48:53 | recreate/2-write-data | ++ local pod=mysql-client logger.go:42: 14:48:53 | recreate/2-write-data | ++ set +o xtrace logger.go:42: 14:48:54 | recreate/2-write-data | mysql-clienttrue logger.go:42: 14:48:54 | recreate/2-write-data | ++ sed -e 's/mysql: //' logger.go:42: 14:48:54 | recreate/2-write-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:48:54 | recreate/2-write-data | ++ kubectl -n kuttl-test-wired-lacewing exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h recreate-mysql-1.recreate-mysql -uroot -proot_password' logger.go:42: 14:48:55 | recreate/2-write-data | + data=100500 logger.go:42: 14:48:55 | recreate/2-write-data | + kubectl create configmap -n kuttl-test-wired-lacewing 02-write-data-1 --from-literal=data=100500 logger.go:42: 14:48:55 | recreate/2-write-data | configmap/02-write-data-1 created logger.go:42: 14:48:55 | recreate/2-write-data | + for i in 0 1 2 logger.go:42: 14:48:55 | recreate/2-write-data | +++ get_cluster_name logger.go:42: 14:48:55 | recreate/2-write-data | +++ kubectl -n kuttl-test-wired-lacewing get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:48:55 | recreate/2-write-data | ++ get_mysql_headless_fqdn recreate 2 logger.go:42: 14:48:55 | recreate/2-write-data | ++ local cluster=recreate logger.go:42: 14:48:55 | recreate/2-write-data | ++ local index=2 logger.go:42: 14:48:55 | recreate/2-write-data | ++ echo recreate-mysql-2.recreate-mysql logger.go:42: 14:48:55 | recreate/2-write-data | + host=recreate-mysql-2.recreate-mysql logger.go:42: 14:48:55 | recreate/2-write-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h recreate-mysql-2.recreate-mysql -uroot -proot_password' logger.go:42: 14:48:55 | recreate/2-write-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:48:55 | recreate/2-write-data | ++ local 'uri=-h recreate-mysql-2.recreate-mysql -uroot -proot_password' logger.go:42: 14:48:55 | recreate/2-write-data | ++ local pod= logger.go:42: 14:48:55 | recreate/2-write-data | +++ get_client_pod logger.go:42: 14:48:55 | recreate/2-write-data | +++ kubectl -n kuttl-test-wired-lacewing get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:48:56 | recreate/2-write-data | ++ client_pod=mysql-client logger.go:42: 14:48:56 | recreate/2-write-data | ++ wait_pod mysql-client logger.go:42: 14:48:56 | recreate/2-write-data | ++ local pod=mysql-client logger.go:42: 14:48:56 | recreate/2-write-data | ++ set +o xtrace logger.go:42: 14:48:56 | recreate/2-write-data | mysql-clienttrue logger.go:42: 14:48:56 | recreate/2-write-data | ++ sed -e 's/mysql: //' logger.go:42: 14:48:56 | recreate/2-write-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:48:56 | recreate/2-write-data | ++ kubectl -n kuttl-test-wired-lacewing exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h recreate-mysql-2.recreate-mysql -uroot -proot_password' logger.go:42: 14:48:57 | recreate/2-write-data | + data=100500 logger.go:42: 14:48:57 | recreate/2-write-data | + kubectl create configmap -n kuttl-test-wired-lacewing 02-write-data-2 --from-literal=data=100500 logger.go:42: 14:48:58 | recreate/2-write-data | configmap/02-write-data-2 created logger.go:42: 14:48:59 | recreate/2-write-data | test step completed 2-write-data logger.go:42: 14:48:59 | recreate/3-pause | starting test step 3-pause logger.go:42: 14:48:59 | recreate/3-pause | running command: [sh -c set -o errexit set -o xtrace source ../../functions get_cr \ | yq eval '.spec.pause=true' - \ | yq eval '.spec.mysql.clusterType="async"' - \ | yq eval '.spec.mysql.size=3' - \ | yq eval '.spec.proxy.haproxy.enabled=true' - \ | yq eval '.spec.proxy.haproxy.size=3' - \ | yq eval '.spec.orchestrator.enabled=true' - \ | yq eval '.spec.orchestrator.size=3' - \ | kubectl -n "${NAMESPACE}" apply -f -] logger.go:42: 14:48:59 | recreate/3-pause | + source ../../functions logger.go:42: 14:48:59 | recreate/3-pause | +++ realpath ../../.. logger.go:42: 14:48:59 | recreate/3-pause | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:48:59 | recreate/3-pause | ++++ pwd logger.go:42: 14:48:59 | recreate/3-pause | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/recreate logger.go:42: 14:48:59 | recreate/3-pause | ++ test_name=recreate logger.go:42: 14:48:59 | recreate/3-pause | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 14:48:59 | recreate/3-pause | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:48:59 | recreate/3-pause | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:48:59 | recreate/3-pause | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:48:59 | recreate/3-pause | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:48:59 | recreate/3-pause | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:48:59 | recreate/3-pause | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:48:59 | recreate/3-pause | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:48:59 | recreate/3-pause | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:48:59 | recreate/3-pause | +++ export TEMP_DIR=/tmp/kuttl/ps/recreate logger.go:42: 14:48:59 | recreate/3-pause | +++ TEMP_DIR=/tmp/kuttl/ps/recreate logger.go:42: 14:48:59 | recreate/3-pause | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:48:59 | recreate/3-pause | +++ export GIT_BRANCH=PR-523 logger.go:42: 14:48:59 | recreate/3-pause | +++ GIT_BRANCH=PR-523 logger.go:42: 14:48:59 | recreate/3-pause | +++ export VERSION=PR-523-f00253e logger.go:42: 14:48:59 | recreate/3-pause | +++ VERSION=PR-523-f00253e logger.go:42: 14:48:59 | recreate/3-pause | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:48:59 | recreate/3-pause | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:48:59 | recreate/3-pause | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:48:59 | recreate/3-pause | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:48:59 | recreate/3-pause | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:48:59 | recreate/3-pause | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:48:59 | recreate/3-pause | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:48:59 | recreate/3-pause | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:48:59 | recreate/3-pause | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:48:59 | recreate/3-pause | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:48:59 | recreate/3-pause | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:48:59 | recreate/3-pause | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:48:59 | recreate/3-pause | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:48:59 | recreate/3-pause | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:48:59 | recreate/3-pause | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:48:59 | recreate/3-pause | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:48:59 | recreate/3-pause | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:48:59 | recreate/3-pause | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:48:59 | recreate/3-pause | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:48:59 | recreate/3-pause | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:48:59 | recreate/3-pause | ++++ which gdate logger.go:42: 14:48:59 | recreate/3-pause | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:48:59 | recreate/3-pause | ++++ which date logger.go:42: 14:48:59 | recreate/3-pause | +++ date=/usr/bin/date logger.go:42: 14:48:59 | recreate/3-pause | +++ command -v oc logger.go:42: 14:48:59 | recreate/3-pause | +++ kubectl get nodes logger.go:42: 14:48:59 | recreate/3-pause | +++ grep '^minikube' logger.go:42: 14:49:00 | recreate/3-pause | + yq eval .spec.pause=true - logger.go:42: 14:49:00 | recreate/3-pause | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 14:49:00 | recreate/3-pause | + yq eval .spec.proxy.haproxy.enabled=true - logger.go:42: 14:49:00 | recreate/3-pause | + get_cr logger.go:42: 14:49:00 | recreate/3-pause | + local name_suffix= logger.go:42: 14:49:00 | recreate/3-pause | + yq eval .spec.proxy.haproxy.size=3 - logger.go:42: 14:49:00 | recreate/3-pause | + yq eval .spec.orchestrator.enabled=true - logger.go:42: 14:49:00 | recreate/3-pause | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 14:49:00 | recreate/3-pause | + yq eval .spec.orchestrator.size=3 - logger.go:42: 14:49:00 | recreate/3-pause | + kubectl -n kuttl-test-wired-lacewing apply -f - logger.go:42: 14:49:00 | recreate/3-pause | + yq eval .spec.mysql.size=3 - logger.go:42: 14:49:00 | recreate/3-pause | + yq eval '.spec.secretsName="test-secrets"' - logger.go:42: 14:49:00 | recreate/3-pause | + yq eval '.spec.sslSecretName="test-ssl"' - logger.go:42: 14:49:00 | recreate/3-pause | + yq eval '.spec.upgradeOptions.apply="disabled"' - logger.go:42: 14:49:00 | recreate/3-pause | ++ printf '.metadata.name="%s"' recreate logger.go:42: 14:49:00 | recreate/3-pause | + yq eval '.metadata.name="recreate"' /mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy/cr.yaml logger.go:42: 14:49:00 | recreate/3-pause | ++ printf '.spec.mysql.image="%s"' perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:49:00 | recreate/3-pause | + yq eval '.spec.mysql.image="perconalab/percona-server-mysql-operator:main-psmysql"' - logger.go:42: 14:49:00 | recreate/3-pause | ++ printf '.spec.initImage="%s"' perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:49:00 | recreate/3-pause | + yq eval '.spec.initImage="perconalab/percona-server-mysql-operator:PR-523-f00253e"' - logger.go:42: 14:49:00 | recreate/3-pause | ++ printf '.spec.backup.image="%s"' perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:49:00 | recreate/3-pause | + yq eval '.spec.backup.image="perconalab/percona-server-mysql-operator:main-backup"' - logger.go:42: 14:49:00 | recreate/3-pause | + '[' -n '' ']' logger.go:42: 14:49:00 | recreate/3-pause | ++ printf '.spec.pmm.image="%s"' perconalab/pmm-client:dev-latest logger.go:42: 14:49:00 | recreate/3-pause | + yq eval - logger.go:42: 14:49:00 | recreate/3-pause | + yq eval '.spec.pmm.image="perconalab/pmm-client:dev-latest"' - logger.go:42: 14:49:00 | recreate/3-pause | ++ printf '.spec.toolkit.image="%s"' perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:49:00 | recreate/3-pause | + yq eval '.spec.toolkit.image="perconalab/percona-server-mysql-operator:main-toolkit"' - logger.go:42: 14:49:00 | recreate/3-pause | ++ printf '.spec.proxy.haproxy.image="%s"' perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:49:00 | recreate/3-pause | + yq eval '.spec.proxy.haproxy.image="perconalab/percona-server-mysql-operator:main-haproxy"' - logger.go:42: 14:49:00 | recreate/3-pause | ++ printf '.spec.proxy.router.image="%s"' perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:49:00 | recreate/3-pause | ++ printf '.spec.orchestrator.image="%s"' perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:49:00 | recreate/3-pause | + yq eval '.spec.proxy.router.image="perconalab/percona-server-mysql-operator:main-router"' - logger.go:42: 14:49:00 | recreate/3-pause | + yq eval '.spec.orchestrator.image="perconalab/percona-server-mysql-operator:main-orchestrator"' - logger.go:42: 14:49:03 | recreate/3-pause | perconaservermysql.ps.percona.com/recreate configured logger.go:42: 14:50:46 | recreate/3-pause | test step completed 3-pause logger.go:42: 14:50:46 | recreate/4-unpause | starting test step 4-unpause logger.go:42: 14:50:46 | recreate/4-unpause | running command: [sh -c set -o errexit set -o xtrace source ../../functions get_cr \ | yq eval '.spec.pause=false' - \ | yq eval '.spec.mysql.clusterType="async"' - \ | yq eval '.spec.mysql.size=3' - \ | yq eval '.spec.proxy.haproxy.enabled=true' - \ | yq eval '.spec.proxy.haproxy.size=3' - \ | yq eval '.spec.orchestrator.enabled=true' - \ | yq eval '.spec.orchestrator.size=3' - \ | kubectl -n "${NAMESPACE}" apply -f -] logger.go:42: 14:50:46 | recreate/4-unpause | + source ../../functions logger.go:42: 14:50:46 | recreate/4-unpause | +++ realpath ../../.. logger.go:42: 14:50:46 | recreate/4-unpause | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:50:46 | recreate/4-unpause | ++++ pwd logger.go:42: 14:50:46 | recreate/4-unpause | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/recreate logger.go:42: 14:50:46 | recreate/4-unpause | ++ test_name=recreate logger.go:42: 14:50:46 | recreate/4-unpause | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 14:50:46 | recreate/4-unpause | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:50:46 | recreate/4-unpause | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:50:46 | recreate/4-unpause | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:50:46 | recreate/4-unpause | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:50:46 | recreate/4-unpause | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:50:46 | recreate/4-unpause | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:50:46 | recreate/4-unpause | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:50:46 | recreate/4-unpause | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:50:46 | recreate/4-unpause | +++ export TEMP_DIR=/tmp/kuttl/ps/recreate logger.go:42: 14:50:46 | recreate/4-unpause | +++ TEMP_DIR=/tmp/kuttl/ps/recreate logger.go:42: 14:50:46 | recreate/4-unpause | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:50:46 | recreate/4-unpause | +++ export GIT_BRANCH=PR-523 logger.go:42: 14:50:46 | recreate/4-unpause | +++ GIT_BRANCH=PR-523 logger.go:42: 14:50:46 | recreate/4-unpause | +++ export VERSION=PR-523-f00253e logger.go:42: 14:50:46 | recreate/4-unpause | +++ VERSION=PR-523-f00253e logger.go:42: 14:50:46 | recreate/4-unpause | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:50:46 | recreate/4-unpause | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:50:46 | recreate/4-unpause | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:50:46 | recreate/4-unpause | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:50:46 | recreate/4-unpause | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:50:46 | recreate/4-unpause | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:50:46 | recreate/4-unpause | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:50:46 | recreate/4-unpause | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:50:46 | recreate/4-unpause | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:50:46 | recreate/4-unpause | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:50:46 | recreate/4-unpause | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:50:46 | recreate/4-unpause | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:50:46 | recreate/4-unpause | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:50:46 | recreate/4-unpause | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:50:46 | recreate/4-unpause | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:50:46 | recreate/4-unpause | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:50:46 | recreate/4-unpause | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:50:46 | recreate/4-unpause | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:50:46 | recreate/4-unpause | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:50:46 | recreate/4-unpause | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:50:46 | recreate/4-unpause | ++++ which gdate logger.go:42: 14:50:46 | recreate/4-unpause | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:50:46 | recreate/4-unpause | ++++ which date logger.go:42: 14:50:46 | recreate/4-unpause | +++ date=/usr/bin/date logger.go:42: 14:50:46 | recreate/4-unpause | +++ command -v oc logger.go:42: 14:50:46 | recreate/4-unpause | +++ kubectl get nodes logger.go:42: 14:50:46 | recreate/4-unpause | +++ grep '^minikube' logger.go:42: 14:50:47 | recreate/4-unpause | + get_cr logger.go:42: 14:50:47 | recreate/4-unpause | + local name_suffix= logger.go:42: 14:50:47 | recreate/4-unpause | + yq eval .spec.pause=false - logger.go:42: 14:50:47 | recreate/4-unpause | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 14:50:47 | recreate/4-unpause | + yq eval .spec.orchestrator.size=3 - logger.go:42: 14:50:47 | recreate/4-unpause | + yq eval .spec.proxy.haproxy.size=3 - logger.go:42: 14:50:47 | recreate/4-unpause | + kubectl -n kuttl-test-wired-lacewing apply -f - logger.go:42: 14:50:47 | recreate/4-unpause | + yq eval .spec.mysql.size=3 - logger.go:42: 14:50:47 | recreate/4-unpause | + yq eval .spec.orchestrator.enabled=true - logger.go:42: 14:50:47 | recreate/4-unpause | + yq eval '.spec.secretsName="test-secrets"' - logger.go:42: 14:50:47 | recreate/4-unpause | + yq eval .spec.proxy.haproxy.enabled=true - logger.go:42: 14:50:47 | recreate/4-unpause | + yq eval '.spec.upgradeOptions.apply="disabled"' - logger.go:42: 14:50:47 | recreate/4-unpause | + '[' -n '' ']' logger.go:42: 14:50:47 | recreate/4-unpause | + yq eval - logger.go:42: 14:50:47 | recreate/4-unpause | ++ printf '.spec.backup.image="%s"' perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:50:47 | recreate/4-unpause | + yq eval '.spec.backup.image="perconalab/percona-server-mysql-operator:main-backup"' - logger.go:42: 14:50:47 | recreate/4-unpause | ++ printf '.spec.proxy.haproxy.image="%s"' perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:50:47 | recreate/4-unpause | + yq eval '.spec.proxy.haproxy.image="perconalab/percona-server-mysql-operator:main-haproxy"' - logger.go:42: 14:50:47 | recreate/4-unpause | + yq eval '.spec.sslSecretName="test-ssl"' - logger.go:42: 14:50:47 | recreate/4-unpause | ++ printf '.spec.initImage="%s"' perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:50:47 | recreate/4-unpause | + yq eval '.spec.initImage="perconalab/percona-server-mysql-operator:PR-523-f00253e"' - logger.go:42: 14:50:47 | recreate/4-unpause | ++ printf '.spec.proxy.router.image="%s"' perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:50:47 | recreate/4-unpause | + yq eval '.spec.proxy.router.image="perconalab/percona-server-mysql-operator:main-router"' - logger.go:42: 14:50:47 | recreate/4-unpause | ++ printf '.metadata.name="%s"' recreate logger.go:42: 14:50:47 | recreate/4-unpause | ++ printf '.spec.pmm.image="%s"' perconalab/pmm-client:dev-latest logger.go:42: 14:50:47 | recreate/4-unpause | + yq eval '.spec.pmm.image="perconalab/pmm-client:dev-latest"' - logger.go:42: 14:50:47 | recreate/4-unpause | ++ printf '.spec.mysql.image="%s"' perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:50:47 | recreate/4-unpause | + yq eval '.spec.mysql.image="perconalab/percona-server-mysql-operator:main-psmysql"' - logger.go:42: 14:50:47 | recreate/4-unpause | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 14:50:47 | recreate/4-unpause | + yq eval '.metadata.name="recreate"' /mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy/cr.yaml logger.go:42: 14:50:47 | recreate/4-unpause | ++ printf '.spec.orchestrator.image="%s"' perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:50:47 | recreate/4-unpause | + yq eval '.spec.orchestrator.image="perconalab/percona-server-mysql-operator:main-orchestrator"' - logger.go:42: 14:50:47 | recreate/4-unpause | ++ printf '.spec.toolkit.image="%s"' perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:50:47 | recreate/4-unpause | + yq eval '.spec.toolkit.image="perconalab/percona-server-mysql-operator:main-toolkit"' - logger.go:42: 14:50:48 | recreate/4-unpause | perconaservermysql.ps.percona.com/recreate configured logger.go:42: 14:53:52 | recreate/4-unpause | test step completed 4-unpause logger.go:42: 14:53:52 | recreate/5-write-data | starting test step 5-write-data logger.go:42: 14:53:52 | recreate/5-write-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "INSERT myDB.myTable (id) VALUES (100501)" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password" for i in 0 1 2; do host=$(get_mysql_headless_fqdn $(get_cluster_name) $i) data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${host} -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 06-write-data-${i} --from-literal=data="${data}" done] logger.go:42: 14:53:52 | recreate/5-write-data | + source ../../functions logger.go:42: 14:53:52 | recreate/5-write-data | +++ realpath ../../.. logger.go:42: 14:53:52 | recreate/5-write-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:53:52 | recreate/5-write-data | ++++ pwd logger.go:42: 14:53:52 | recreate/5-write-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/recreate logger.go:42: 14:53:52 | recreate/5-write-data | ++ test_name=recreate logger.go:42: 14:53:52 | recreate/5-write-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 14:53:52 | recreate/5-write-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:53:52 | recreate/5-write-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:53:52 | recreate/5-write-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:53:52 | recreate/5-write-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:53:52 | recreate/5-write-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:53:52 | recreate/5-write-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:53:52 | recreate/5-write-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:53:52 | recreate/5-write-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:53:52 | recreate/5-write-data | +++ export TEMP_DIR=/tmp/kuttl/ps/recreate logger.go:42: 14:53:52 | recreate/5-write-data | +++ TEMP_DIR=/tmp/kuttl/ps/recreate logger.go:42: 14:53:52 | recreate/5-write-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:53:52 | recreate/5-write-data | +++ export GIT_BRANCH=PR-523 logger.go:42: 14:53:52 | recreate/5-write-data | +++ GIT_BRANCH=PR-523 logger.go:42: 14:53:52 | recreate/5-write-data | +++ export VERSION=PR-523-f00253e logger.go:42: 14:53:52 | recreate/5-write-data | +++ VERSION=PR-523-f00253e logger.go:42: 14:53:52 | recreate/5-write-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:53:52 | recreate/5-write-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:53:52 | recreate/5-write-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:53:52 | recreate/5-write-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:53:52 | recreate/5-write-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:53:52 | recreate/5-write-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:53:52 | recreate/5-write-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:53:52 | recreate/5-write-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:53:52 | recreate/5-write-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:53:52 | recreate/5-write-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:53:52 | recreate/5-write-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:53:52 | recreate/5-write-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:53:52 | recreate/5-write-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:53:52 | recreate/5-write-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:53:52 | recreate/5-write-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:53:52 | recreate/5-write-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:53:52 | recreate/5-write-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:53:52 | recreate/5-write-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:53:52 | recreate/5-write-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:53:52 | recreate/5-write-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:53:52 | recreate/5-write-data | ++++ which gdate logger.go:42: 14:53:52 | recreate/5-write-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:53:52 | recreate/5-write-data | ++++ which date logger.go:42: 14:53:52 | recreate/5-write-data | +++ date=/usr/bin/date logger.go:42: 14:53:52 | recreate/5-write-data | +++ command -v oc logger.go:42: 14:53:52 | recreate/5-write-data | +++ kubectl get nodes logger.go:42: 14:53:52 | recreate/5-write-data | +++ grep '^minikube' logger.go:42: 14:53:54 | recreate/5-write-data | +++ get_cluster_name logger.go:42: 14:53:54 | recreate/5-write-data | +++ kubectl -n kuttl-test-wired-lacewing get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:53:54 | recreate/5-write-data | ++ get_haproxy_svc recreate logger.go:42: 14:53:54 | recreate/5-write-data | ++ local cluster=recreate logger.go:42: 14:53:54 | recreate/5-write-data | ++ echo recreate-haproxy logger.go:42: 14:53:54 | recreate/5-write-data | + run_mysql 'INSERT myDB.myTable (id) VALUES (100501)' '-h recreate-haproxy -uroot -proot_password' logger.go:42: 14:53:54 | recreate/5-write-data | + local 'command=INSERT myDB.myTable (id) VALUES (100501)' logger.go:42: 14:53:54 | recreate/5-write-data | + local 'uri=-h recreate-haproxy -uroot -proot_password' logger.go:42: 14:53:54 | recreate/5-write-data | + local pod= logger.go:42: 14:53:54 | recreate/5-write-data | ++ get_client_pod logger.go:42: 14:53:54 | recreate/5-write-data | ++ kubectl -n kuttl-test-wired-lacewing get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:53:54 | recreate/5-write-data | + client_pod=mysql-client logger.go:42: 14:53:54 | recreate/5-write-data | + wait_pod mysql-client logger.go:42: 14:53:54 | recreate/5-write-data | + local pod=mysql-client logger.go:42: 14:53:54 | recreate/5-write-data | + set +o xtrace logger.go:42: 14:53:55 | recreate/5-write-data | mysql-clienttrue logger.go:42: 14:53:55 | recreate/5-write-data | + kubectl -n kuttl-test-wired-lacewing exec mysql-client -- bash -c 'printf '\''%s\n'\'' "INSERT myDB.myTable (id) VALUES (100501)" | mysql -sN -h recreate-haproxy -uroot -proot_password' logger.go:42: 14:53:55 | recreate/5-write-data | + sed -e 's/mysql: //' logger.go:42: 14:53:55 | recreate/5-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:53:56 | recreate/5-write-data | + : logger.go:42: 14:53:56 | recreate/5-write-data | + for i in 0 1 2 logger.go:42: 14:53:56 | recreate/5-write-data | +++ get_cluster_name logger.go:42: 14:53:56 | recreate/5-write-data | +++ kubectl -n kuttl-test-wired-lacewing get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:53:56 | recreate/5-write-data | ++ get_mysql_headless_fqdn recreate 0 logger.go:42: 14:53:56 | recreate/5-write-data | ++ local cluster=recreate logger.go:42: 14:53:56 | recreate/5-write-data | ++ local index=0 logger.go:42: 14:53:56 | recreate/5-write-data | ++ echo recreate-mysql-0.recreate-mysql logger.go:42: 14:53:56 | recreate/5-write-data | + host=recreate-mysql-0.recreate-mysql logger.go:42: 14:53:56 | recreate/5-write-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h recreate-mysql-0.recreate-mysql -uroot -proot_password' logger.go:42: 14:53:56 | recreate/5-write-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:53:56 | recreate/5-write-data | ++ local 'uri=-h recreate-mysql-0.recreate-mysql -uroot -proot_password' logger.go:42: 14:53:56 | recreate/5-write-data | ++ local pod= logger.go:42: 14:53:56 | recreate/5-write-data | +++ get_client_pod logger.go:42: 14:53:56 | recreate/5-write-data | +++ kubectl -n kuttl-test-wired-lacewing get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:53:57 | recreate/5-write-data | ++ client_pod=mysql-client logger.go:42: 14:53:57 | recreate/5-write-data | ++ wait_pod mysql-client logger.go:42: 14:53:57 | recreate/5-write-data | ++ local pod=mysql-client logger.go:42: 14:53:57 | recreate/5-write-data | ++ set +o xtrace logger.go:42: 14:53:57 | recreate/5-write-data | mysql-clienttrue logger.go:42: 14:53:57 | recreate/5-write-data | ++ kubectl -n kuttl-test-wired-lacewing exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h recreate-mysql-0.recreate-mysql -uroot -proot_password' logger.go:42: 14:53:57 | recreate/5-write-data | ++ sed -e 's/mysql: //' logger.go:42: 14:53:57 | recreate/5-write-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:53:58 | recreate/5-write-data | + data='100500 logger.go:42: 14:53:58 | recreate/5-write-data | 100501' logger.go:42: 14:53:58 | recreate/5-write-data | + kubectl create configmap -n kuttl-test-wired-lacewing 06-write-data-0 '--from-literal=data=100500 logger.go:42: 14:53:58 | recreate/5-write-data | 100501' logger.go:42: 14:53:59 | recreate/5-write-data | configmap/06-write-data-0 created logger.go:42: 14:53:59 | recreate/5-write-data | + for i in 0 1 2 logger.go:42: 14:53:59 | recreate/5-write-data | +++ get_cluster_name logger.go:42: 14:53:59 | recreate/5-write-data | +++ kubectl -n kuttl-test-wired-lacewing get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:53:59 | recreate/5-write-data | ++ get_mysql_headless_fqdn recreate 1 logger.go:42: 14:53:59 | recreate/5-write-data | ++ local cluster=recreate logger.go:42: 14:53:59 | recreate/5-write-data | ++ local index=1 logger.go:42: 14:53:59 | recreate/5-write-data | ++ echo recreate-mysql-1.recreate-mysql logger.go:42: 14:53:59 | recreate/5-write-data | + host=recreate-mysql-1.recreate-mysql logger.go:42: 14:53:59 | recreate/5-write-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h recreate-mysql-1.recreate-mysql -uroot -proot_password' logger.go:42: 14:53:59 | recreate/5-write-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:53:59 | recreate/5-write-data | ++ local 'uri=-h recreate-mysql-1.recreate-mysql -uroot -proot_password' logger.go:42: 14:53:59 | recreate/5-write-data | ++ local pod= logger.go:42: 14:53:59 | recreate/5-write-data | +++ get_client_pod logger.go:42: 14:53:59 | recreate/5-write-data | +++ kubectl -n kuttl-test-wired-lacewing get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:54:00 | recreate/5-write-data | ++ client_pod=mysql-client logger.go:42: 14:54:00 | recreate/5-write-data | ++ wait_pod mysql-client logger.go:42: 14:54:00 | recreate/5-write-data | ++ local pod=mysql-client logger.go:42: 14:54:00 | recreate/5-write-data | ++ set +o xtrace logger.go:42: 14:54:00 | recreate/5-write-data | mysql-clienttrue logger.go:42: 14:54:00 | recreate/5-write-data | ++ kubectl -n kuttl-test-wired-lacewing exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h recreate-mysql-1.recreate-mysql -uroot -proot_password' logger.go:42: 14:54:00 | recreate/5-write-data | ++ sed -e 's/mysql: //' logger.go:42: 14:54:00 | recreate/5-write-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:54:01 | recreate/5-write-data | + data='100500 logger.go:42: 14:54:01 | recreate/5-write-data | 100501' logger.go:42: 14:54:01 | recreate/5-write-data | + kubectl create configmap -n kuttl-test-wired-lacewing 06-write-data-1 '--from-literal=data=100500 logger.go:42: 14:54:01 | recreate/5-write-data | 100501' logger.go:42: 14:54:01 | recreate/5-write-data | configmap/06-write-data-1 created logger.go:42: 14:54:01 | recreate/5-write-data | + for i in 0 1 2 logger.go:42: 14:54:01 | recreate/5-write-data | +++ get_cluster_name logger.go:42: 14:54:01 | recreate/5-write-data | +++ kubectl -n kuttl-test-wired-lacewing get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:54:02 | recreate/5-write-data | ++ get_mysql_headless_fqdn recreate 2 logger.go:42: 14:54:02 | recreate/5-write-data | ++ local cluster=recreate logger.go:42: 14:54:02 | recreate/5-write-data | ++ local index=2 logger.go:42: 14:54:02 | recreate/5-write-data | ++ echo recreate-mysql-2.recreate-mysql logger.go:42: 14:54:02 | recreate/5-write-data | + host=recreate-mysql-2.recreate-mysql logger.go:42: 14:54:02 | recreate/5-write-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h recreate-mysql-2.recreate-mysql -uroot -proot_password' logger.go:42: 14:54:02 | recreate/5-write-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:54:02 | recreate/5-write-data | ++ local 'uri=-h recreate-mysql-2.recreate-mysql -uroot -proot_password' logger.go:42: 14:54:02 | recreate/5-write-data | ++ local pod= logger.go:42: 14:54:02 | recreate/5-write-data | +++ get_client_pod logger.go:42: 14:54:02 | recreate/5-write-data | +++ kubectl -n kuttl-test-wired-lacewing get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:54:02 | recreate/5-write-data | ++ client_pod=mysql-client logger.go:42: 14:54:02 | recreate/5-write-data | ++ wait_pod mysql-client logger.go:42: 14:54:02 | recreate/5-write-data | ++ local pod=mysql-client logger.go:42: 14:54:02 | recreate/5-write-data | ++ set +o xtrace logger.go:42: 14:54:03 | recreate/5-write-data | mysql-clienttrue logger.go:42: 14:54:03 | recreate/5-write-data | ++ sed -e 's/mysql: //' logger.go:42: 14:54:03 | recreate/5-write-data | ++ kubectl -n kuttl-test-wired-lacewing exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h recreate-mysql-2.recreate-mysql -uroot -proot_password' logger.go:42: 14:54:03 | recreate/5-write-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:54:04 | recreate/5-write-data | + data='100500 logger.go:42: 14:54:04 | recreate/5-write-data | 100501' logger.go:42: 14:54:04 | recreate/5-write-data | + kubectl create configmap -n kuttl-test-wired-lacewing 06-write-data-2 '--from-literal=data=100500 logger.go:42: 14:54:04 | recreate/5-write-data | 100501' logger.go:42: 14:54:04 | recreate/5-write-data | configmap/06-write-data-2 created logger.go:42: 14:54:06 | recreate/5-write-data | test step completed 5-write-data logger.go:42: 14:54:06 | recreate/7-delete-cluster | starting test step 7-delete-cluster logger.go:42: 14:54:06 | recreate/7-delete-cluster | running command: [sh -c set -o errexit set -o xtrace source ../../functions kubectl delete ps -n ${NAMESPACE} recreate] logger.go:42: 14:54:06 | recreate/7-delete-cluster | + source ../../functions logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ realpath ../../.. logger.go:42: 14:54:06 | recreate/7-delete-cluster | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:54:06 | recreate/7-delete-cluster | ++++ pwd logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/recreate logger.go:42: 14:54:06 | recreate/7-delete-cluster | ++ test_name=recreate logger.go:42: 14:54:06 | recreate/7-delete-cluster | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ export TEMP_DIR=/tmp/kuttl/ps/recreate logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ TEMP_DIR=/tmp/kuttl/ps/recreate logger.go:42: 14:54:06 | recreate/7-delete-cluster | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ export GIT_BRANCH=PR-523 logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ GIT_BRANCH=PR-523 logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ export VERSION=PR-523-f00253e logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ VERSION=PR-523-f00253e logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:54:06 | recreate/7-delete-cluster | ++++ which gdate logger.go:42: 14:54:06 | recreate/7-delete-cluster | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:54:06 | recreate/7-delete-cluster | ++++ which date logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ date=/usr/bin/date logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ command -v oc logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ kubectl get nodes logger.go:42: 14:54:06 | recreate/7-delete-cluster | +++ grep '^minikube' logger.go:42: 14:54:06 | recreate/7-delete-cluster | + kubectl delete ps -n kuttl-test-wired-lacewing recreate logger.go:42: 14:54:07 | recreate/7-delete-cluster | perconaservermysql.ps.percona.com "recreate" deleted logger.go:42: 14:54:25 | recreate/7-delete-cluster | test step completed 7-delete-cluster logger.go:42: 14:54:25 | recreate/8- | starting test step 8- logger.go:42: 14:54:27 | recreate/8- | test step completed 8- logger.go:42: 14:54:27 | recreate/9-recreate-cluster | starting test step 9-recreate-cluster logger.go:42: 14:54:27 | recreate/9-recreate-cluster | running command: [sh -c set -o errexit set -o xtrace source ../../functions get_cr \ | yq eval '.spec.pause=false' - \ | yq eval '.spec.mysql.clusterType="async"' - \ | yq eval '.spec.mysql.size=3' - \ | yq eval '.spec.proxy.haproxy.enabled=true' - \ | yq eval '.spec.proxy.haproxy.size=3' - \ | yq eval '.spec.orchestrator.enabled=true' - \ | yq eval '.spec.orchestrator.size=3' - \ | kubectl -n "${NAMESPACE}" apply -f -] logger.go:42: 14:54:27 | recreate/9-recreate-cluster | + source ../../functions logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ realpath ../../.. logger.go:42: 14:54:27 | recreate/9-recreate-cluster | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:54:27 | recreate/9-recreate-cluster | ++++ pwd logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/recreate logger.go:42: 14:54:27 | recreate/9-recreate-cluster | ++ test_name=recreate logger.go:42: 14:54:27 | recreate/9-recreate-cluster | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ export TEMP_DIR=/tmp/kuttl/ps/recreate logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ TEMP_DIR=/tmp/kuttl/ps/recreate logger.go:42: 14:54:27 | recreate/9-recreate-cluster | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ export GIT_BRANCH=PR-523 logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ GIT_BRANCH=PR-523 logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ export VERSION=PR-523-f00253e logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ VERSION=PR-523-f00253e logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:54:27 | recreate/9-recreate-cluster | ++++ which gdate logger.go:42: 14:54:27 | recreate/9-recreate-cluster | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:54:27 | recreate/9-recreate-cluster | ++++ which date logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ date=/usr/bin/date logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ command -v oc logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ kubectl get nodes logger.go:42: 14:54:27 | recreate/9-recreate-cluster | +++ grep '^minikube' logger.go:42: 14:54:28 | recreate/9-recreate-cluster | + yq eval .spec.pause=false - logger.go:42: 14:54:28 | recreate/9-recreate-cluster | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 14:54:28 | recreate/9-recreate-cluster | + yq eval .spec.proxy.haproxy.enabled=true - logger.go:42: 14:54:28 | recreate/9-recreate-cluster | + get_cr logger.go:42: 14:54:28 | recreate/9-recreate-cluster | + local name_suffix= logger.go:42: 14:54:28 | recreate/9-recreate-cluster | + '[' -n '' ']' logger.go:42: 14:54:28 | recreate/9-recreate-cluster | ++ printf '.spec.pmm.image="%s"' perconalab/pmm-client:dev-latest logger.go:42: 14:54:28 | recreate/9-recreate-cluster | + yq eval - logger.go:42: 14:54:28 | recreate/9-recreate-cluster | + yq eval '.spec.pmm.image="perconalab/pmm-client:dev-latest"' - logger.go:42: 14:54:28 | recreate/9-recreate-cluster | ++ printf '.spec.proxy.haproxy.image="%s"' perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:54:28 | recreate/9-recreate-cluster | + yq eval '.spec.proxy.haproxy.image="perconalab/percona-server-mysql-operator:main-haproxy"' - logger.go:42: 14:54:28 | recreate/9-recreate-cluster | + yq eval .spec.proxy.haproxy.size=3 - logger.go:42: 14:54:28 | recreate/9-recreate-cluster | + yq eval .spec.mysql.size=3 - logger.go:42: 14:54:28 | recreate/9-recreate-cluster | + yq eval .spec.orchestrator.enabled=true - logger.go:42: 14:54:28 | recreate/9-recreate-cluster | + yq eval .spec.orchestrator.size=3 - logger.go:42: 14:54:28 | recreate/9-recreate-cluster | + kubectl -n kuttl-test-wired-lacewing apply -f - logger.go:42: 14:54:28 | recreate/9-recreate-cluster | + yq eval '.spec.secretsName="test-secrets"' - logger.go:42: 14:54:28 | recreate/9-recreate-cluster | + yq eval '.spec.sslSecretName="test-ssl"' - logger.go:42: 14:54:28 | recreate/9-recreate-cluster | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 14:54:28 | recreate/9-recreate-cluster | + yq eval '.spec.upgradeOptions.apply="disabled"' - logger.go:42: 14:54:28 | recreate/9-recreate-cluster | ++ printf '.spec.mysql.image="%s"' perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:54:28 | recreate/9-recreate-cluster | ++ printf '.spec.proxy.router.image="%s"' perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:54:28 | recreate/9-recreate-cluster | ++ printf '.spec.orchestrator.image="%s"' perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:54:28 | recreate/9-recreate-cluster | + yq eval '.spec.proxy.router.image="perconalab/percona-server-mysql-operator:main-router"' - logger.go:42: 14:54:28 | recreate/9-recreate-cluster | + yq eval '.spec.mysql.image="perconalab/percona-server-mysql-operator:main-psmysql"' - logger.go:42: 14:54:28 | recreate/9-recreate-cluster | ++ printf '.metadata.name="%s"' recreate logger.go:42: 14:54:28 | recreate/9-recreate-cluster | + yq eval '.spec.orchestrator.image="perconalab/percona-server-mysql-operator:main-orchestrator"' - logger.go:42: 14:54:28 | recreate/9-recreate-cluster | + yq eval '.metadata.name="recreate"' /mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy/cr.yaml logger.go:42: 14:54:28 | recreate/9-recreate-cluster | ++ printf '.spec.initImage="%s"' perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:54:28 | recreate/9-recreate-cluster | + yq eval '.spec.initImage="perconalab/percona-server-mysql-operator:PR-523-f00253e"' - logger.go:42: 14:54:28 | recreate/9-recreate-cluster | ++ printf '.spec.toolkit.image="%s"' perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:54:28 | recreate/9-recreate-cluster | + yq eval '.spec.toolkit.image="perconalab/percona-server-mysql-operator:main-toolkit"' - logger.go:42: 14:54:28 | recreate/9-recreate-cluster | ++ printf '.spec.backup.image="%s"' perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:54:28 | recreate/9-recreate-cluster | + yq eval '.spec.backup.image="perconalab/percona-server-mysql-operator:main-backup"' - logger.go:42: 14:54:29 | recreate/9-recreate-cluster | perconaservermysql.ps.percona.com/recreate created logger.go:42: 14:57:36 | recreate/9-recreate-cluster | test step completed 9-recreate-cluster logger.go:42: 14:57:36 | recreate/10-write-data | starting test step 10-write-data logger.go:42: 14:57:36 | recreate/10-write-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "INSERT myDB.myTable (id) VALUES (100502)" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password" for i in 0 1 2; do host=$(get_mysql_headless_fqdn $(get_cluster_name) $i) data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${host} -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 11-write-data-${i} --from-literal=data="${data}" done] logger.go:42: 14:57:36 | recreate/10-write-data | + source ../../functions logger.go:42: 14:57:36 | recreate/10-write-data | +++ realpath ../../.. logger.go:42: 14:57:36 | recreate/10-write-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:57:36 | recreate/10-write-data | ++++ pwd logger.go:42: 14:57:36 | recreate/10-write-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/recreate logger.go:42: 14:57:36 | recreate/10-write-data | ++ test_name=recreate logger.go:42: 14:57:36 | recreate/10-write-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 14:57:36 | recreate/10-write-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:57:36 | recreate/10-write-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 14:57:36 | recreate/10-write-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:57:36 | recreate/10-write-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 14:57:36 | recreate/10-write-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:57:36 | recreate/10-write-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 14:57:36 | recreate/10-write-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:57:36 | recreate/10-write-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 14:57:36 | recreate/10-write-data | +++ export TEMP_DIR=/tmp/kuttl/ps/recreate logger.go:42: 14:57:36 | recreate/10-write-data | +++ TEMP_DIR=/tmp/kuttl/ps/recreate logger.go:42: 14:57:36 | recreate/10-write-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 14:57:36 | recreate/10-write-data | +++ export GIT_BRANCH=PR-523 logger.go:42: 14:57:36 | recreate/10-write-data | +++ GIT_BRANCH=PR-523 logger.go:42: 14:57:36 | recreate/10-write-data | +++ export VERSION=PR-523-f00253e logger.go:42: 14:57:36 | recreate/10-write-data | +++ VERSION=PR-523-f00253e logger.go:42: 14:57:36 | recreate/10-write-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:57:36 | recreate/10-write-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 14:57:36 | recreate/10-write-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:57:36 | recreate/10-write-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 14:57:36 | recreate/10-write-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:57:36 | recreate/10-write-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 14:57:36 | recreate/10-write-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:57:36 | recreate/10-write-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 14:57:36 | recreate/10-write-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:57:36 | recreate/10-write-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 14:57:36 | recreate/10-write-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:57:36 | recreate/10-write-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 14:57:36 | recreate/10-write-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:57:36 | recreate/10-write-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 14:57:36 | recreate/10-write-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:57:36 | recreate/10-write-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 14:57:36 | recreate/10-write-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:57:36 | recreate/10-write-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 14:57:36 | recreate/10-write-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:57:36 | recreate/10-write-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 14:57:36 | recreate/10-write-data | ++++ which gdate logger.go:42: 14:57:36 | recreate/10-write-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 14:57:36 | recreate/10-write-data | ++++ which date logger.go:42: 14:57:36 | recreate/10-write-data | +++ date=/usr/bin/date logger.go:42: 14:57:36 | recreate/10-write-data | +++ command -v oc logger.go:42: 14:57:36 | recreate/10-write-data | +++ kubectl get nodes logger.go:42: 14:57:36 | recreate/10-write-data | +++ grep '^minikube' logger.go:42: 14:57:37 | recreate/10-write-data | +++ get_cluster_name logger.go:42: 14:57:37 | recreate/10-write-data | +++ kubectl -n kuttl-test-wired-lacewing get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:57:37 | recreate/10-write-data | ++ get_haproxy_svc recreate logger.go:42: 14:57:37 | recreate/10-write-data | ++ local cluster=recreate logger.go:42: 14:57:37 | recreate/10-write-data | ++ echo recreate-haproxy logger.go:42: 14:57:37 | recreate/10-write-data | + run_mysql 'INSERT myDB.myTable (id) VALUES (100502)' '-h recreate-haproxy -uroot -proot_password' logger.go:42: 14:57:37 | recreate/10-write-data | + local 'command=INSERT myDB.myTable (id) VALUES (100502)' logger.go:42: 14:57:37 | recreate/10-write-data | + local 'uri=-h recreate-haproxy -uroot -proot_password' logger.go:42: 14:57:37 | recreate/10-write-data | + local pod= logger.go:42: 14:57:37 | recreate/10-write-data | ++ get_client_pod logger.go:42: 14:57:37 | recreate/10-write-data | ++ kubectl -n kuttl-test-wired-lacewing get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:57:37 | recreate/10-write-data | + client_pod=mysql-client logger.go:42: 14:57:37 | recreate/10-write-data | + wait_pod mysql-client logger.go:42: 14:57:37 | recreate/10-write-data | + local pod=mysql-client logger.go:42: 14:57:37 | recreate/10-write-data | + set +o xtrace logger.go:42: 14:57:38 | recreate/10-write-data | mysql-clienttrue logger.go:42: 14:57:38 | recreate/10-write-data | + kubectl -n kuttl-test-wired-lacewing exec mysql-client -- bash -c 'printf '\''%s\n'\'' "INSERT myDB.myTable (id) VALUES (100502)" | mysql -sN -h recreate-haproxy -uroot -proot_password' logger.go:42: 14:57:38 | recreate/10-write-data | + sed -e 's/mysql: //' logger.go:42: 14:57:38 | recreate/10-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:57:39 | recreate/10-write-data | + : logger.go:42: 14:57:39 | recreate/10-write-data | + for i in 0 1 2 logger.go:42: 14:57:39 | recreate/10-write-data | +++ get_cluster_name logger.go:42: 14:57:39 | recreate/10-write-data | +++ kubectl -n kuttl-test-wired-lacewing get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:57:39 | recreate/10-write-data | ++ get_mysql_headless_fqdn recreate 0 logger.go:42: 14:57:39 | recreate/10-write-data | ++ local cluster=recreate logger.go:42: 14:57:39 | recreate/10-write-data | ++ local index=0 logger.go:42: 14:57:39 | recreate/10-write-data | ++ echo recreate-mysql-0.recreate-mysql logger.go:42: 14:57:39 | recreate/10-write-data | + host=recreate-mysql-0.recreate-mysql logger.go:42: 14:57:39 | recreate/10-write-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h recreate-mysql-0.recreate-mysql -uroot -proot_password' logger.go:42: 14:57:39 | recreate/10-write-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:57:39 | recreate/10-write-data | ++ local 'uri=-h recreate-mysql-0.recreate-mysql -uroot -proot_password' logger.go:42: 14:57:39 | recreate/10-write-data | ++ local pod= logger.go:42: 14:57:39 | recreate/10-write-data | +++ get_client_pod logger.go:42: 14:57:39 | recreate/10-write-data | +++ kubectl -n kuttl-test-wired-lacewing get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:57:40 | recreate/10-write-data | ++ client_pod=mysql-client logger.go:42: 14:57:40 | recreate/10-write-data | ++ wait_pod mysql-client logger.go:42: 14:57:40 | recreate/10-write-data | ++ local pod=mysql-client logger.go:42: 14:57:40 | recreate/10-write-data | ++ set +o xtrace logger.go:42: 14:57:40 | recreate/10-write-data | mysql-clienttrue logger.go:42: 14:57:40 | recreate/10-write-data | ++ kubectl -n kuttl-test-wired-lacewing exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h recreate-mysql-0.recreate-mysql -uroot -proot_password' logger.go:42: 14:57:40 | recreate/10-write-data | ++ sed -e 's/mysql: //' logger.go:42: 14:57:40 | recreate/10-write-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:57:41 | recreate/10-write-data | + data='100500 logger.go:42: 14:57:41 | recreate/10-write-data | 100501 logger.go:42: 14:57:41 | recreate/10-write-data | 100502' logger.go:42: 14:57:41 | recreate/10-write-data | + kubectl create configmap -n kuttl-test-wired-lacewing 11-write-data-0 '--from-literal=data=100500 logger.go:42: 14:57:41 | recreate/10-write-data | 100501 logger.go:42: 14:57:41 | recreate/10-write-data | 100502' logger.go:42: 14:57:42 | recreate/10-write-data | configmap/11-write-data-0 created logger.go:42: 14:57:42 | recreate/10-write-data | + for i in 0 1 2 logger.go:42: 14:57:42 | recreate/10-write-data | +++ get_cluster_name logger.go:42: 14:57:42 | recreate/10-write-data | +++ kubectl -n kuttl-test-wired-lacewing get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:57:42 | recreate/10-write-data | ++ get_mysql_headless_fqdn recreate 1 logger.go:42: 14:57:42 | recreate/10-write-data | ++ local cluster=recreate logger.go:42: 14:57:42 | recreate/10-write-data | ++ local index=1 logger.go:42: 14:57:42 | recreate/10-write-data | ++ echo recreate-mysql-1.recreate-mysql logger.go:42: 14:57:42 | recreate/10-write-data | + host=recreate-mysql-1.recreate-mysql logger.go:42: 14:57:42 | recreate/10-write-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h recreate-mysql-1.recreate-mysql -uroot -proot_password' logger.go:42: 14:57:42 | recreate/10-write-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:57:42 | recreate/10-write-data | ++ local 'uri=-h recreate-mysql-1.recreate-mysql -uroot -proot_password' logger.go:42: 14:57:42 | recreate/10-write-data | ++ local pod= logger.go:42: 14:57:42 | recreate/10-write-data | +++ get_client_pod logger.go:42: 14:57:42 | recreate/10-write-data | +++ kubectl -n kuttl-test-wired-lacewing get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:57:42 | recreate/10-write-data | ++ client_pod=mysql-client logger.go:42: 14:57:42 | recreate/10-write-data | ++ wait_pod mysql-client logger.go:42: 14:57:42 | recreate/10-write-data | ++ local pod=mysql-client logger.go:42: 14:57:42 | recreate/10-write-data | ++ set +o xtrace logger.go:42: 14:57:43 | recreate/10-write-data | mysql-clienttrue logger.go:42: 14:57:43 | recreate/10-write-data | ++ kubectl -n kuttl-test-wired-lacewing exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h recreate-mysql-1.recreate-mysql -uroot -proot_password' logger.go:42: 14:57:43 | recreate/10-write-data | ++ sed -e 's/mysql: //' logger.go:42: 14:57:43 | recreate/10-write-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:57:44 | recreate/10-write-data | + data='100500 logger.go:42: 14:57:44 | recreate/10-write-data | 100501 logger.go:42: 14:57:44 | recreate/10-write-data | 100502' logger.go:42: 14:57:44 | recreate/10-write-data | + kubectl create configmap -n kuttl-test-wired-lacewing 11-write-data-1 '--from-literal=data=100500 logger.go:42: 14:57:44 | recreate/10-write-data | 100501 logger.go:42: 14:57:44 | recreate/10-write-data | 100502' logger.go:42: 14:57:44 | recreate/10-write-data | configmap/11-write-data-1 created logger.go:42: 14:57:44 | recreate/10-write-data | + for i in 0 1 2 logger.go:42: 14:57:44 | recreate/10-write-data | +++ get_cluster_name logger.go:42: 14:57:44 | recreate/10-write-data | +++ kubectl -n kuttl-test-wired-lacewing get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 14:57:45 | recreate/10-write-data | ++ get_mysql_headless_fqdn recreate 2 logger.go:42: 14:57:45 | recreate/10-write-data | ++ local cluster=recreate logger.go:42: 14:57:45 | recreate/10-write-data | ++ local index=2 logger.go:42: 14:57:45 | recreate/10-write-data | ++ echo recreate-mysql-2.recreate-mysql logger.go:42: 14:57:45 | recreate/10-write-data | + host=recreate-mysql-2.recreate-mysql logger.go:42: 14:57:45 | recreate/10-write-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h recreate-mysql-2.recreate-mysql -uroot -proot_password' logger.go:42: 14:57:45 | recreate/10-write-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 14:57:45 | recreate/10-write-data | ++ local 'uri=-h recreate-mysql-2.recreate-mysql -uroot -proot_password' logger.go:42: 14:57:45 | recreate/10-write-data | ++ local pod= logger.go:42: 14:57:45 | recreate/10-write-data | +++ get_client_pod logger.go:42: 14:57:45 | recreate/10-write-data | +++ kubectl -n kuttl-test-wired-lacewing get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 14:57:45 | recreate/10-write-data | ++ client_pod=mysql-client logger.go:42: 14:57:45 | recreate/10-write-data | ++ wait_pod mysql-client logger.go:42: 14:57:45 | recreate/10-write-data | ++ local pod=mysql-client logger.go:42: 14:57:45 | recreate/10-write-data | ++ set +o xtrace logger.go:42: 14:57:46 | recreate/10-write-data | mysql-clienttrue logger.go:42: 14:57:46 | recreate/10-write-data | ++ sed -e 's/mysql: //' logger.go:42: 14:57:46 | recreate/10-write-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 14:57:46 | recreate/10-write-data | ++ kubectl -n kuttl-test-wired-lacewing exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h recreate-mysql-2.recreate-mysql -uroot -proot_password' logger.go:42: 14:57:47 | recreate/10-write-data | + data='100500 logger.go:42: 14:57:47 | recreate/10-write-data | 100501 logger.go:42: 14:57:47 | recreate/10-write-data | 100502' logger.go:42: 14:57:47 | recreate/10-write-data | + kubectl create configmap -n kuttl-test-wired-lacewing 11-write-data-2 '--from-literal=data=100500 logger.go:42: 14:57:47 | recreate/10-write-data | 100501 logger.go:42: 14:57:47 | recreate/10-write-data | 100502' logger.go:42: 14:57:47 | recreate/10-write-data | configmap/11-write-data-2 created logger.go:42: 14:57:49 | recreate/10-write-data | test step completed 10-write-data logger.go:42: 14:57:49 | recreate/99-drop-finalizer | starting test step 99-drop-finalizer logger.go:42: 14:57:50 | recreate/99-drop-finalizer | PerconaServerMySQL:kuttl-test-wired-lacewing/recreate updated logger.go:42: 14:57:50 | recreate/99-drop-finalizer | test step completed 99-drop-finalizer logger.go:42: 14:57:50 | recreate | recreate events from ns kuttl-test-wired-lacewing: logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:16 +0000 UTC Normal Deployment.apps percona-server-mysql-operator ScalingReplicaSet Scaled up replica set percona-server-mysql-operator-b4c599bbb to 1 deployment-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:17 +0000 UTC Normal Pod percona-server-mysql-operator-b4c599bbb-5zgb2 Scheduled Successfully assigned kuttl-test-wired-lacewing/percona-server-mysql-operator-b4c599bbb-5zgb2 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-s4z3 default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:17 +0000 UTC Normal ReplicaSet.apps percona-server-mysql-operator-b4c599bbb SuccessfulCreate Created pod: percona-server-mysql-operator-b4c599bbb-5zgb2 replicaset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:19 +0000 UTC Normal Lease.coordination.k8s.io 08db2feb.percona.com LeaderElection percona-server-mysql-operator-b4c599bbb-5zgb2_cf032e42-6414-4641-a9cf-18e0f0ef5dd9 became leader percona-server-mysql-operator-b4c599bbb-5zgb2_cf032e42-6414-4641-a9cf-18e0f0ef5dd9 logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:19 +0000 UTC Normal Pod mysql-client Scheduled Successfully assigned kuttl-test-wired-lacewing/mysql-client to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-c4j1 default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:19 +0000 UTC Normal Pod percona-server-mysql-operator-b4c599bbb-5zgb2.spec.containers{manager} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:19 +0000 UTC Normal Pod percona-server-mysql-operator-b4c599bbb-5zgb2.spec.containers{manager} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 114.954004ms (114.96882ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:19 +0000 UTC Normal Pod percona-server-mysql-operator-b4c599bbb-5zgb2.spec.containers{manager} Created Created container manager kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:19 +0000 UTC Normal Pod percona-server-mysql-operator-b4c599bbb-5zgb2.spec.containers{manager} Started Started container manager kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:20 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Pulled Container image "percona/percona-server:8.0.33" already present on machine kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:20 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Created Created container mysql-client kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:20 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Started Started container mysql-client kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:30 +0000 UTC Normal Certificate.cert-manager.io recreate-ca-cert Issuing Issuing certificate as Secret does not exist logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:31 +0000 UTC Normal CertificateRequest.cert-manager.io recreate-ca-cert-l78pq WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:31 +0000 UTC Normal CertificateRequest.cert-manager.io recreate-ca-cert-l78pq WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:31 +0000 UTC Normal CertificateRequest.cert-manager.io recreate-ca-cert-l78pq WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:31 +0000 UTC Normal CertificateRequest.cert-manager.io recreate-ca-cert-l78pq WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:31 +0000 UTC Normal CertificateRequest.cert-manager.io recreate-ca-cert-l78pq WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:31 +0000 UTC Normal CertificateRequest.cert-manager.io recreate-ca-cert-l78pq cert-manager.io Certificate request has been approved by cert-manager.io logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:31 +0000 UTC Normal CertificateRequest.cert-manager.io recreate-ca-cert-l78pq CertificateIssued Certificate fetched from issuer successfully logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:31 +0000 UTC Normal Certificate.cert-manager.io recreate-ca-cert Generated Stored new private key in temporary Secret resource "recreate-ca-cert-kwq4c" logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:31 +0000 UTC Normal Certificate.cert-manager.io recreate-ca-cert Requested Created new CertificateRequest resource "recreate-ca-cert-l78pq" logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:31 +0000 UTC Normal Certificate.cert-manager.io recreate-ca-cert Issuing The certificate has been successfully issued logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:33 +0000 UTC Normal Issuer.cert-manager.io recreate-pso-issuer KeyPairVerified Signing CA verified logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:33 +0000 UTC Normal CertificateRequest.cert-manager.io recreate-ssl-qp4cq WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:33 +0000 UTC Normal CertificateRequest.cert-manager.io recreate-ssl-qp4cq WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:33 +0000 UTC Normal CertificateRequest.cert-manager.io recreate-ssl-qp4cq WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:33 +0000 UTC Normal CertificateRequest.cert-manager.io recreate-ssl-qp4cq WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:33 +0000 UTC Normal CertificateRequest.cert-manager.io recreate-ssl-qp4cq WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:33 +0000 UTC Normal CertificateRequest.cert-manager.io recreate-ssl-qp4cq cert-manager.io Certificate request has been approved by cert-manager.io logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:33 +0000 UTC Normal CertificateRequest.cert-manager.io recreate-ssl-qp4cq CertificateIssued Certificate fetched from issuer successfully logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:33 +0000 UTC Normal Certificate.cert-manager.io recreate-ssl Issuing Existing issued Secret is not up to date for spec: [spec.dnsNames] logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:33 +0000 UTC Normal Certificate.cert-manager.io recreate-ssl Reused Reusing private key stored in existing Secret resource "test-ssl" logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:33 +0000 UTC Normal Certificate.cert-manager.io recreate-ssl Requested Created new CertificateRequest resource "recreate-ssl-qp4cq" logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:34 +0000 UTC Normal Certificate.cert-manager.io recreate-ssl Issuing The certificate has been successfully issued logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:37 +0000 UTC Normal PersistentVolumeClaim datadir-recreate-mysql-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:37 +0000 UTC Normal PersistentVolumeClaim datadir-recreate-mysql-0 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:37 +0000 UTC Normal PersistentVolumeClaim datadir-recreate-mysql-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-wired-lacewing/datadir-recreate-mysql-0" pd.csi.storage.gke.io_gke-46a60d3caf524388b9ed-2266-f72b-vm_574702e6-7624-4023-8c8f-9ab089ab1643 logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:37 +0000 UTC Normal StatefulSet.apps recreate-mysql SuccessfulCreate create Claim datadir-recreate-mysql-0 Pod recreate-mysql-0 in StatefulSet recreate-mysql success statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:37 +0000 UTC Normal StatefulSet.apps recreate-mysql SuccessfulCreate create Pod recreate-mysql-0 in StatefulSet recreate-mysql successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:37 +0000 UTC Normal Pod recreate-orc-0 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-orc-0 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-s4z3 default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:37 +0000 UTC Normal StatefulSet.apps recreate-orc SuccessfulCreate create Pod recreate-orc-0 in StatefulSet recreate-orc successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:38 +0000 UTC Warning Pod recreate-orc-0 FailedMount MountVolume.SetUp failed for volume "config" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:39 +0000 UTC Warning Pod recreate-orc-0 FailedMount MountVolume.SetUp failed for volume "tls" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:40 +0000 UTC Normal Pod recreate-orc-0.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:40 +0000 UTC Normal Pod recreate-orc-0.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 98.615296ms (98.621902ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:40 +0000 UTC Normal Pod recreate-orc-0.spec.initContainers{orc-init} Created Created container orc-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:40 +0000 UTC Normal Pod recreate-orc-0.spec.initContainers{orc-init} Started Started container orc-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:41 +0000 UTC Normal PersistentVolumeClaim datadir-recreate-mysql-0 ProvisioningSucceeded Successfully provisioned volume pvc-9a16fcb8-2ed1-41a7-a364-1a03464361d3 pd.csi.storage.gke.io_gke-46a60d3caf524388b9ed-2266-f72b-vm_574702e6-7624-4023-8c8f-9ab089ab1643 logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:41 +0000 UTC Normal Pod recreate-mysql-0 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-mysql-0 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-c4j1 default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:41 +0000 UTC Normal Pod recreate-orc-0.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:41 +0000 UTC Normal Pod recreate-orc-0.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 91.674525ms (91.68769ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:41 +0000 UTC Normal Pod recreate-orc-0.spec.containers{orc} Created Created container orc kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:41 +0000 UTC Normal Pod recreate-orc-0.spec.containers{orc} Started Started container orc kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:42 +0000 UTC Normal Pod recreate-orc-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:42 +0000 UTC Normal Pod recreate-orc-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 73.413609ms (73.425363ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:42 +0000 UTC Normal Pod recreate-orc-0.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:42 +0000 UTC Normal Pod recreate-orc-0.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:49 +0000 UTC Normal Pod recreate-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-9a16fcb8-2ed1-41a7-a364-1a03464361d3" attachdetach-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:51 +0000 UTC Normal Pod recreate-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:51 +0000 UTC Normal Pod recreate-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 95.306458ms (95.325505ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:51 +0000 UTC Normal Pod recreate-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:51 +0000 UTC Normal Pod recreate-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:52 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:52 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 80.738411ms (80.747205ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:52 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:52 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:52 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:52 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 79.506109ms (79.513737ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:52 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:53 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:53 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:53 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 87.695049ms (87.711488ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:53 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{pt-heartbeat} Created Created container pt-heartbeat kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:45:53 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{pt-heartbeat} Started Started container pt-heartbeat kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:14 +0000 UTC Normal Pod recreate-orc-1 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-orc-1 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-jtft default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:14 +0000 UTC Normal StatefulSet.apps recreate-orc SuccessfulCreate create Pod recreate-orc-1 in StatefulSet recreate-orc successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:15 +0000 UTC Normal Pod recreate-orc-1.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:15 +0000 UTC Normal Pod recreate-orc-1.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 94.122208ms (94.137078ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:15 +0000 UTC Normal Pod recreate-orc-1.spec.initContainers{orc-init} Created Created container orc-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:15 +0000 UTC Normal Pod recreate-orc-1.spec.initContainers{orc-init} Started Started container orc-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:17 +0000 UTC Normal Pod recreate-orc-1.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:17 +0000 UTC Normal Pod recreate-orc-1.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 81.310359ms (81.327328ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:17 +0000 UTC Normal Pod recreate-orc-1.spec.containers{orc} Created Created container orc kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:17 +0000 UTC Normal Pod recreate-orc-1.spec.containers{orc} Started Started container orc kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:17 +0000 UTC Normal Pod recreate-orc-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:17 +0000 UTC Normal Pod recreate-orc-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 85.750849ms (85.764018ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:17 +0000 UTC Normal Pod recreate-orc-1.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:17 +0000 UTC Normal Pod recreate-orc-1.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:25 +0000 UTC Normal PersistentVolumeClaim datadir-recreate-mysql-1 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:25 +0000 UTC Normal PersistentVolumeClaim datadir-recreate-mysql-1 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:25 +0000 UTC Normal PersistentVolumeClaim datadir-recreate-mysql-1 Provisioning External provisioner is provisioning volume for claim "kuttl-test-wired-lacewing/datadir-recreate-mysql-1" pd.csi.storage.gke.io_gke-46a60d3caf524388b9ed-2266-f72b-vm_574702e6-7624-4023-8c8f-9ab089ab1643 logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:25 +0000 UTC Normal StatefulSet.apps recreate-mysql SuccessfulCreate create Claim datadir-recreate-mysql-1 Pod recreate-mysql-1 in StatefulSet recreate-mysql success statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:25 +0000 UTC Normal StatefulSet.apps recreate-mysql SuccessfulCreate create Pod recreate-mysql-1 in StatefulSet recreate-mysql successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:26 +0000 UTC Normal Pod recreate-haproxy-0 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-haproxy-0 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-s4z3 default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:26 +0000 UTC Normal StatefulSet.apps recreate-haproxy SuccessfulCreate create Pod recreate-haproxy-0 in StatefulSet recreate-haproxy successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:27 +0000 UTC Normal Pod recreate-haproxy-0.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:27 +0000 UTC Normal Pod recreate-haproxy-0.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 82.005843ms (82.017916ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:27 +0000 UTC Normal Pod recreate-haproxy-0.spec.initContainers{haproxy-init} Created Created container haproxy-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:27 +0000 UTC Normal Pod recreate-haproxy-0.spec.initContainers{haproxy-init} Started Started container haproxy-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:28 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:28 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 89.759756ms (89.774034ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:28 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{haproxy} Created Created container haproxy kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:28 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{haproxy} Started Started container haproxy kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:28 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:29 +0000 UTC Normal PersistentVolumeClaim datadir-recreate-mysql-1 ProvisioningSucceeded Successfully provisioned volume pvc-daeea10d-e100-4ef6-ad24-74ce329aa622 pd.csi.storage.gke.io_gke-46a60d3caf524388b9ed-2266-f72b-vm_574702e6-7624-4023-8c8f-9ab089ab1643 logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:29 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 105.143511ms (105.151426ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:29 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:29 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:30 +0000 UTC Normal Pod recreate-haproxy-1 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-haproxy-1 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-c4j1 default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:30 +0000 UTC Normal Pod recreate-haproxy-1.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:30 +0000 UTC Normal Pod recreate-haproxy-1.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 105.485306ms (105.500619ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:30 +0000 UTC Normal Pod recreate-haproxy-1.spec.initContainers{haproxy-init} Created Created container haproxy-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:30 +0000 UTC Normal StatefulSet.apps recreate-haproxy SuccessfulCreate create Pod recreate-haproxy-1 in StatefulSet recreate-haproxy successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:30 +0000 UTC Normal Pod recreate-mysql-1 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-mysql-1 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-jtft default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:31 +0000 UTC Normal Pod recreate-haproxy-1.spec.initContainers{haproxy-init} Started Started container haproxy-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:32 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:32 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 75.292435ms (75.311385ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:32 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{haproxy} Created Created container haproxy kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:32 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{haproxy} Started Started container haproxy kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:32 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:33 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 85.725679ms (85.743011ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:33 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:33 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:33 +0000 UTC Normal Pod recreate-haproxy-2 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-haproxy-2 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-jtft default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:33 +0000 UTC Normal StatefulSet.apps recreate-haproxy SuccessfulCreate create Pod recreate-haproxy-2 in StatefulSet recreate-haproxy successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:34 +0000 UTC Normal Pod recreate-haproxy-2.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:34 +0000 UTC Normal Pod recreate-haproxy-2.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 104.763345ms (104.781086ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:34 +0000 UTC Normal Pod recreate-haproxy-2.spec.initContainers{haproxy-init} Created Created container haproxy-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:34 +0000 UTC Normal Pod recreate-haproxy-2.spec.initContainers{haproxy-init} Started Started container haproxy-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:36 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:36 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 100.879171ms (100.895991ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:36 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{haproxy} Created Created container haproxy kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:36 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{haproxy} Started Started container haproxy kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:36 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:36 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 75.332777ms (75.345396ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:36 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:36 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:38 +0000 UTC Normal Pod recreate-mysql-1 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-daeea10d-e100-4ef6-ad24-74ce329aa622" attachdetach-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:39 +0000 UTC Normal Pod recreate-mysql-1.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:39 +0000 UTC Normal Pod recreate-mysql-1.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 83.014387ms (83.037369ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:39 +0000 UTC Normal Pod recreate-mysql-1.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:39 +0000 UTC Normal Pod recreate-mysql-1.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:41 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:41 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 112.348464ms (112.356902ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:41 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:41 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:41 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:41 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 85.462924ms (85.47085ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:41 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:41 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:41 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:41 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 171.610126ms (171.616479ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:41 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{pt-heartbeat} Created Created container pt-heartbeat kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:41 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{pt-heartbeat} Started Started container pt-heartbeat kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:50 +0000 UTC Normal Pod recreate-orc-2 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-orc-2 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-c4j1 default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:50 +0000 UTC Normal StatefulSet.apps recreate-orc SuccessfulCreate create Pod recreate-orc-2 in StatefulSet recreate-orc successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:51 +0000 UTC Normal Pod recreate-orc-2.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:51 +0000 UTC Normal Pod recreate-orc-2.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 83.774082ms (83.788185ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:51 +0000 UTC Normal Pod recreate-orc-2.spec.initContainers{orc-init} Created Created container orc-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:51 +0000 UTC Normal Pod recreate-orc-2.spec.initContainers{orc-init} Started Started container orc-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:52 +0000 UTC Normal Pod recreate-orc-2.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:52 +0000 UTC Normal Pod recreate-orc-2.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 89.18454ms (89.199828ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:52 +0000 UTC Normal Pod recreate-orc-2.spec.containers{orc} Created Created container orc kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:53 +0000 UTC Normal Pod recreate-orc-2.spec.containers{orc} Started Started container orc kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:53 +0000 UTC Normal Pod recreate-orc-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:53 +0000 UTC Normal Pod recreate-orc-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 98.7495ms (98.762337ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:53 +0000 UTC Normal Pod recreate-orc-2.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:46:53 +0000 UTC Normal Pod recreate-orc-2.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:00 +0000 UTC Warning Pod recreate-mysql-1.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:00 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:03 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 92.026593ms (92.051555ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:34 +0000 UTC Normal PersistentVolumeClaim datadir-recreate-mysql-2 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:34 +0000 UTC Normal PersistentVolumeClaim datadir-recreate-mysql-2 Provisioning External provisioner is provisioning volume for claim "kuttl-test-wired-lacewing/datadir-recreate-mysql-2" pd.csi.storage.gke.io_gke-46a60d3caf524388b9ed-2266-f72b-vm_574702e6-7624-4023-8c8f-9ab089ab1643 logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:34 +0000 UTC Normal PersistentVolumeClaim datadir-recreate-mysql-2 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:34 +0000 UTC Normal StatefulSet.apps recreate-mysql SuccessfulCreate create Claim datadir-recreate-mysql-2 Pod recreate-mysql-2 in StatefulSet recreate-mysql success statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:34 +0000 UTC Normal StatefulSet.apps recreate-mysql SuccessfulCreate create Pod recreate-mysql-2 in StatefulSet recreate-mysql successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:38 +0000 UTC Normal PersistentVolumeClaim datadir-recreate-mysql-2 ProvisioningSucceeded Successfully provisioned volume pvc-e41a5f87-b905-4d4b-8353-33babdd062fd pd.csi.storage.gke.io_gke-46a60d3caf524388b9ed-2266-f72b-vm_574702e6-7624-4023-8c8f-9ab089ab1643 logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:38 +0000 UTC Normal Pod recreate-mysql-2 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-mysql-2 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-s4z3 default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:46 +0000 UTC Normal Pod recreate-mysql-2 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-e41a5f87-b905-4d4b-8353-33babdd062fd" attachdetach-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:47 +0000 UTC Normal Pod recreate-mysql-2.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:47 +0000 UTC Normal Pod recreate-mysql-2.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 119.419169ms (119.432948ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:47 +0000 UTC Normal Pod recreate-mysql-2.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:47 +0000 UTC Normal Pod recreate-mysql-2.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:48 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:49 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 76.806304ms (76.824657ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:49 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:49 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:49 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:49 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 79.097068ms (79.104205ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:49 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:49 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:49 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:49 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 70.345032ms (70.361667ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:49 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{pt-heartbeat} Created Created container pt-heartbeat kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:47:49 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{pt-heartbeat} Started Started container pt-heartbeat kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:48:07 +0000 UTC Warning Pod recreate-mysql-2.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:48:07 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:48:10 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 86.299256ms (86.307627ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:09 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{xtrabackup} Killing Stopping container xtrabackup kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:09 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:09 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{pt-heartbeat} Killing Stopping container pt-heartbeat kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:09 +0000 UTC Normal StatefulSet.apps recreate-mysql SuccessfulDelete delete Pod recreate-mysql-2 in StatefulSet recreate-mysql successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:09 +0000 UTC Normal Pod recreate-orc-2.spec.containers{orc} Killing Stopping container orc kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:09 +0000 UTC Normal Pod recreate-orc-2.spec.containers{mysql-monit} Killing Stopping container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:09 +0000 UTC Normal StatefulSet.apps recreate-orc SuccessfulDelete delete Pod recreate-orc-2 in StatefulSet recreate-orc successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:10 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{haproxy} Killing Stopping container haproxy kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:10 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{mysql-monit} Killing Stopping container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:10 +0000 UTC Normal StatefulSet.apps recreate-haproxy SuccessfulDelete delete Pod recreate-haproxy-2 in StatefulSet recreate-haproxy successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:11 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{haproxy} Killing Stopping container haproxy kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:11 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{mysql-monit} Killing Stopping container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:11 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{haproxy} Killing Stopping container haproxy kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:11 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{mysql-monit} Killing Stopping container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:11 +0000 UTC Normal StatefulSet.apps recreate-haproxy SuccessfulDelete delete Pod recreate-haproxy-1 in StatefulSet recreate-haproxy successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:11 +0000 UTC Normal StatefulSet.apps recreate-haproxy SuccessfulDelete delete Pod recreate-haproxy-0 in StatefulSet recreate-haproxy successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:14 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{xtrabackup} Killing Stopping container xtrabackup kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:14 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:14 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{pt-heartbeat} Killing Stopping container pt-heartbeat kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:14 +0000 UTC Warning Endpoints recreate-mysql-unready FailedToUpdateEndpoint Failed to update endpoint kuttl-test-wired-lacewing/recreate-mysql-unready: Operation cannot be fulfilled on endpoints "recreate-mysql-unready": the object has been modified; please apply your changes to the latest version and try again endpoint-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:14 +0000 UTC Normal StatefulSet.apps recreate-mysql SuccessfulDelete delete Pod recreate-mysql-1 in StatefulSet recreate-mysql successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:19 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:19 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{pt-heartbeat} Killing Stopping container pt-heartbeat kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:19 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{xtrabackup} Killing Stopping container xtrabackup kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:19 +0000 UTC Normal StatefulSet.apps recreate-mysql SuccessfulDelete delete Pod recreate-mysql-0 in StatefulSet recreate-mysql successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:20 +0000 UTC Warning Pod recreate-mysql-0.spec.containers{mysql} Unhealthy Readiness probe failed: 2024/02/05 14:49:20 readiness check failed: connect to db: ping DB: dial tcp 10.248.88.30:33062: connect: connection refused kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:25 +0000 UTC Warning Pod recreate-mysql-0.spec.containers{mysql} Unhealthy Readiness probe errored: rpc error: code = Unknown desc = failed to exec in container: container is in CONTAINER_EXITED state kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:40 +0000 UTC Normal Pod recreate-orc-1.spec.containers{orc} Killing Stopping container orc kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:40 +0000 UTC Normal Pod recreate-orc-1.spec.containers{mysql-monit} Killing Stopping container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:40 +0000 UTC Warning Endpoints recreate-orc FailedToUpdateEndpoint Failed to update endpoint kuttl-test-wired-lacewing/recreate-orc: Operation cannot be fulfilled on endpoints "recreate-orc": the object has been modified; please apply your changes to the latest version and try again endpoint-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:49:40 +0000 UTC Normal StatefulSet.apps recreate-orc SuccessfulDelete delete Pod recreate-orc-1 in StatefulSet recreate-orc successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:50:12 +0000 UTC Normal Pod recreate-orc-0.spec.containers{orc} Killing Stopping container orc kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:50:12 +0000 UTC Normal Pod recreate-orc-0.spec.containers{mysql-monit} Killing Stopping container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:50:12 +0000 UTC Normal StatefulSet.apps recreate-orc SuccessfulDelete delete Pod recreate-orc-0 in StatefulSet recreate-orc successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:50:58 +0000 UTC Normal Pod recreate-mysql-0 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-mysql-0 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-c4j1 default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:50:58 +0000 UTC Normal Pod recreate-orc-0 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-orc-0 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-s4z3 default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:50:58 +0000 UTC Normal Pod recreate-orc-0.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:50:59 +0000 UTC Normal Pod recreate-orc-0.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 132.472425ms (132.489519ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:50:59 +0000 UTC Normal Pod recreate-orc-0.spec.initContainers{orc-init} Created Created container orc-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:50:59 +0000 UTC Normal Pod recreate-orc-0.spec.initContainers{orc-init} Started Started container orc-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:00 +0000 UTC Normal Pod recreate-orc-0.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:00 +0000 UTC Normal Pod recreate-orc-0.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 79.790967ms (79.802226ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:00 +0000 UTC Normal Pod recreate-orc-0.spec.containers{orc} Created Created container orc kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:00 +0000 UTC Normal Pod recreate-orc-0.spec.containers{orc} Started Started container orc kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:00 +0000 UTC Normal Pod recreate-orc-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:00 +0000 UTC Normal Pod recreate-orc-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 72.732696ms (72.739838ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:00 +0000 UTC Normal Pod recreate-orc-0.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:00 +0000 UTC Normal Pod recreate-orc-0.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:05 +0000 UTC Normal Pod recreate-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-9a16fcb8-2ed1-41a7-a364-1a03464361d3" attachdetach-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:07 +0000 UTC Normal Pod recreate-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:07 +0000 UTC Normal Pod recreate-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 104.387205ms (104.401369ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:07 +0000 UTC Normal Pod recreate-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:08 +0000 UTC Normal Pod recreate-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:09 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:09 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 88.473603ms (88.489692ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:09 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:09 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:09 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:09 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 83.799183ms (83.806738ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:09 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:09 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:09 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:09 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 80.980161ms (80.990645ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:09 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{pt-heartbeat} Created Created container pt-heartbeat kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:10 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{pt-heartbeat} Started Started container pt-heartbeat kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:33 +0000 UTC Normal Pod recreate-orc-1 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-orc-1 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-jtft default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:34 +0000 UTC Normal Pod recreate-orc-1.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:34 +0000 UTC Normal Pod recreate-orc-1.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 91.417101ms (91.425154ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:34 +0000 UTC Normal Pod recreate-orc-1.spec.initContainers{orc-init} Created Created container orc-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:34 +0000 UTC Normal Pod recreate-orc-1.spec.initContainers{orc-init} Started Started container orc-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:36 +0000 UTC Normal Pod recreate-orc-1.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:36 +0000 UTC Normal Pod recreate-orc-1.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 84.337823ms (84.355127ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:36 +0000 UTC Normal Pod recreate-orc-1.spec.containers{orc} Created Created container orc kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:36 +0000 UTC Normal Pod recreate-orc-1.spec.containers{orc} Started Started container orc kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:36 +0000 UTC Normal Pod recreate-orc-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:36 +0000 UTC Normal Pod recreate-orc-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 76.976665ms (76.994072ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:36 +0000 UTC Normal Pod recreate-orc-1.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:36 +0000 UTC Normal Pod recreate-orc-1.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:42 +0000 UTC Normal Pod recreate-mysql-1 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-mysql-1 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-jtft default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:43 +0000 UTC Warning Pod recreate-mysql-1 FailedMount MountVolume.SetUp failed for volume "config" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:46 +0000 UTC Normal Pod recreate-haproxy-0 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-haproxy-0 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-s4z3 default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:47 +0000 UTC Normal Pod recreate-haproxy-0.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:47 +0000 UTC Normal Pod recreate-haproxy-0.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 97.385274ms (97.399648ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:47 +0000 UTC Normal Pod recreate-haproxy-0.spec.initContainers{haproxy-init} Created Created container haproxy-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:47 +0000 UTC Normal Pod recreate-haproxy-0.spec.initContainers{haproxy-init} Started Started container haproxy-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:48 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:48 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 79.665312ms (79.671645ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:48 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{haproxy} Created Created container haproxy kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:48 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{haproxy} Started Started container haproxy kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:48 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:48 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 83.131197ms (83.144269ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:48 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:48 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:49 +0000 UTC Normal Pod recreate-haproxy-1 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-haproxy-1 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-c4j1 default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:50 +0000 UTC Normal Pod recreate-haproxy-1.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:50 +0000 UTC Normal Pod recreate-haproxy-1.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 83.703042ms (83.712012ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:50 +0000 UTC Normal Pod recreate-haproxy-1.spec.initContainers{haproxy-init} Created Created container haproxy-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:50 +0000 UTC Normal Pod recreate-haproxy-1.spec.initContainers{haproxy-init} Started Started container haproxy-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:50 +0000 UTC Normal Pod recreate-mysql-1 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-daeea10d-e100-4ef6-ad24-74ce329aa622" attachdetach-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:51 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:51 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 76.27718ms (76.292655ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:51 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{haproxy} Created Created container haproxy kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:51 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{haproxy} Started Started container haproxy kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:51 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:51 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 99.033888ms (99.052278ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:51 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:51 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:51 +0000 UTC Normal Pod recreate-mysql-1.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:51 +0000 UTC Normal Pod recreate-mysql-1.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 92.319544ms (92.357408ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:51 +0000 UTC Normal Pod recreate-mysql-1.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:51 +0000 UTC Normal Pod recreate-mysql-1.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:52 +0000 UTC Normal Pod recreate-haproxy-2 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-haproxy-2 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-jtft default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:53 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:53 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 76.690435ms (76.702892ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:53 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:53 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:53 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:53 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 80.163975ms (80.170468ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:53 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:53 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:53 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:53 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 94.980436ms (94.987274ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:53 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{pt-heartbeat} Created Created container pt-heartbeat kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:53 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{pt-heartbeat} Started Started container pt-heartbeat kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:54 +0000 UTC Warning Pod recreate-haproxy-2 FailedMount MountVolume.SetUp failed for volume "config" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:55 +0000 UTC Normal Pod recreate-haproxy-2.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:55 +0000 UTC Normal Pod recreate-haproxy-2.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 76.512414ms (76.528997ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:55 +0000 UTC Normal Pod recreate-haproxy-2.spec.initContainers{haproxy-init} Created Created container haproxy-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:55 +0000 UTC Normal Pod recreate-haproxy-2.spec.initContainers{haproxy-init} Started Started container haproxy-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:57 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:57 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 79.482898ms (79.499117ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:57 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{haproxy} Created Created container haproxy kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:57 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{haproxy} Started Started container haproxy kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:57 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:57 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 85.029804ms (85.043121ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:57 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:51:57 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:09 +0000 UTC Normal Pod recreate-orc-2 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-orc-2 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-c4j1 default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:09 +0000 UTC Normal Pod recreate-orc-2.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:09 +0000 UTC Normal Pod recreate-orc-2.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 89.114693ms (89.131234ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:09 +0000 UTC Normal Pod recreate-orc-2.spec.initContainers{orc-init} Created Created container orc-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:10 +0000 UTC Normal Pod recreate-orc-2.spec.initContainers{orc-init} Started Started container orc-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:11 +0000 UTC Normal Pod recreate-orc-2.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:11 +0000 UTC Normal Pod recreate-orc-2.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 99.662464ms (99.704605ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:11 +0000 UTC Normal Pod recreate-orc-2.spec.containers{orc} Created Created container orc kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:11 +0000 UTC Normal Pod recreate-orc-2.spec.containers{orc} Started Started container orc kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:11 +0000 UTC Normal Pod recreate-orc-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:11 +0000 UTC Normal Pod recreate-orc-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 90.90056ms (90.915482ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:12 +0000 UTC Warning Pod recreate-mysql-1.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:12 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:12 +0000 UTC Normal Pod recreate-orc-2.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:12 +0000 UTC Normal Pod recreate-orc-2.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:16 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 75.704209ms (75.72107ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:46 +0000 UTC Normal Pod recreate-mysql-2 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-mysql-2 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-s4z3 default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:54 +0000 UTC Normal Pod recreate-mysql-2 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-e41a5f87-b905-4d4b-8353-33babdd062fd" attachdetach-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:55 +0000 UTC Normal Pod recreate-mysql-2.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:55 +0000 UTC Normal Pod recreate-mysql-2.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 93.27355ms (93.279574ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:55 +0000 UTC Normal Pod recreate-mysql-2.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:55 +0000 UTC Normal Pod recreate-mysql-2.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:57 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:57 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 105.731962ms (105.748287ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:57 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:57 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:57 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:58 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 73.905446ms (73.920568ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:58 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:58 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:58 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:58 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 83.728244ms (83.735222ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:58 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{pt-heartbeat} Created Created container pt-heartbeat kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:52:58 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{pt-heartbeat} Started Started container pt-heartbeat kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:53:16 +0000 UTC Warning Pod recreate-mysql-2.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:53:16 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:53:19 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 75.397153ms (75.405212ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:10 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{xtrabackup} Killing Stopping container xtrabackup kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:10 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:10 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{pt-heartbeat} Killing Stopping container pt-heartbeat kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:15 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{xtrabackup} Killing Stopping container xtrabackup kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:15 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:23 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{haproxy} Killing Stopping container haproxy kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:23 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{mysql-monit} Killing Stopping container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:23 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:23 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{pt-heartbeat} Killing Stopping container pt-heartbeat kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:23 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{xtrabackup} Killing Stopping container xtrabackup kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:24 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{haproxy} Killing Stopping container haproxy kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:24 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{mysql-monit} Killing Stopping container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:24 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{haproxy} Killing Stopping container haproxy kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:24 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{mysql-monit} Killing Stopping container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:24 +0000 UTC Normal Pod recreate-orc-0.spec.containers{orc} Killing Stopping container orc kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:24 +0000 UTC Normal Pod recreate-orc-0.spec.containers{mysql-monit} Killing Stopping container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:24 +0000 UTC Normal Pod recreate-orc-1.spec.containers{orc} Killing Stopping container orc kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:24 +0000 UTC Normal Pod recreate-orc-1.spec.containers{mysql-monit} Killing Stopping container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:24 +0000 UTC Normal Pod recreate-orc-2.spec.containers{orc} Killing Stopping container orc kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:24 +0000 UTC Normal Pod recreate-orc-2.spec.containers{mysql-monit} Killing Stopping container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:27 +0000 UTC Warning Pod recreate-mysql-0.spec.containers{mysql} Unhealthy Readiness probe failed: 2024/02/05 14:54:27 readiness check failed: connect to db: ping DB: dial tcp 10.248.88.33:33062: connect: connection refused kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:32 +0000 UTC Warning Pod recreate-mysql-0.spec.containers{mysql} Unhealthy Readiness probe errored: rpc error: code = Unknown desc = failed to exec in container: container is in CONTAINER_EXITED state kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:35 +0000 UTC Normal Pod recreate-haproxy-0 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-haproxy-0 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-s4z3 default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:35 +0000 UTC Normal StatefulSet.apps recreate-haproxy SuccessfulCreate create Pod recreate-haproxy-0 in StatefulSet recreate-haproxy successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:36 +0000 UTC Normal Pod recreate-haproxy-0.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:36 +0000 UTC Normal Pod recreate-haproxy-0.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 96.681461ms (96.701679ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:36 +0000 UTC Normal Pod recreate-haproxy-0.spec.initContainers{haproxy-init} Created Created container haproxy-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:36 +0000 UTC Normal Pod recreate-haproxy-0.spec.initContainers{haproxy-init} Started Started container haproxy-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:38 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:38 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 101.570545ms (101.584691ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:38 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{haproxy} Created Created container haproxy kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:38 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{haproxy} Started Started container haproxy kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:38 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:38 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 143.867446ms (143.882436ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:38 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:38 +0000 UTC Normal Pod recreate-haproxy-0.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:39 +0000 UTC Normal Pod recreate-haproxy-1 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-haproxy-1 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-jtft default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:39 +0000 UTC Normal StatefulSet.apps recreate-haproxy SuccessfulCreate create Pod recreate-haproxy-1 in StatefulSet recreate-haproxy successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:40 +0000 UTC Normal Pod recreate-haproxy-1.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:40 +0000 UTC Normal Pod recreate-haproxy-1.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 94.213983ms (94.228524ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:40 +0000 UTC Normal Pod recreate-haproxy-1.spec.initContainers{haproxy-init} Created Created container haproxy-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:40 +0000 UTC Normal Pod recreate-haproxy-1.spec.initContainers{haproxy-init} Started Started container haproxy-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:41 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:41 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 81.618799ms (81.633432ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:41 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{haproxy} Created Created container haproxy kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:41 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{haproxy} Started Started container haproxy kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:41 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:41 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 84.954905ms (84.969022ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:41 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:41 +0000 UTC Normal Pod recreate-haproxy-1.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:42 +0000 UTC Normal Pod recreate-haproxy-2 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-haproxy-2 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-c4j1 default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:42 +0000 UTC Normal StatefulSet.apps recreate-haproxy SuccessfulCreate create Pod recreate-haproxy-2 in StatefulSet recreate-haproxy successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:43 +0000 UTC Normal Pod recreate-haproxy-2.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:43 +0000 UTC Normal Pod recreate-haproxy-2.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 341.83643ms (341.851239ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:43 +0000 UTC Normal Pod recreate-haproxy-2.spec.initContainers{haproxy-init} Created Created container haproxy-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:43 +0000 UTC Normal Pod recreate-haproxy-2.spec.initContainers{haproxy-init} Started Started container haproxy-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:45 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:45 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 106.836022ms (106.852803ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:45 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{haproxy} Created Created container haproxy kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:45 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{haproxy} Started Started container haproxy kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:45 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:45 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 86.603706ms (86.62577ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:45 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:45 +0000 UTC Normal Pod recreate-haproxy-2.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:46 +0000 UTC Normal Pod recreate-mysql-0 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-mysql-0 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-c4j1 default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:46 +0000 UTC Normal StatefulSet.apps recreate-mysql SuccessfulCreate create Pod recreate-mysql-0 in StatefulSet recreate-mysql successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:51 +0000 UTC Normal Pod recreate-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:51 +0000 UTC Normal Pod recreate-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 77.316545ms (77.340806ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:51 +0000 UTC Normal Pod recreate-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:51 +0000 UTC Normal Pod recreate-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:53 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:53 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 100.53231ms (100.546835ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:53 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:53 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:53 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:53 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 84.425527ms (84.434391ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:53 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:53 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:53 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:53 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 112.275378ms (112.283956ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:53 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{pt-heartbeat} Created Created container pt-heartbeat kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:53 +0000 UTC Normal Pod recreate-mysql-0.spec.containers{pt-heartbeat} Started Started container pt-heartbeat kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:56 +0000 UTC Normal StatefulSet.apps recreate-orc SuccessfulCreate create Pod recreate-orc-0 in StatefulSet recreate-orc successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:57 +0000 UTC Normal Pod recreate-orc-0 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-orc-0 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-s4z3 default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:57 +0000 UTC Normal Pod recreate-orc-0.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:57 +0000 UTC Normal Pod recreate-orc-0.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 71.876396ms (71.892781ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:57 +0000 UTC Normal Pod recreate-orc-0.spec.initContainers{orc-init} Created Created container orc-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:57 +0000 UTC Normal Pod recreate-orc-0.spec.initContainers{orc-init} Started Started container orc-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:57 +0000 UTC Warning Pod recreate-orc-2.spec.containers{orc} Unhealthy Readiness probe failed: Get "http://10.248.88.35:3000/api/health": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:59 +0000 UTC Normal Pod recreate-orc-0.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:59 +0000 UTC Normal Pod recreate-orc-0.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 82.778103ms (82.784863ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:59 +0000 UTC Normal Pod recreate-orc-0.spec.containers{orc} Created Created container orc kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:59 +0000 UTC Normal Pod recreate-orc-0.spec.containers{orc} Started Started container orc kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:59 +0000 UTC Normal Pod recreate-orc-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:59 +0000 UTC Normal Pod recreate-orc-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 77.720095ms (77.735266ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:59 +0000 UTC Normal Pod recreate-orc-0.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:54:59 +0000 UTC Normal Pod recreate-orc-0.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:26 +0000 UTC Normal Pod recreate-mysql-1 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-mysql-1 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-jtft default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:26 +0000 UTC Normal StatefulSet.apps recreate-mysql SuccessfulCreate create Pod recreate-mysql-1 in StatefulSet recreate-mysql successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:32 +0000 UTC Normal Pod recreate-orc-1 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-orc-1 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-c4j1 default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:32 +0000 UTC Normal StatefulSet.apps recreate-orc SuccessfulCreate create Pod recreate-orc-1 in StatefulSet recreate-orc successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:33 +0000 UTC Normal Pod recreate-orc-1.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:33 +0000 UTC Normal Pod recreate-orc-1.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 98.518027ms (98.537259ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:33 +0000 UTC Normal Pod recreate-orc-1.spec.initContainers{orc-init} Created Created container orc-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:33 +0000 UTC Normal Pod recreate-orc-1.spec.initContainers{orc-init} Started Started container orc-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:34 +0000 UTC Normal Pod recreate-mysql-1 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-daeea10d-e100-4ef6-ad24-74ce329aa622" attachdetach-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:35 +0000 UTC Normal Pod recreate-mysql-1.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:35 +0000 UTC Normal Pod recreate-mysql-1.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 92.082872ms (92.09924ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:35 +0000 UTC Normal Pod recreate-mysql-1.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:35 +0000 UTC Normal Pod recreate-mysql-1.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:35 +0000 UTC Normal Pod recreate-orc-1.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:35 +0000 UTC Normal Pod recreate-orc-1.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 157.285083ms (157.300973ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:35 +0000 UTC Normal Pod recreate-orc-1.spec.containers{orc} Created Created container orc kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:35 +0000 UTC Normal Pod recreate-orc-1.spec.containers{orc} Started Started container orc kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:35 +0000 UTC Normal Pod recreate-orc-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:35 +0000 UTC Normal Pod recreate-orc-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 104.810966ms (104.819476ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:35 +0000 UTC Normal Pod recreate-orc-1.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:35 +0000 UTC Normal Pod recreate-orc-1.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:36 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:36 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 84.110963ms (84.138875ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:36 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:36 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:36 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:36 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 93.493049ms (93.506458ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:36 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:37 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:37 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:37 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 102.202451ms (102.209815ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:37 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{pt-heartbeat} Created Created container pt-heartbeat kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:37 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{pt-heartbeat} Started Started container pt-heartbeat kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:56 +0000 UTC Warning Pod recreate-mysql-1.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:56 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:55:59 +0000 UTC Normal Pod recreate-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 93.584251ms (93.597052ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:07 +0000 UTC Normal Pod recreate-orc-2 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-orc-2 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-jtft default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:07 +0000 UTC Normal StatefulSet.apps recreate-orc SuccessfulCreate create Pod recreate-orc-2 in StatefulSet recreate-orc successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:08 +0000 UTC Normal Pod recreate-orc-2.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:08 +0000 UTC Normal Pod recreate-orc-2.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 74.474582ms (74.490685ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:08 +0000 UTC Normal Pod recreate-orc-2.spec.initContainers{orc-init} Created Created container orc-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:08 +0000 UTC Normal Pod recreate-orc-2.spec.initContainers{orc-init} Started Started container orc-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:09 +0000 UTC Normal Pod recreate-orc-2.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:09 +0000 UTC Normal Pod recreate-orc-2.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 70.515679ms (70.532734ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:09 +0000 UTC Normal Pod recreate-orc-2.spec.containers{orc} Created Created container orc kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:09 +0000 UTC Normal Pod recreate-orc-2.spec.containers{orc} Started Started container orc kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:09 +0000 UTC Normal Pod recreate-orc-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:10 +0000 UTC Normal Pod recreate-orc-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 82.579879ms (82.59869ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:10 +0000 UTC Normal Pod recreate-orc-2.spec.containers{mysql-monit} Created Created container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:10 +0000 UTC Normal Pod recreate-orc-2.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:30 +0000 UTC Normal Pod recreate-mysql-2 Scheduled Successfully assigned kuttl-test-wired-lacewing/recreate-mysql-2 to gke-jen-ps-523-f00253e-4-default-pool-ea5b7f5d-s4z3 default-scheduler logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:30 +0000 UTC Normal StatefulSet.apps recreate-mysql SuccessfulCreate create Pod recreate-mysql-2 in StatefulSet recreate-mysql successful statefulset-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:37 +0000 UTC Normal Pod recreate-mysql-2 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-e41a5f87-b905-4d4b-8353-33babdd062fd" attachdetach-controller logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:39 +0000 UTC Normal Pod recreate-mysql-2.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:39 +0000 UTC Normal Pod recreate-mysql-2.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 110.392766ms (110.416663ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:39 +0000 UTC Normal Pod recreate-mysql-2.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:39 +0000 UTC Normal Pod recreate-mysql-2.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:40 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:40 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 270.806863ms (270.814017ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:40 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:40 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:40 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:40 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 87.563881ms (87.572117ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:40 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:40 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:40 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:40 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 95.042223ms (95.047582ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:41 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{pt-heartbeat} Created Created container pt-heartbeat kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:41 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{pt-heartbeat} Started Started container pt-heartbeat kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:59 +0000 UTC Warning Pod recreate-mysql-2.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:56:59 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 14:57:50 | recreate | 2024-02-05 14:57:03 +0000 UTC Normal Pod recreate-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 80.165759ms (80.17896ms including waiting) kubelet logger.go:42: 14:57:50 | recreate | Deleting namespace: kuttl-test-wired-lacewing === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (804.59s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/recreate (801.61s) PASS