=== RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://35.202.81.226 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 180 seconds for each step harness.go:372: testsuite: e2e-tests/tests has 26 tests === RUN kuttl/harness === RUN kuttl/harness/service-per-pod === PAUSE kuttl/harness/service-per-pod === CONT kuttl/harness/service-per-pod logger.go:42: 12:58:29 | service-per-pod | Creating namespace: kuttl-test-moving-shepherd logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | starting test step 0-deploy-operator logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | running command: [sh -c set -o errexit set -o xtrace source ../../functions deploy_operator deploy_non_tls_cluster_secrets deploy_tls_cluster_secrets deploy_client] logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | + source ../../functions logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ realpath ../../.. logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | ++++ mktemp -d logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ export TEMP_DIR=/tmp/tmp.y7gr47zMhU logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ TEMP_DIR=/tmp/tmp.y7gr47zMhU logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ GIT_BRANCH=PR-339 logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ VERSION=PR-339-3a8483b logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | ++++ which gdate logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | ++++ which date logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ date=/usr/bin/date logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ command -v oc logger.go:42: 12:58:29 | service-per-pod/0-deploy-operator | +++ oc get projects logger.go:42: 12:58:35 | service-per-pod/0-deploy-operator | error: the server doesn't have a resource type "projects" logger.go:42: 12:58:35 | service-per-pod/0-deploy-operator | +++ kubectl get nodes logger.go:42: 12:58:35 | service-per-pod/0-deploy-operator | +++ grep '^minikube' logger.go:42: 12:58:35 | service-per-pod/0-deploy-operator | ++++ pwd logger.go:42: 12:58:35 | service-per-pod/0-deploy-operator | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/service-per-pod logger.go:42: 12:58:35 | service-per-pod/0-deploy-operator | ++ test_name=service-per-pod logger.go:42: 12:58:35 | service-per-pod/0-deploy-operator | + deploy_operator logger.go:42: 12:58:35 | service-per-pod/0-deploy-operator | + kubectl -n kuttl-test-moving-shepherd apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/crd.yaml logger.go:42: 12:58:37 | service-per-pod/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlbackups.ps.percona.com serverside-applied logger.go:42: 12:58:37 | service-per-pod/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlrestores.ps.percona.com serverside-applied logger.go:42: 12:58:38 | service-per-pod/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqls.ps.percona.com serverside-applied logger.go:42: 12:58:38 | service-per-pod/0-deploy-operator | + kubectl -n kuttl-test-moving-shepherd apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/rbac.yaml logger.go:42: 12:58:39 | service-per-pod/0-deploy-operator | serviceaccount/percona-server-mysql-operator created logger.go:42: 12:58:39 | service-per-pod/0-deploy-operator | serviceaccount/percona-server-mysql-operator-orchestrator created logger.go:42: 12:58:39 | service-per-pod/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 12:58:40 | service-per-pod/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 12:58:40 | service-per-pod/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 12:58:40 | service-per-pod/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 12:58:41 | service-per-pod/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 12:58:41 | service-per-pod/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 12:58:41 | service-per-pod/0-deploy-operator | + kubectl -n kuttl-test-moving-shepherd apply -f - logger.go:42: 12:58:41 | service-per-pod/0-deploy-operator | + yq eval '(select(documentIndex==1).spec.template.spec.containers[] | select(.name=="manager").env[] | select(.name=="DISABLE_TELEMETRY").value) = "true"' logger.go:42: 12:58:41 | service-per-pod/0-deploy-operator | ++ printf 'select(documentIndex==1).spec.template.spec.containers[0].image="%s"' perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:58:41 | service-per-pod/0-deploy-operator | + yq eval 'select(documentIndex==1).spec.template.spec.containers[0].image="perconalab/percona-server-mysql-operator:PR-339-3a8483b"' /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/operator.yaml logger.go:42: 12:58:43 | service-per-pod/0-deploy-operator | configmap/percona-server-mysql-operator-config created logger.go:42: 12:58:43 | service-per-pod/0-deploy-operator | deployment.apps/percona-server-mysql-operator created logger.go:42: 12:58:43 | service-per-pod/0-deploy-operator | + deploy_non_tls_cluster_secrets logger.go:42: 12:58:43 | service-per-pod/0-deploy-operator | + kubectl -n kuttl-test-moving-shepherd apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/secrets.yaml logger.go:42: 12:58:44 | service-per-pod/0-deploy-operator | secret/test-secrets created logger.go:42: 12:58:44 | service-per-pod/0-deploy-operator | + deploy_tls_cluster_secrets logger.go:42: 12:58:44 | service-per-pod/0-deploy-operator | + kubectl -n kuttl-test-moving-shepherd apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/ssl-secret.yaml logger.go:42: 12:58:45 | service-per-pod/0-deploy-operator | secret/test-ssl created logger.go:42: 12:58:45 | service-per-pod/0-deploy-operator | + deploy_client logger.go:42: 12:58:45 | service-per-pod/0-deploy-operator | + kubectl -n kuttl-test-moving-shepherd apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/client.yaml logger.go:42: 12:58:46 | service-per-pod/0-deploy-operator | pod/mysql-client created logger.go:42: 12:58:54 | service-per-pod/0-deploy-operator | test step completed 0-deploy-operator logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | starting test step 1-create-cluster-clusterip logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | running command: [sh -c set -o errexit set -o xtrace source ../../functions get_cr \ | yq eval '.spec.mysql.clusterType="async"' - \ | yq eval '.spec.mysql.expose.enabled = true' - \ | yq eval '.spec.mysql.expose.type = "ClusterIP"' - \ | kubectl -n "${NAMESPACE}" apply -f -] logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | + source ../../functions logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ realpath ../../.. logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | ++++ mktemp -d logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ export TEMP_DIR=/tmp/tmp.2Va0LD4BJ4 logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ TEMP_DIR=/tmp/tmp.2Va0LD4BJ4 logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ GIT_BRANCH=PR-339 logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ VERSION=PR-339-3a8483b logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | ++++ which gdate logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | ++++ which date logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ date=/usr/bin/date logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ command -v oc logger.go:42: 12:58:54 | service-per-pod/1-create-cluster-clusterip | +++ oc get projects logger.go:42: 12:59:00 | service-per-pod/1-create-cluster-clusterip | error: the server doesn't have a resource type "projects" logger.go:42: 12:59:00 | service-per-pod/1-create-cluster-clusterip | +++ kubectl get nodes logger.go:42: 12:59:00 | service-per-pod/1-create-cluster-clusterip | +++ grep '^minikube' logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | ++++ pwd logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/service-per-pod logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | ++ test_name=service-per-pod logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | + get_cr logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | + local name_suffix= logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | + yq eval '.spec.mysql.expose.enabled = true' - logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | + yq eval '.spec.mysql.expose.type = "ClusterIP"' - logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | ++ printf '.metadata.name="%s"' service-per-pod logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | + yq eval '.spec.upgradeOptions.apply="disabled"' - logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | + yq eval '.spec.sslSecretName="test-ssl"' - logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | ++ printf '.spec.proxy.router.image="%s"' perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | + yq eval '.spec.proxy.router.image="perconalab/percona-server-mysql-operator:main-router"' - logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | + kubectl -n kuttl-test-moving-shepherd apply -f - logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | ++ printf '.spec.backup.image="%s"' perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | + yq eval '.spec.backup.image="perconalab/percona-server-mysql-operator:main-backup"' - logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | + yq eval '.metadata.name="service-per-pod"' /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/cr.yaml logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | + yq eval '.spec.secretsName="test-secrets"' - logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | ++ printf '.spec.initImage="%s"' perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | + yq eval '.spec.initImage="perconalab/percona-server-mysql-operator:PR-339-3a8483b"' - logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | + '[' -n '' ']' logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | + yq eval - logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | ++ printf '.spec.toolkit.image="%s"' perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | + yq eval '.spec.toolkit.image="perconalab/percona-server-mysql-operator:main-toolkit"' - logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | ++ printf '.spec.proxy.haproxy.image="%s"' perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | ++ printf '.spec.pmm.image="%s"' perconalab/pmm-client:dev-latest logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | ++ printf '.spec.mysql.image="%s"' perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | + yq eval '.spec.proxy.haproxy.image="perconalab/percona-server-mysql-operator:main-haproxy"' - logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | ++ printf '.spec.orchestrator.image="%s"' perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | + yq eval '.spec.orchestrator.image="perconalab/percona-server-mysql-operator:main-orchestrator"' - logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | + yq eval '.spec.pmm.image="perconalab/pmm-client:dev-latest"' - logger.go:42: 12:59:01 | service-per-pod/1-create-cluster-clusterip | + yq eval '.spec.mysql.image="perconalab/percona-server-mysql-operator:main-psmysql"' - logger.go:42: 12:59:02 | service-per-pod/1-create-cluster-clusterip | perconaservermysql.ps.percona.com/service-per-pod created logger.go:42: 13:02:44 | service-per-pod/1-create-cluster-clusterip | test step completed 1-create-cluster-clusterip logger.go:42: 13:02:44 | service-per-pod/2-write-data | starting test step 2-write-data logger.go:42: 13:02:44 | service-per-pod/2-write-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password" run_mysql \ "INSERT myDB.myTable (id) VALUES (100500)" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password"] logger.go:42: 13:02:44 | service-per-pod/2-write-data | + source ../../functions logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ realpath ../../.. logger.go:42: 13:02:44 | service-per-pod/2-write-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:02:44 | service-per-pod/2-write-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:02:44 | service-per-pod/2-write-data | ++++ mktemp -d logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ export TEMP_DIR=/tmp/tmp.dYrH861de9 logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ TEMP_DIR=/tmp/tmp.dYrH861de9 logger.go:42: 13:02:44 | service-per-pod/2-write-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ export GIT_BRANCH=PR-339 logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ GIT_BRANCH=PR-339 logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ export VERSION=PR-339-3a8483b logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ VERSION=PR-339-3a8483b logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:02:44 | service-per-pod/2-write-data | ++++ which gdate logger.go:42: 13:02:44 | service-per-pod/2-write-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 13:02:44 | service-per-pod/2-write-data | ++++ which date logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ date=/usr/bin/date logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ command -v oc logger.go:42: 13:02:44 | service-per-pod/2-write-data | +++ oc get projects logger.go:42: 13:02:50 | service-per-pod/2-write-data | error: the server doesn't have a resource type "projects" logger.go:42: 13:02:50 | service-per-pod/2-write-data | +++ kubectl get nodes logger.go:42: 13:02:50 | service-per-pod/2-write-data | +++ grep '^minikube' logger.go:42: 13:02:51 | service-per-pod/2-write-data | ++++ pwd logger.go:42: 13:02:51 | service-per-pod/2-write-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/service-per-pod logger.go:42: 13:02:51 | service-per-pod/2-write-data | ++ test_name=service-per-pod logger.go:42: 13:02:51 | service-per-pod/2-write-data | +++ get_cluster_name logger.go:42: 13:02:51 | service-per-pod/2-write-data | +++ kubectl -n kuttl-test-moving-shepherd get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 13:02:52 | service-per-pod/2-write-data | ++ get_haproxy_svc service-per-pod logger.go:42: 13:02:52 | service-per-pod/2-write-data | ++ local cluster=service-per-pod logger.go:42: 13:02:52 | service-per-pod/2-write-data | ++ echo service-per-pod-haproxy logger.go:42: 13:02:52 | service-per-pod/2-write-data | + run_mysql 'CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' '-h service-per-pod-haproxy -uroot -proot_password' logger.go:42: 13:02:52 | service-per-pod/2-write-data | + local 'command=CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' logger.go:42: 13:02:52 | service-per-pod/2-write-data | + local 'uri=-h service-per-pod-haproxy -uroot -proot_password' logger.go:42: 13:02:52 | service-per-pod/2-write-data | + local pod= logger.go:42: 13:02:52 | service-per-pod/2-write-data | ++ get_client_pod logger.go:42: 13:02:52 | service-per-pod/2-write-data | ++ kubectl -n kuttl-test-moving-shepherd get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:02:52 | service-per-pod/2-write-data | + client_pod=mysql-client logger.go:42: 13:02:52 | service-per-pod/2-write-data | + wait_pod mysql-client logger.go:42: 13:02:52 | service-per-pod/2-write-data | + local pod=mysql-client logger.go:42: 13:02:52 | service-per-pod/2-write-data | + set +o xtrace logger.go:42: 13:02:53 | service-per-pod/2-write-data | mysql-clienttrue logger.go:42: 13:02:53 | service-per-pod/2-write-data | + sed -e 's/mysql: //' logger.go:42: 13:02:53 | service-per-pod/2-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:02:53 | service-per-pod/2-write-data | + kubectl -n kuttl-test-moving-shepherd exec mysql-client -- bash -c 'printf '\''%s\n'\'' "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" | mysql -sN -h service-per-pod-haproxy -uroot -proot_password' logger.go:42: 13:02:55 | service-per-pod/2-write-data | + : logger.go:42: 13:02:55 | service-per-pod/2-write-data | +++ get_cluster_name logger.go:42: 13:02:55 | service-per-pod/2-write-data | +++ kubectl -n kuttl-test-moving-shepherd get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 13:02:55 | service-per-pod/2-write-data | ++ get_haproxy_svc service-per-pod logger.go:42: 13:02:55 | service-per-pod/2-write-data | ++ local cluster=service-per-pod logger.go:42: 13:02:55 | service-per-pod/2-write-data | ++ echo service-per-pod-haproxy logger.go:42: 13:02:55 | service-per-pod/2-write-data | + run_mysql 'INSERT myDB.myTable (id) VALUES (100500)' '-h service-per-pod-haproxy -uroot -proot_password' logger.go:42: 13:02:55 | service-per-pod/2-write-data | + local 'command=INSERT myDB.myTable (id) VALUES (100500)' logger.go:42: 13:02:55 | service-per-pod/2-write-data | + local 'uri=-h service-per-pod-haproxy -uroot -proot_password' logger.go:42: 13:02:55 | service-per-pod/2-write-data | + local pod= logger.go:42: 13:02:55 | service-per-pod/2-write-data | ++ get_client_pod logger.go:42: 13:02:55 | service-per-pod/2-write-data | ++ kubectl -n kuttl-test-moving-shepherd get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:02:56 | service-per-pod/2-write-data | + client_pod=mysql-client logger.go:42: 13:02:56 | service-per-pod/2-write-data | + wait_pod mysql-client logger.go:42: 13:02:56 | service-per-pod/2-write-data | + local pod=mysql-client logger.go:42: 13:02:56 | service-per-pod/2-write-data | + set +o xtrace logger.go:42: 13:02:57 | service-per-pod/2-write-data | mysql-clienttrue logger.go:42: 13:02:57 | service-per-pod/2-write-data | + sed -e 's/mysql: //' logger.go:42: 13:02:57 | service-per-pod/2-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:02:57 | service-per-pod/2-write-data | + kubectl -n kuttl-test-moving-shepherd exec mysql-client -- bash -c 'printf '\''%s\n'\'' "INSERT myDB.myTable (id) VALUES (100500)" | mysql -sN -h service-per-pod-haproxy -uroot -proot_password' logger.go:42: 13:02:58 | service-per-pod/2-write-data | + : logger.go:42: 13:03:00 | service-per-pod/2-write-data | test step completed 2-write-data logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | starting test step 3-read-from-clusterip logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | running command: [sh -c set -o errexit set -o xtrace source ../../functions args='' size=$(kubectl -n ${NAMESPACE} get ps ${test_name} -o jsonpath='{.spec.mysql.size}') for i in $(seq 0 $((size - 1))); do ip=$(get_service_ip ${test_name}-mysql-${i}) host=$(get_mysql_headless_fqdn ${test_name} $i) data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${ip} -uroot -proot_password") args="${args} --from-literal=${host}=${data}" done kubectl create configmap -n "${NAMESPACE}" 04-read-from-clusterip ${args}] logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | + source ../../functions logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ realpath ../../.. logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | ++++ mktemp -d logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ export TEMP_DIR=/tmp/tmp.h3cJPPqvnZ logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ TEMP_DIR=/tmp/tmp.h3cJPPqvnZ logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ export GIT_BRANCH=PR-339 logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ GIT_BRANCH=PR-339 logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ export VERSION=PR-339-3a8483b logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ VERSION=PR-339-3a8483b logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | ++++ which gdate logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | ++++ which date logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ date=/usr/bin/date logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ command -v oc logger.go:42: 13:03:00 | service-per-pod/3-read-from-clusterip | +++ oc get projects logger.go:42: 13:03:05 | service-per-pod/3-read-from-clusterip | error: the server doesn't have a resource type "projects" logger.go:42: 13:03:05 | service-per-pod/3-read-from-clusterip | +++ kubectl get nodes logger.go:42: 13:03:05 | service-per-pod/3-read-from-clusterip | +++ grep '^minikube' logger.go:42: 13:03:06 | service-per-pod/3-read-from-clusterip | ++++ pwd logger.go:42: 13:03:06 | service-per-pod/3-read-from-clusterip | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/service-per-pod logger.go:42: 13:03:06 | service-per-pod/3-read-from-clusterip | ++ test_name=service-per-pod logger.go:42: 13:03:06 | service-per-pod/3-read-from-clusterip | + args= logger.go:42: 13:03:06 | service-per-pod/3-read-from-clusterip | ++ kubectl -n kuttl-test-moving-shepherd get ps service-per-pod -o 'jsonpath={.spec.mysql.size}' logger.go:42: 13:03:07 | service-per-pod/3-read-from-clusterip | + size=3 logger.go:42: 13:03:07 | service-per-pod/3-read-from-clusterip | ++ seq 0 2 logger.go:42: 13:03:07 | service-per-pod/3-read-from-clusterip | + for i in '$(seq 0 $((size - 1)))' logger.go:42: 13:03:07 | service-per-pod/3-read-from-clusterip | ++ get_service_ip service-per-pod-mysql-0 logger.go:42: 13:03:07 | service-per-pod/3-read-from-clusterip | ++ local service=service-per-pod-mysql-0 logger.go:42: 13:03:07 | service-per-pod/3-read-from-clusterip | ++ kubectl get service/service-per-pod-mysql-0 -n kuttl-test-moving-shepherd -o 'jsonpath={.spec.type}' logger.go:42: 13:03:07 | service-per-pod/3-read-from-clusterip | ++ grep -q NotFound logger.go:42: 13:03:08 | service-per-pod/3-read-from-clusterip | +++ kubectl get service/service-per-pod-mysql-0 -n kuttl-test-moving-shepherd -o 'jsonpath={.spec.type}' logger.go:42: 13:03:08 | service-per-pod/3-read-from-clusterip | ++ '[' ClusterIP = ClusterIP ']' logger.go:42: 13:03:08 | service-per-pod/3-read-from-clusterip | ++ kubectl get service/service-per-pod-mysql-0 -n kuttl-test-moving-shepherd -o 'jsonpath={.spec.clusterIP}' logger.go:42: 13:03:09 | service-per-pod/3-read-from-clusterip | ++ return logger.go:42: 13:03:09 | service-per-pod/3-read-from-clusterip | + ip=10.21.85.135 logger.go:42: 13:03:09 | service-per-pod/3-read-from-clusterip | ++ get_mysql_headless_fqdn service-per-pod 0 logger.go:42: 13:03:09 | service-per-pod/3-read-from-clusterip | ++ local cluster=service-per-pod logger.go:42: 13:03:09 | service-per-pod/3-read-from-clusterip | ++ local index=0 logger.go:42: 13:03:09 | service-per-pod/3-read-from-clusterip | ++ echo service-per-pod-mysql-0.service-per-pod-mysql logger.go:42: 13:03:09 | service-per-pod/3-read-from-clusterip | + host=service-per-pod-mysql-0.service-per-pod-mysql logger.go:42: 13:03:09 | service-per-pod/3-read-from-clusterip | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h 10.21.85.135 -uroot -proot_password' logger.go:42: 13:03:09 | service-per-pod/3-read-from-clusterip | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 13:03:09 | service-per-pod/3-read-from-clusterip | ++ local 'uri=-h 10.21.85.135 -uroot -proot_password' logger.go:42: 13:03:09 | service-per-pod/3-read-from-clusterip | ++ local pod= logger.go:42: 13:03:09 | service-per-pod/3-read-from-clusterip | +++ get_client_pod logger.go:42: 13:03:09 | service-per-pod/3-read-from-clusterip | +++ kubectl -n kuttl-test-moving-shepherd get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:03:10 | service-per-pod/3-read-from-clusterip | ++ client_pod=mysql-client logger.go:42: 13:03:10 | service-per-pod/3-read-from-clusterip | ++ wait_pod mysql-client logger.go:42: 13:03:10 | service-per-pod/3-read-from-clusterip | ++ local pod=mysql-client logger.go:42: 13:03:10 | service-per-pod/3-read-from-clusterip | ++ set +o xtrace logger.go:42: 13:03:10 | service-per-pod/3-read-from-clusterip | mysql-clienttrue logger.go:42: 13:03:10 | service-per-pod/3-read-from-clusterip | ++ kubectl -n kuttl-test-moving-shepherd exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h 10.21.85.135 -uroot -proot_password' logger.go:42: 13:03:10 | service-per-pod/3-read-from-clusterip | ++ sed -e 's/mysql: //' logger.go:42: 13:03:10 | service-per-pod/3-read-from-clusterip | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:03:12 | service-per-pod/3-read-from-clusterip | + data=100500 logger.go:42: 13:03:12 | service-per-pod/3-read-from-clusterip | + args=' --from-literal=service-per-pod-mysql-0.service-per-pod-mysql=100500' logger.go:42: 13:03:12 | service-per-pod/3-read-from-clusterip | + for i in '$(seq 0 $((size - 1)))' logger.go:42: 13:03:12 | service-per-pod/3-read-from-clusterip | ++ get_service_ip service-per-pod-mysql-1 logger.go:42: 13:03:12 | service-per-pod/3-read-from-clusterip | ++ local service=service-per-pod-mysql-1 logger.go:42: 13:03:12 | service-per-pod/3-read-from-clusterip | ++ kubectl get service/service-per-pod-mysql-1 -n kuttl-test-moving-shepherd -o 'jsonpath={.spec.type}' logger.go:42: 13:03:12 | service-per-pod/3-read-from-clusterip | ++ grep -q NotFound logger.go:42: 13:03:13 | service-per-pod/3-read-from-clusterip | +++ kubectl get service/service-per-pod-mysql-1 -n kuttl-test-moving-shepherd -o 'jsonpath={.spec.type}' logger.go:42: 13:03:13 | service-per-pod/3-read-from-clusterip | ++ '[' ClusterIP = ClusterIP ']' logger.go:42: 13:03:13 | service-per-pod/3-read-from-clusterip | ++ kubectl get service/service-per-pod-mysql-1 -n kuttl-test-moving-shepherd -o 'jsonpath={.spec.clusterIP}' logger.go:42: 13:03:14 | service-per-pod/3-read-from-clusterip | ++ return logger.go:42: 13:03:14 | service-per-pod/3-read-from-clusterip | + ip=10.21.92.60 logger.go:42: 13:03:14 | service-per-pod/3-read-from-clusterip | ++ get_mysql_headless_fqdn service-per-pod 1 logger.go:42: 13:03:14 | service-per-pod/3-read-from-clusterip | ++ local cluster=service-per-pod logger.go:42: 13:03:14 | service-per-pod/3-read-from-clusterip | ++ local index=1 logger.go:42: 13:03:14 | service-per-pod/3-read-from-clusterip | ++ echo service-per-pod-mysql-1.service-per-pod-mysql logger.go:42: 13:03:14 | service-per-pod/3-read-from-clusterip | + host=service-per-pod-mysql-1.service-per-pod-mysql logger.go:42: 13:03:14 | service-per-pod/3-read-from-clusterip | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h 10.21.92.60 -uroot -proot_password' logger.go:42: 13:03:14 | service-per-pod/3-read-from-clusterip | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 13:03:14 | service-per-pod/3-read-from-clusterip | ++ local 'uri=-h 10.21.92.60 -uroot -proot_password' logger.go:42: 13:03:14 | service-per-pod/3-read-from-clusterip | ++ local pod= logger.go:42: 13:03:14 | service-per-pod/3-read-from-clusterip | +++ get_client_pod logger.go:42: 13:03:14 | service-per-pod/3-read-from-clusterip | +++ kubectl -n kuttl-test-moving-shepherd get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:03:15 | service-per-pod/3-read-from-clusterip | ++ client_pod=mysql-client logger.go:42: 13:03:15 | service-per-pod/3-read-from-clusterip | ++ wait_pod mysql-client logger.go:42: 13:03:15 | service-per-pod/3-read-from-clusterip | ++ local pod=mysql-client logger.go:42: 13:03:15 | service-per-pod/3-read-from-clusterip | ++ set +o xtrace logger.go:42: 13:03:16 | service-per-pod/3-read-from-clusterip | mysql-clienttrue logger.go:42: 13:03:16 | service-per-pod/3-read-from-clusterip | ++ kubectl -n kuttl-test-moving-shepherd exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h 10.21.92.60 -uroot -proot_password' logger.go:42: 13:03:16 | service-per-pod/3-read-from-clusterip | ++ sed -e 's/mysql: //' logger.go:42: 13:03:16 | service-per-pod/3-read-from-clusterip | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:03:17 | service-per-pod/3-read-from-clusterip | + data=100500 logger.go:42: 13:03:17 | service-per-pod/3-read-from-clusterip | + args=' --from-literal=service-per-pod-mysql-0.service-per-pod-mysql=100500 --from-literal=service-per-pod-mysql-1.service-per-pod-mysql=100500' logger.go:42: 13:03:17 | service-per-pod/3-read-from-clusterip | + for i in '$(seq 0 $((size - 1)))' logger.go:42: 13:03:17 | service-per-pod/3-read-from-clusterip | ++ get_service_ip service-per-pod-mysql-2 logger.go:42: 13:03:17 | service-per-pod/3-read-from-clusterip | ++ local service=service-per-pod-mysql-2 logger.go:42: 13:03:17 | service-per-pod/3-read-from-clusterip | ++ kubectl get service/service-per-pod-mysql-2 -n kuttl-test-moving-shepherd -o 'jsonpath={.spec.type}' logger.go:42: 13:03:17 | service-per-pod/3-read-from-clusterip | ++ grep -q NotFound logger.go:42: 13:03:18 | service-per-pod/3-read-from-clusterip | +++ kubectl get service/service-per-pod-mysql-2 -n kuttl-test-moving-shepherd -o 'jsonpath={.spec.type}' logger.go:42: 13:03:18 | service-per-pod/3-read-from-clusterip | ++ '[' ClusterIP = ClusterIP ']' logger.go:42: 13:03:18 | service-per-pod/3-read-from-clusterip | ++ kubectl get service/service-per-pod-mysql-2 -n kuttl-test-moving-shepherd -o 'jsonpath={.spec.clusterIP}' logger.go:42: 13:03:19 | service-per-pod/3-read-from-clusterip | ++ return logger.go:42: 13:03:19 | service-per-pod/3-read-from-clusterip | + ip=10.21.90.24 logger.go:42: 13:03:19 | service-per-pod/3-read-from-clusterip | ++ get_mysql_headless_fqdn service-per-pod 2 logger.go:42: 13:03:19 | service-per-pod/3-read-from-clusterip | ++ local cluster=service-per-pod logger.go:42: 13:03:19 | service-per-pod/3-read-from-clusterip | ++ local index=2 logger.go:42: 13:03:19 | service-per-pod/3-read-from-clusterip | ++ echo service-per-pod-mysql-2.service-per-pod-mysql logger.go:42: 13:03:19 | service-per-pod/3-read-from-clusterip | + host=service-per-pod-mysql-2.service-per-pod-mysql logger.go:42: 13:03:19 | service-per-pod/3-read-from-clusterip | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h 10.21.90.24 -uroot -proot_password' logger.go:42: 13:03:19 | service-per-pod/3-read-from-clusterip | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 13:03:19 | service-per-pod/3-read-from-clusterip | ++ local 'uri=-h 10.21.90.24 -uroot -proot_password' logger.go:42: 13:03:19 | service-per-pod/3-read-from-clusterip | ++ local pod= logger.go:42: 13:03:19 | service-per-pod/3-read-from-clusterip | +++ get_client_pod logger.go:42: 13:03:19 | service-per-pod/3-read-from-clusterip | +++ kubectl -n kuttl-test-moving-shepherd get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:03:20 | service-per-pod/3-read-from-clusterip | ++ client_pod=mysql-client logger.go:42: 13:03:20 | service-per-pod/3-read-from-clusterip | ++ wait_pod mysql-client logger.go:42: 13:03:20 | service-per-pod/3-read-from-clusterip | ++ local pod=mysql-client logger.go:42: 13:03:20 | service-per-pod/3-read-from-clusterip | ++ set +o xtrace logger.go:42: 13:03:21 | service-per-pod/3-read-from-clusterip | mysql-clienttrue logger.go:42: 13:03:21 | service-per-pod/3-read-from-clusterip | ++ kubectl -n kuttl-test-moving-shepherd exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h 10.21.90.24 -uroot -proot_password' logger.go:42: 13:03:21 | service-per-pod/3-read-from-clusterip | ++ sed -e 's/mysql: //' logger.go:42: 13:03:21 | service-per-pod/3-read-from-clusterip | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:03:22 | service-per-pod/3-read-from-clusterip | + data=100500 logger.go:42: 13:03:22 | service-per-pod/3-read-from-clusterip | + args=' --from-literal=service-per-pod-mysql-0.service-per-pod-mysql=100500 --from-literal=service-per-pod-mysql-1.service-per-pod-mysql=100500 --from-literal=service-per-pod-mysql-2.service-per-pod-mysql=100500' logger.go:42: 13:03:22 | service-per-pod/3-read-from-clusterip | + kubectl create configmap -n kuttl-test-moving-shepherd 04-read-from-clusterip --from-literal=service-per-pod-mysql-0.service-per-pod-mysql=100500 --from-literal=service-per-pod-mysql-1.service-per-pod-mysql=100500 --from-literal=service-per-pod-mysql-2.service-per-pod-mysql=100500 logger.go:42: 13:03:23 | service-per-pod/3-read-from-clusterip | configmap/04-read-from-clusterip created logger.go:42: 13:03:24 | service-per-pod/3-read-from-clusterip | test step completed 3-read-from-clusterip logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | starting test step 4-create-cluster-loadbalancer logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | running command: [sh -c set -o errexit set -o xtrace source ../../functions get_cr \ | yq eval '.spec.mysql.clusterType="async"' - \ | yq eval '.spec.mysql.expose.enabled = true' - \ | yq eval '.spec.mysql.expose.type = "LoadBalancer"' - \ | kubectl -n "${NAMESPACE}" apply -f -] logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | + source ../../functions logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ realpath ../../.. logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | ++++ mktemp -d logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ export TEMP_DIR=/tmp/tmp.OF3958kPrZ logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ TEMP_DIR=/tmp/tmp.OF3958kPrZ logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ export GIT_BRANCH=PR-339 logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ GIT_BRANCH=PR-339 logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ export VERSION=PR-339-3a8483b logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ VERSION=PR-339-3a8483b logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | ++++ which gdate logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | ++++ which date logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ date=/usr/bin/date logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ command -v oc logger.go:42: 13:03:24 | service-per-pod/4-create-cluster-loadbalancer | +++ oc get projects logger.go:42: 13:03:30 | service-per-pod/4-create-cluster-loadbalancer | error: the server doesn't have a resource type "projects" logger.go:42: 13:03:30 | service-per-pod/4-create-cluster-loadbalancer | +++ kubectl get nodes logger.go:42: 13:03:30 | service-per-pod/4-create-cluster-loadbalancer | +++ grep '^minikube' logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | ++++ pwd logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/service-per-pod logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | ++ test_name=service-per-pod logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | + get_cr logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | + local name_suffix= logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | + yq eval '.spec.mysql.expose.type = "LoadBalancer"' - logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | ++ printf '.spec.backup.image="%s"' perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | ++ printf '.spec.initImage="%s"' perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | + yq eval '.spec.mysql.expose.enabled = true' - logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | + yq eval '.spec.backup.image="perconalab/percona-server-mysql-operator:main-backup"' - logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | ++ printf '.spec.proxy.haproxy.image="%s"' perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | + yq eval '.spec.initImage="perconalab/percona-server-mysql-operator:PR-339-3a8483b"' - logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | ++ printf '.spec.pmm.image="%s"' perconalab/pmm-client:dev-latest logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | + yq eval '.spec.proxy.haproxy.image="perconalab/percona-server-mysql-operator:main-haproxy"' - logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | + yq eval '.spec.pmm.image="perconalab/pmm-client:dev-latest"' - logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | + kubectl -n kuttl-test-moving-shepherd apply -f - logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | + '[' -n '' ']' logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | + yq eval - logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | ++ printf '.spec.mysql.image="%s"' perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | + yq eval '.spec.mysql.image="perconalab/percona-server-mysql-operator:main-psmysql"' - logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | ++ printf '.spec.proxy.router.image="%s"' perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | + yq eval '.spec.secretsName="test-secrets"' - logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | + yq eval '.spec.proxy.router.image="perconalab/percona-server-mysql-operator:main-router"' - logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | ++ printf '.metadata.name="%s"' service-per-pod logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | + yq eval '.metadata.name="service-per-pod"' /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/cr.yaml logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | + yq eval '.spec.sslSecretName="test-ssl"' - logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | + yq eval '.spec.upgradeOptions.apply="disabled"' - logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | ++ printf '.spec.orchestrator.image="%s"' perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | + yq eval '.spec.orchestrator.image="perconalab/percona-server-mysql-operator:main-orchestrator"' - logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | ++ printf '.spec.toolkit.image="%s"' perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:03:31 | service-per-pod/4-create-cluster-loadbalancer | + yq eval '.spec.toolkit.image="perconalab/percona-server-mysql-operator:main-toolkit"' - logger.go:42: 13:03:32 | service-per-pod/4-create-cluster-loadbalancer | perconaservermysql.ps.percona.com/service-per-pod configured logger.go:42: 13:03:43 | service-per-pod/4-create-cluster-loadbalancer | test step completed 4-create-cluster-loadbalancer logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | starting test step 5-read-from-loadbalancer logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | running command: [sh -c set -o errexit set -o xtrace source ../../functions args='' size=$(kubectl -n ${NAMESPACE} get ps ${test_name} -o jsonpath='{.spec.mysql.size}') sleep 75 for i in $(seq 0 $((size - 1))); do ip=$(get_service_ip ${test_name}-mysql-$i) host=$(get_mysql_headless_fqdn ${test_name} $i) data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${ip} -uroot -proot_password") args="${args} --from-literal=${host}=${data}" done kubectl create configmap -n "${NAMESPACE}" 04-read-from-loadbalancer ${args}] logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | + source ../../functions logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ realpath ../../.. logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | ++++ mktemp -d logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ export TEMP_DIR=/tmp/tmp.UULhkhwQec logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ TEMP_DIR=/tmp/tmp.UULhkhwQec logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ export GIT_BRANCH=PR-339 logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ GIT_BRANCH=PR-339 logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ export VERSION=PR-339-3a8483b logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ VERSION=PR-339-3a8483b logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | ++++ which gdate logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | ++++ which date logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ date=/usr/bin/date logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ command -v oc logger.go:42: 13:03:43 | service-per-pod/5-read-from-loadbalancer | +++ oc get projects logger.go:42: 13:03:49 | service-per-pod/5-read-from-loadbalancer | error: the server doesn't have a resource type "projects" logger.go:42: 13:03:49 | service-per-pod/5-read-from-loadbalancer | +++ kubectl get nodes logger.go:42: 13:03:49 | service-per-pod/5-read-from-loadbalancer | +++ grep '^minikube' logger.go:42: 13:03:49 | service-per-pod/5-read-from-loadbalancer | ++++ pwd logger.go:42: 13:03:49 | service-per-pod/5-read-from-loadbalancer | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/service-per-pod logger.go:42: 13:03:49 | service-per-pod/5-read-from-loadbalancer | ++ test_name=service-per-pod logger.go:42: 13:03:49 | service-per-pod/5-read-from-loadbalancer | + args= logger.go:42: 13:03:49 | service-per-pod/5-read-from-loadbalancer | ++ kubectl -n kuttl-test-moving-shepherd get ps service-per-pod -o 'jsonpath={.spec.mysql.size}' logger.go:42: 13:03:50 | service-per-pod/5-read-from-loadbalancer | + size=3 logger.go:42: 13:03:50 | service-per-pod/5-read-from-loadbalancer | + sleep 75 logger.go:42: 13:05:05 | service-per-pod/5-read-from-loadbalancer | ++ seq 0 2 logger.go:42: 13:05:05 | service-per-pod/5-read-from-loadbalancer | + for i in '$(seq 0 $((size - 1)))' logger.go:42: 13:05:05 | service-per-pod/5-read-from-loadbalancer | ++ get_service_ip service-per-pod-mysql-0 logger.go:42: 13:05:05 | service-per-pod/5-read-from-loadbalancer | ++ local service=service-per-pod-mysql-0 logger.go:42: 13:05:05 | service-per-pod/5-read-from-loadbalancer | ++ kubectl get service/service-per-pod-mysql-0 -n kuttl-test-moving-shepherd -o 'jsonpath={.spec.type}' logger.go:42: 13:05:05 | service-per-pod/5-read-from-loadbalancer | ++ grep -q NotFound logger.go:42: 13:05:06 | service-per-pod/5-read-from-loadbalancer | +++ kubectl get service/service-per-pod-mysql-0 -n kuttl-test-moving-shepherd -o 'jsonpath={.spec.type}' logger.go:42: 13:05:07 | service-per-pod/5-read-from-loadbalancer | ++ '[' LoadBalancer = ClusterIP ']' logger.go:42: 13:05:07 | service-per-pod/5-read-from-loadbalancer | ++ kubectl get service/service-per-pod-mysql-0 -n kuttl-test-moving-shepherd -o 'jsonpath={.status.loadBalancer.ingress[]}' logger.go:42: 13:05:07 | service-per-pod/5-read-from-loadbalancer | ++ egrep -q 'hostname|ip' logger.go:42: 13:05:07 | service-per-pod/5-read-from-loadbalancer | ++ kubectl get service/service-per-pod-mysql-0 -n kuttl-test-moving-shepherd -o 'jsonpath={.status.loadBalancer.ingress[].ip}' logger.go:42: 13:05:08 | service-per-pod/5-read-from-loadbalancer | ++ kubectl get service/service-per-pod-mysql-0 -n kuttl-test-moving-shepherd -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' logger.go:42: 13:05:09 | service-per-pod/5-read-from-loadbalancer | + ip=35.202.145.56 logger.go:42: 13:05:09 | service-per-pod/5-read-from-loadbalancer | ++ get_mysql_headless_fqdn service-per-pod 0 logger.go:42: 13:05:09 | service-per-pod/5-read-from-loadbalancer | ++ local cluster=service-per-pod logger.go:42: 13:05:09 | service-per-pod/5-read-from-loadbalancer | ++ local index=0 logger.go:42: 13:05:09 | service-per-pod/5-read-from-loadbalancer | ++ echo service-per-pod-mysql-0.service-per-pod-mysql logger.go:42: 13:05:09 | service-per-pod/5-read-from-loadbalancer | + host=service-per-pod-mysql-0.service-per-pod-mysql logger.go:42: 13:05:09 | service-per-pod/5-read-from-loadbalancer | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h 35.202.145.56 -uroot -proot_password' logger.go:42: 13:05:09 | service-per-pod/5-read-from-loadbalancer | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 13:05:09 | service-per-pod/5-read-from-loadbalancer | ++ local 'uri=-h 35.202.145.56 -uroot -proot_password' logger.go:42: 13:05:09 | service-per-pod/5-read-from-loadbalancer | ++ local pod= logger.go:42: 13:05:09 | service-per-pod/5-read-from-loadbalancer | +++ get_client_pod logger.go:42: 13:05:09 | service-per-pod/5-read-from-loadbalancer | +++ kubectl -n kuttl-test-moving-shepherd get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:05:10 | service-per-pod/5-read-from-loadbalancer | ++ client_pod=mysql-client logger.go:42: 13:05:10 | service-per-pod/5-read-from-loadbalancer | ++ wait_pod mysql-client logger.go:42: 13:05:10 | service-per-pod/5-read-from-loadbalancer | ++ local pod=mysql-client logger.go:42: 13:05:10 | service-per-pod/5-read-from-loadbalancer | ++ set +o xtrace logger.go:42: 13:05:10 | service-per-pod/5-read-from-loadbalancer | mysql-clienttrue logger.go:42: 13:05:10 | service-per-pod/5-read-from-loadbalancer | ++ sed -e 's/mysql: //' logger.go:42: 13:05:10 | service-per-pod/5-read-from-loadbalancer | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:05:10 | service-per-pod/5-read-from-loadbalancer | ++ kubectl -n kuttl-test-moving-shepherd exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h 35.202.145.56 -uroot -proot_password' logger.go:42: 13:05:12 | service-per-pod/5-read-from-loadbalancer | + data=100500 logger.go:42: 13:05:12 | service-per-pod/5-read-from-loadbalancer | + args=' --from-literal=service-per-pod-mysql-0.service-per-pod-mysql=100500' logger.go:42: 13:05:12 | service-per-pod/5-read-from-loadbalancer | + for i in '$(seq 0 $((size - 1)))' logger.go:42: 13:05:12 | service-per-pod/5-read-from-loadbalancer | ++ get_service_ip service-per-pod-mysql-1 logger.go:42: 13:05:12 | service-per-pod/5-read-from-loadbalancer | ++ local service=service-per-pod-mysql-1 logger.go:42: 13:05:12 | service-per-pod/5-read-from-loadbalancer | ++ kubectl get service/service-per-pod-mysql-1 -n kuttl-test-moving-shepherd -o 'jsonpath={.spec.type}' logger.go:42: 13:05:12 | service-per-pod/5-read-from-loadbalancer | ++ grep -q NotFound logger.go:42: 13:05:12 | service-per-pod/5-read-from-loadbalancer | +++ kubectl get service/service-per-pod-mysql-1 -n kuttl-test-moving-shepherd -o 'jsonpath={.spec.type}' logger.go:42: 13:05:13 | service-per-pod/5-read-from-loadbalancer | ++ '[' LoadBalancer = ClusterIP ']' logger.go:42: 13:05:13 | service-per-pod/5-read-from-loadbalancer | ++ kubectl get service/service-per-pod-mysql-1 -n kuttl-test-moving-shepherd -o 'jsonpath={.status.loadBalancer.ingress[]}' logger.go:42: 13:05:13 | service-per-pod/5-read-from-loadbalancer | ++ egrep -q 'hostname|ip' logger.go:42: 13:05:14 | service-per-pod/5-read-from-loadbalancer | ++ kubectl get service/service-per-pod-mysql-1 -n kuttl-test-moving-shepherd -o 'jsonpath={.status.loadBalancer.ingress[].ip}' logger.go:42: 13:05:15 | service-per-pod/5-read-from-loadbalancer | ++ kubectl get service/service-per-pod-mysql-1 -n kuttl-test-moving-shepherd -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' logger.go:42: 13:05:15 | service-per-pod/5-read-from-loadbalancer | + ip=35.222.183.124 logger.go:42: 13:05:15 | service-per-pod/5-read-from-loadbalancer | ++ get_mysql_headless_fqdn service-per-pod 1 logger.go:42: 13:05:15 | service-per-pod/5-read-from-loadbalancer | ++ local cluster=service-per-pod logger.go:42: 13:05:15 | service-per-pod/5-read-from-loadbalancer | ++ local index=1 logger.go:42: 13:05:15 | service-per-pod/5-read-from-loadbalancer | ++ echo service-per-pod-mysql-1.service-per-pod-mysql logger.go:42: 13:05:15 | service-per-pod/5-read-from-loadbalancer | + host=service-per-pod-mysql-1.service-per-pod-mysql logger.go:42: 13:05:15 | service-per-pod/5-read-from-loadbalancer | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h 35.222.183.124 -uroot -proot_password' logger.go:42: 13:05:15 | service-per-pod/5-read-from-loadbalancer | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 13:05:15 | service-per-pod/5-read-from-loadbalancer | ++ local 'uri=-h 35.222.183.124 -uroot -proot_password' logger.go:42: 13:05:15 | service-per-pod/5-read-from-loadbalancer | ++ local pod= logger.go:42: 13:05:15 | service-per-pod/5-read-from-loadbalancer | +++ get_client_pod logger.go:42: 13:05:15 | service-per-pod/5-read-from-loadbalancer | +++ kubectl -n kuttl-test-moving-shepherd get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:05:16 | service-per-pod/5-read-from-loadbalancer | ++ client_pod=mysql-client logger.go:42: 13:05:16 | service-per-pod/5-read-from-loadbalancer | ++ wait_pod mysql-client logger.go:42: 13:05:16 | service-per-pod/5-read-from-loadbalancer | ++ local pod=mysql-client logger.go:42: 13:05:16 | service-per-pod/5-read-from-loadbalancer | ++ set +o xtrace logger.go:42: 13:05:17 | service-per-pod/5-read-from-loadbalancer | mysql-clienttrue logger.go:42: 13:05:17 | service-per-pod/5-read-from-loadbalancer | ++ kubectl -n kuttl-test-moving-shepherd exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h 35.222.183.124 -uroot -proot_password' logger.go:42: 13:05:17 | service-per-pod/5-read-from-loadbalancer | ++ sed -e 's/mysql: //' logger.go:42: 13:05:17 | service-per-pod/5-read-from-loadbalancer | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:05:18 | service-per-pod/5-read-from-loadbalancer | + data=100500 logger.go:42: 13:05:18 | service-per-pod/5-read-from-loadbalancer | + args=' --from-literal=service-per-pod-mysql-0.service-per-pod-mysql=100500 --from-literal=service-per-pod-mysql-1.service-per-pod-mysql=100500' logger.go:42: 13:05:18 | service-per-pod/5-read-from-loadbalancer | + for i in '$(seq 0 $((size - 1)))' logger.go:42: 13:05:18 | service-per-pod/5-read-from-loadbalancer | ++ get_service_ip service-per-pod-mysql-2 logger.go:42: 13:05:18 | service-per-pod/5-read-from-loadbalancer | ++ local service=service-per-pod-mysql-2 logger.go:42: 13:05:18 | service-per-pod/5-read-from-loadbalancer | ++ kubectl get service/service-per-pod-mysql-2 -n kuttl-test-moving-shepherd -o 'jsonpath={.spec.type}' logger.go:42: 13:05:18 | service-per-pod/5-read-from-loadbalancer | ++ grep -q NotFound logger.go:42: 13:05:19 | service-per-pod/5-read-from-loadbalancer | +++ kubectl get service/service-per-pod-mysql-2 -n kuttl-test-moving-shepherd -o 'jsonpath={.spec.type}' logger.go:42: 13:05:20 | service-per-pod/5-read-from-loadbalancer | ++ '[' LoadBalancer = ClusterIP ']' logger.go:42: 13:05:20 | service-per-pod/5-read-from-loadbalancer | ++ kubectl get service/service-per-pod-mysql-2 -n kuttl-test-moving-shepherd -o 'jsonpath={.status.loadBalancer.ingress[]}' logger.go:42: 13:05:20 | service-per-pod/5-read-from-loadbalancer | ++ egrep -q 'hostname|ip' logger.go:42: 13:05:20 | service-per-pod/5-read-from-loadbalancer | ++ kubectl get service/service-per-pod-mysql-2 -n kuttl-test-moving-shepherd -o 'jsonpath={.status.loadBalancer.ingress[].ip}' logger.go:42: 13:05:21 | service-per-pod/5-read-from-loadbalancer | ++ kubectl get service/service-per-pod-mysql-2 -n kuttl-test-moving-shepherd -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' logger.go:42: 13:05:22 | service-per-pod/5-read-from-loadbalancer | + ip=34.132.221.154 logger.go:42: 13:05:22 | service-per-pod/5-read-from-loadbalancer | ++ get_mysql_headless_fqdn service-per-pod 2 logger.go:42: 13:05:22 | service-per-pod/5-read-from-loadbalancer | ++ local cluster=service-per-pod logger.go:42: 13:05:22 | service-per-pod/5-read-from-loadbalancer | ++ local index=2 logger.go:42: 13:05:22 | service-per-pod/5-read-from-loadbalancer | ++ echo service-per-pod-mysql-2.service-per-pod-mysql logger.go:42: 13:05:22 | service-per-pod/5-read-from-loadbalancer | + host=service-per-pod-mysql-2.service-per-pod-mysql logger.go:42: 13:05:22 | service-per-pod/5-read-from-loadbalancer | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h 34.132.221.154 -uroot -proot_password' logger.go:42: 13:05:22 | service-per-pod/5-read-from-loadbalancer | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 13:05:22 | service-per-pod/5-read-from-loadbalancer | ++ local 'uri=-h 34.132.221.154 -uroot -proot_password' logger.go:42: 13:05:22 | service-per-pod/5-read-from-loadbalancer | ++ local pod= logger.go:42: 13:05:22 | service-per-pod/5-read-from-loadbalancer | +++ get_client_pod logger.go:42: 13:05:22 | service-per-pod/5-read-from-loadbalancer | +++ kubectl -n kuttl-test-moving-shepherd get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:05:23 | service-per-pod/5-read-from-loadbalancer | ++ client_pod=mysql-client logger.go:42: 13:05:23 | service-per-pod/5-read-from-loadbalancer | ++ wait_pod mysql-client logger.go:42: 13:05:23 | service-per-pod/5-read-from-loadbalancer | ++ local pod=mysql-client logger.go:42: 13:05:23 | service-per-pod/5-read-from-loadbalancer | ++ set +o xtrace logger.go:42: 13:05:23 | service-per-pod/5-read-from-loadbalancer | mysql-clienttrue logger.go:42: 13:05:23 | service-per-pod/5-read-from-loadbalancer | ++ kubectl -n kuttl-test-moving-shepherd exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h 34.132.221.154 -uroot -proot_password' logger.go:42: 13:05:23 | service-per-pod/5-read-from-loadbalancer | ++ sed -e 's/mysql: //' logger.go:42: 13:05:23 | service-per-pod/5-read-from-loadbalancer | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:05:25 | service-per-pod/5-read-from-loadbalancer | + data=100500 logger.go:42: 13:05:25 | service-per-pod/5-read-from-loadbalancer | + args=' --from-literal=service-per-pod-mysql-0.service-per-pod-mysql=100500 --from-literal=service-per-pod-mysql-1.service-per-pod-mysql=100500 --from-literal=service-per-pod-mysql-2.service-per-pod-mysql=100500' logger.go:42: 13:05:25 | service-per-pod/5-read-from-loadbalancer | + kubectl create configmap -n kuttl-test-moving-shepherd 04-read-from-loadbalancer --from-literal=service-per-pod-mysql-0.service-per-pod-mysql=100500 --from-literal=service-per-pod-mysql-1.service-per-pod-mysql=100500 --from-literal=service-per-pod-mysql-2.service-per-pod-mysql=100500 logger.go:42: 13:05:26 | service-per-pod/5-read-from-loadbalancer | configmap/04-read-from-loadbalancer created logger.go:42: 13:05:27 | service-per-pod/5-read-from-loadbalancer | test step completed 5-read-from-loadbalancer logger.go:42: 13:05:27 | service-per-pod/6-drop-finalizer | starting test step 6-drop-finalizer logger.go:42: 13:05:29 | service-per-pod/6-drop-finalizer | PerconaServerMySQL:kuttl-test-moving-shepherd/service-per-pod updated logger.go:42: 13:05:29 | service-per-pod/6-drop-finalizer | test step completed 6-drop-finalizer logger.go:42: 13:05:29 | service-per-pod | service-per-pod events from ns kuttl-test-moving-shepherd: logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:58:43 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-995qn Scheduled Successfully assigned kuttl-test-moving-shepherd/percona-server-mysql-operator-6dcd78d857-995qn to gke-jen-ps-339-3a8483b-1-default-pool-9abd8684-f1l9 logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:58:43 +0000 UTC Normal ReplicaSet.apps percona-server-mysql-operator-6dcd78d857 SuccessfulCreate Created pod: percona-server-mysql-operator-6dcd78d857-995qn logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:58:43 +0000 UTC Normal Deployment.apps percona-server-mysql-operator ScalingReplicaSet Scaled up replica set percona-server-mysql-operator-6dcd78d857 to 1 logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:58:45 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-995qn.spec.containers{manager} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:58:45 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-995qn.spec.containers{manager} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 420.620508ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:58:45 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-995qn.spec.containers{manager} Created Created container manager logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:58:45 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-995qn.spec.containers{manager} Started Started container manager logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:58:46 +0000 UTC Normal Lease.coordination.k8s.io 08db2feb.percona.com LeaderElection percona-server-mysql-operator-6dcd78d857-995qn_c83a28ed-f13a-444d-b910-8bf5be182b94 became leader logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:58:46 +0000 UTC Normal Pod mysql-client Scheduled Successfully assigned kuttl-test-moving-shepherd/mysql-client to gke-jen-ps-339-3a8483b-1-default-pool-9abd8684-p5sl logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:58:47 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Pulled Container image "percona/percona-server:8.0.25" already present on machine logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:58:47 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Created Created container mysql-client logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:58:47 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Started Started container mysql-client logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:03 +0000 UTC Normal CertificateRequest.cert-manager.io service-per-pod-ca-cert-5mchf WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:03 +0000 UTC Normal CertificateRequest.cert-manager.io service-per-pod-ca-cert-5mchf WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:03 +0000 UTC Normal CertificateRequest.cert-manager.io service-per-pod-ca-cert-5mchf WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:03 +0000 UTC Normal CertificateRequest.cert-manager.io service-per-pod-ca-cert-5mchf WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:03 +0000 UTC Normal CertificateRequest.cert-manager.io service-per-pod-ca-cert-5mchf WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:03 +0000 UTC Normal CertificateRequest.cert-manager.io service-per-pod-ca-cert-5mchf cert-manager.io Certificate request has been approved by cert-manager.io logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:03 +0000 UTC Normal CertificateRequest.cert-manager.io service-per-pod-ca-cert-5mchf CertificateIssued Certificate fetched from issuer successfully logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:03 +0000 UTC Normal Certificate.cert-manager.io service-per-pod-ca-cert Issuing Issuing certificate as Secret does not exist logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:03 +0000 UTC Normal Certificate.cert-manager.io service-per-pod-ca-cert Generated Stored new private key in temporary Secret resource "service-per-pod-ca-cert-njfg8" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:03 +0000 UTC Normal Certificate.cert-manager.io service-per-pod-ca-cert Requested Created new CertificateRequest resource "service-per-pod-ca-cert-5mchf" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:03 +0000 UTC Normal Certificate.cert-manager.io service-per-pod-ca-cert Issuing The certificate has been successfully issued logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:06 +0000 UTC Normal Issuer.cert-manager.io service-per-pod-pso-issuer KeyPairVerified Signing CA verified logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:06 +0000 UTC Normal CertificateRequest.cert-manager.io service-per-pod-ssl-zx5dz WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:06 +0000 UTC Normal CertificateRequest.cert-manager.io service-per-pod-ssl-zx5dz WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:06 +0000 UTC Normal CertificateRequest.cert-manager.io service-per-pod-ssl-zx5dz WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:06 +0000 UTC Normal CertificateRequest.cert-manager.io service-per-pod-ssl-zx5dz WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:06 +0000 UTC Normal CertificateRequest.cert-manager.io service-per-pod-ssl-zx5dz WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:06 +0000 UTC Normal CertificateRequest.cert-manager.io service-per-pod-ssl-zx5dz cert-manager.io Certificate request has been approved by cert-manager.io logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:06 +0000 UTC Normal CertificateRequest.cert-manager.io service-per-pod-ssl-zx5dz CertificateIssued Certificate fetched from issuer successfully logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:06 +0000 UTC Normal Certificate.cert-manager.io service-per-pod-ssl Issuing Issuing certificate as Secret was previously issued by Issuer.cert-manager.io/ logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:06 +0000 UTC Normal Certificate.cert-manager.io service-per-pod-ssl Reused Reusing private key stored in existing Secret resource "test-ssl" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:06 +0000 UTC Normal Certificate.cert-manager.io service-per-pod-ssl Requested Created new CertificateRequest resource "service-per-pod-ssl-zx5dz" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:06 +0000 UTC Normal Certificate.cert-manager.io service-per-pod-ssl Issuing The certificate has been successfully issued logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:09 +0000 UTC Normal PersistentVolumeClaim datadir-service-per-pod-mysql-0 WaitForFirstConsumer waiting for first consumer to be created before binding logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:09 +0000 UTC Normal StatefulSet.apps service-per-pod-mysql SuccessfulCreate create Claim datadir-service-per-pod-mysql-0 Pod service-per-pod-mysql-0 in StatefulSet service-per-pod-mysql success logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:09 +0000 UTC Normal StatefulSet.apps service-per-pod-mysql SuccessfulCreate create Pod service-per-pod-mysql-0 in StatefulSet service-per-pod-mysql successful logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:10 +0000 UTC Normal PersistentVolumeClaim datadir-service-per-pod-mysql-0 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:10 +0000 UTC Normal PersistentVolumeClaim datadir-service-per-pod-mysql-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-moving-shepherd/datadir-service-per-pod-mysql-0" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:10 +0000 UTC Normal Pod service-per-pod-orc-0 Scheduled Successfully assigned kuttl-test-moving-shepherd/service-per-pod-orc-0 to gke-jen-ps-339-3a8483b-1-default-pool-9abd8684-f1l9 logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:10 +0000 UTC Normal Pod service-per-pod-orc-0.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:10 +0000 UTC Normal StatefulSet.apps service-per-pod-orc SuccessfulCreate create Pod service-per-pod-orc-0 in StatefulSet service-per-pod-orc successful logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:11 +0000 UTC Normal Pod service-per-pod-orc-0.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 391.436534ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:11 +0000 UTC Normal Pod service-per-pod-orc-0.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:11 +0000 UTC Normal Pod service-per-pod-orc-0.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:13 +0000 UTC Normal PersistentVolumeClaim datadir-service-per-pod-mysql-0 ProvisioningSucceeded Successfully provisioned volume pvc-d02b5f74-4d4d-48ff-bf6f-6d231bcc95a2 logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:13 +0000 UTC Normal Pod service-per-pod-orc-0.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:13 +0000 UTC Normal Pod service-per-pod-orc-0.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 641.446043ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:13 +0000 UTC Normal Pod service-per-pod-orc-0.spec.containers{orc} Created Created container orc logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:13 +0000 UTC Normal Pod service-per-pod-orc-0.spec.containers{orc} Started Started container orc logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:13 +0000 UTC Normal Pod service-per-pod-orc-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:14 +0000 UTC Normal Pod service-per-pod-orc-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 399.07257ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:14 +0000 UTC Normal Pod service-per-pod-orc-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:14 +0000 UTC Normal Pod service-per-pod-orc-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:15 +0000 UTC Normal Pod service-per-pod-mysql-0 Scheduled Successfully assigned kuttl-test-moving-shepherd/service-per-pod-mysql-0 to gke-jen-ps-339-3a8483b-1-default-pool-9abd8684-hccm logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:22 +0000 UTC Normal Pod service-per-pod-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-d02b5f74-4d4d-48ff-bf6f-6d231bcc95a2" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:24 +0000 UTC Normal Pod service-per-pod-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:24 +0000 UTC Normal Pod service-per-pod-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 451.647896ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:24 +0000 UTC Normal Pod service-per-pod-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:24 +0000 UTC Normal Pod service-per-pod-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:26 +0000 UTC Normal Pod service-per-pod-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:27 +0000 UTC Normal Pod service-per-pod-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 412.648445ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:27 +0000 UTC Normal Pod service-per-pod-mysql-0.spec.containers{mysql} Created Created container mysql logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:27 +0000 UTC Normal Pod service-per-pod-mysql-0.spec.containers{mysql} Started Started container mysql logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:27 +0000 UTC Normal Pod service-per-pod-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:27 +0000 UTC Normal Pod service-per-pod-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 436.009131ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:27 +0000 UTC Normal Pod service-per-pod-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:28 +0000 UTC Normal Pod service-per-pod-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:28 +0000 UTC Normal Pod service-per-pod-mysql-0.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:28 +0000 UTC Normal Pod service-per-pod-mysql-0.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 372.434586ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:28 +0000 UTC Normal Pod service-per-pod-mysql-0.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:28 +0000 UTC Normal Pod service-per-pod-mysql-0.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:45 +0000 UTC Normal Pod service-per-pod-orc-1 Scheduled Successfully assigned kuttl-test-moving-shepherd/service-per-pod-orc-1 to gke-jen-ps-339-3a8483b-1-default-pool-9abd8684-p5sl logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:45 +0000 UTC Normal StatefulSet.apps service-per-pod-orc SuccessfulCreate create Pod service-per-pod-orc-1 in StatefulSet service-per-pod-orc successful logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:46 +0000 UTC Normal Pod service-per-pod-orc-1.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:46 +0000 UTC Normal Pod service-per-pod-orc-1.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 396.862692ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:46 +0000 UTC Normal Pod service-per-pod-orc-1.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:46 +0000 UTC Normal Pod service-per-pod-orc-1.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:49 +0000 UTC Normal Pod service-per-pod-orc-1.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:49 +0000 UTC Normal Pod service-per-pod-orc-1.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 372.100106ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:49 +0000 UTC Normal Pod service-per-pod-orc-1.spec.containers{orc} Created Created container orc logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:49 +0000 UTC Normal Pod service-per-pod-orc-1.spec.containers{orc} Started Started container orc logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:49 +0000 UTC Normal Pod service-per-pod-orc-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:50 +0000 UTC Normal Pod service-per-pod-orc-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 481.964241ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:50 +0000 UTC Normal Pod service-per-pod-orc-1.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:50 +0000 UTC Normal Pod service-per-pod-orc-1.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:59 +0000 UTC Normal PersistentVolumeClaim datadir-service-per-pod-mysql-1 WaitForFirstConsumer waiting for first consumer to be created before binding logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:59 +0000 UTC Normal PersistentVolumeClaim datadir-service-per-pod-mysql-1 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:59 +0000 UTC Normal PersistentVolumeClaim datadir-service-per-pod-mysql-1 Provisioning External provisioner is provisioning volume for claim "kuttl-test-moving-shepherd/datadir-service-per-pod-mysql-1" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:59 +0000 UTC Normal StatefulSet.apps service-per-pod-mysql SuccessfulCreate create Claim datadir-service-per-pod-mysql-1 Pod service-per-pod-mysql-1 in StatefulSet service-per-pod-mysql success logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 12:59:59 +0000 UTC Normal StatefulSet.apps service-per-pod-mysql SuccessfulCreate create Pod service-per-pod-mysql-1 in StatefulSet service-per-pod-mysql successful logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:03 +0000 UTC Normal PersistentVolumeClaim datadir-service-per-pod-mysql-1 ProvisioningSucceeded Successfully provisioned volume pvc-290dc8b5-cdc5-4b61-9da4-9899a4e9a00d logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:03 +0000 UTC Normal Pod service-per-pod-mysql-1 Scheduled Successfully assigned kuttl-test-moving-shepherd/service-per-pod-mysql-1 to gke-jen-ps-339-3a8483b-1-default-pool-9abd8684-f1l9 logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:04 +0000 UTC Normal Pod service-per-pod-haproxy-0 Scheduled Successfully assigned kuttl-test-moving-shepherd/service-per-pod-haproxy-0 to gke-jen-ps-339-3a8483b-1-default-pool-9abd8684-p5sl logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:04 +0000 UTC Normal StatefulSet.apps service-per-pod-haproxy SuccessfulCreate create Pod service-per-pod-haproxy-0 in StatefulSet service-per-pod-haproxy successful logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:05 +0000 UTC Normal Pod service-per-pod-haproxy-0.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:07 +0000 UTC Normal Pod service-per-pod-haproxy-0.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 2.137095183s logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:07 +0000 UTC Normal Pod service-per-pod-haproxy-0.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:08 +0000 UTC Normal Pod service-per-pod-haproxy-0.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:10 +0000 UTC Normal Pod service-per-pod-haproxy-0.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:10 +0000 UTC Normal Pod service-per-pod-mysql-1 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-290dc8b5-cdc5-4b61-9da4-9899a4e9a00d" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:11 +0000 UTC Normal Pod service-per-pod-haproxy-0.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 1.30286289s logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:11 +0000 UTC Normal Pod service-per-pod-haproxy-0.spec.containers{haproxy} Created Created container haproxy logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:11 +0000 UTC Normal Pod service-per-pod-haproxy-0.spec.containers{haproxy} Started Started container haproxy logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:11 +0000 UTC Normal Pod service-per-pod-haproxy-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:12 +0000 UTC Normal Pod service-per-pod-haproxy-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 501.569837ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:12 +0000 UTC Normal Pod service-per-pod-haproxy-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:12 +0000 UTC Normal Pod service-per-pod-haproxy-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:12 +0000 UTC Normal Pod service-per-pod-haproxy-1 Scheduled Successfully assigned kuttl-test-moving-shepherd/service-per-pod-haproxy-1 to gke-jen-ps-339-3a8483b-1-default-pool-9abd8684-hccm logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:12 +0000 UTC Normal StatefulSet.apps service-per-pod-haproxy SuccessfulCreate create Pod service-per-pod-haproxy-1 in StatefulSet service-per-pod-haproxy successful logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:13 +0000 UTC Normal Pod service-per-pod-haproxy-1.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:13 +0000 UTC Normal Pod service-per-pod-haproxy-1.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 453.929202ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:13 +0000 UTC Normal Pod service-per-pod-haproxy-1.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:13 +0000 UTC Normal Pod service-per-pod-haproxy-1.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:13 +0000 UTC Normal Pod service-per-pod-mysql-1.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:14 +0000 UTC Normal Pod service-per-pod-mysql-1.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 418.753486ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:14 +0000 UTC Normal Pod service-per-pod-mysql-1.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:14 +0000 UTC Normal Pod service-per-pod-mysql-1.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:15 +0000 UTC Normal Pod service-per-pod-mysql-1.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:16 +0000 UTC Normal Pod service-per-pod-haproxy-1.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:16 +0000 UTC Normal Pod service-per-pod-haproxy-1.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 502.392124ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:16 +0000 UTC Normal Pod service-per-pod-haproxy-1.spec.containers{haproxy} Created Created container haproxy logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:16 +0000 UTC Normal Pod service-per-pod-haproxy-1.spec.containers{haproxy} Started Started container haproxy logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:16 +0000 UTC Normal Pod service-per-pod-haproxy-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:16 +0000 UTC Normal Pod service-per-pod-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 432.444074ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:16 +0000 UTC Normal Pod service-per-pod-mysql-1.spec.containers{mysql} Created Created container mysql logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:16 +0000 UTC Normal Pod service-per-pod-mysql-1.spec.containers{mysql} Started Started container mysql logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:16 +0000 UTC Normal Pod service-per-pod-mysql-1.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:16 +0000 UTC Normal Pod service-per-pod-mysql-1.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 414.812973ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:16 +0000 UTC Normal Pod service-per-pod-mysql-1.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:16 +0000 UTC Normal Pod service-per-pod-mysql-1.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:16 +0000 UTC Normal Pod service-per-pod-mysql-1.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:17 +0000 UTC Normal Pod service-per-pod-haproxy-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 399.687941ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:17 +0000 UTC Normal Pod service-per-pod-haproxy-1.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:17 +0000 UTC Normal Pod service-per-pod-haproxy-1.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:17 +0000 UTC Normal Pod service-per-pod-mysql-1.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 426.214065ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:17 +0000 UTC Normal Pod service-per-pod-mysql-1.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:17 +0000 UTC Normal Pod service-per-pod-mysql-1.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:18 +0000 UTC Normal Pod service-per-pod-haproxy-2 Scheduled Successfully assigned kuttl-test-moving-shepherd/service-per-pod-haproxy-2 to gke-jen-ps-339-3a8483b-1-default-pool-9abd8684-f1l9 logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:18 +0000 UTC Normal StatefulSet.apps service-per-pod-haproxy SuccessfulCreate create Pod service-per-pod-haproxy-2 in StatefulSet service-per-pod-haproxy successful logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:19 +0000 UTC Normal Pod service-per-pod-haproxy-2.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:19 +0000 UTC Normal Pod service-per-pod-haproxy-2.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 417.933752ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:19 +0000 UTC Normal Pod service-per-pod-haproxy-2.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:19 +0000 UTC Normal Pod service-per-pod-haproxy-2.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:20 +0000 UTC Normal Pod service-per-pod-orc-2 Scheduled Successfully assigned kuttl-test-moving-shepherd/service-per-pod-orc-2 to gke-jen-ps-339-3a8483b-1-default-pool-9abd8684-hccm logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:20 +0000 UTC Normal StatefulSet.apps service-per-pod-orc SuccessfulCreate create Pod service-per-pod-orc-2 in StatefulSet service-per-pod-orc successful logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:21 +0000 UTC Normal Pod service-per-pod-haproxy-2.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:21 +0000 UTC Normal Pod service-per-pod-orc-2.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:22 +0000 UTC Normal Pod service-per-pod-haproxy-2.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 406.167054ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:22 +0000 UTC Normal Pod service-per-pod-haproxy-2.spec.containers{haproxy} Created Created container haproxy logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:22 +0000 UTC Normal Pod service-per-pod-haproxy-2.spec.containers{haproxy} Started Started container haproxy logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:22 +0000 UTC Normal Pod service-per-pod-haproxy-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:22 +0000 UTC Normal Pod service-per-pod-haproxy-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 500.00261ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:22 +0000 UTC Normal Pod service-per-pod-haproxy-2.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:22 +0000 UTC Normal Pod service-per-pod-orc-2.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 391.732672ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:22 +0000 UTC Normal Pod service-per-pod-orc-2.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:22 +0000 UTC Normal Pod service-per-pod-orc-2.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:23 +0000 UTC Normal Pod service-per-pod-haproxy-2.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:24 +0000 UTC Normal Pod service-per-pod-orc-2.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:24 +0000 UTC Normal Pod service-per-pod-orc-2.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 401.760617ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:24 +0000 UTC Normal Pod service-per-pod-orc-2.spec.containers{orc} Created Created container orc logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:24 +0000 UTC Normal Pod service-per-pod-orc-2.spec.containers{orc} Started Started container orc logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:24 +0000 UTC Normal Pod service-per-pod-orc-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:25 +0000 UTC Normal Pod service-per-pod-orc-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 385.134996ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:25 +0000 UTC Normal Pod service-per-pod-orc-2.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:25 +0000 UTC Normal Pod service-per-pod-orc-2.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:34 +0000 UTC Warning Pod service-per-pod-mysql-1.spec.containers{mysql} Unhealthy Startup probe failed: logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:34 +0000 UTC Normal Pod service-per-pod-mysql-1.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:00:38 +0000 UTC Normal Pod service-per-pod-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 579.410973ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:01:08 +0000 UTC Normal PersistentVolumeClaim datadir-service-per-pod-mysql-2 WaitForFirstConsumer waiting for first consumer to be created before binding logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:01:08 +0000 UTC Normal PersistentVolumeClaim datadir-service-per-pod-mysql-2 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:01:08 +0000 UTC Normal StatefulSet.apps service-per-pod-mysql SuccessfulCreate create Claim datadir-service-per-pod-mysql-2 Pod service-per-pod-mysql-2 in StatefulSet service-per-pod-mysql success logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:01:08 +0000 UTC Normal StatefulSet.apps service-per-pod-mysql SuccessfulCreate create Pod service-per-pod-mysql-2 in StatefulSet service-per-pod-mysql successful logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:01:09 +0000 UTC Normal PersistentVolumeClaim datadir-service-per-pod-mysql-2 Provisioning External provisioner is provisioning volume for claim "kuttl-test-moving-shepherd/datadir-service-per-pod-mysql-2" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:01:13 +0000 UTC Normal PersistentVolumeClaim datadir-service-per-pod-mysql-2 ProvisioningSucceeded Successfully provisioned volume pvc-61a07065-a93b-42ed-bbce-97ed6ac12d44 logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:01:13 +0000 UTC Normal Pod service-per-pod-mysql-2 Scheduled Successfully assigned kuttl-test-moving-shepherd/service-per-pod-mysql-2 to gke-jen-ps-339-3a8483b-1-default-pool-9abd8684-p5sl logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:01:21 +0000 UTC Normal Pod service-per-pod-mysql-2 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-61a07065-a93b-42ed-bbce-97ed6ac12d44" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:01:23 +0000 UTC Normal Pod service-per-pod-mysql-2.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:01:23 +0000 UTC Normal Pod service-per-pod-mysql-2.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 409.060489ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:01:23 +0000 UTC Normal Pod service-per-pod-mysql-2.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:01:23 +0000 UTC Normal Pod service-per-pod-mysql-2.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:01:25 +0000 UTC Normal Pod service-per-pod-mysql-2.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:01:25 +0000 UTC Normal Pod service-per-pod-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 395.234133ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:01:25 +0000 UTC Normal Pod service-per-pod-mysql-2.spec.containers{mysql} Created Created container mysql logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:01:25 +0000 UTC Normal Pod service-per-pod-mysql-2.spec.containers{mysql} Started Started container mysql logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:01:25 +0000 UTC Normal Pod service-per-pod-mysql-2.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:01:56 +0000 UTC Normal Pod service-per-pod-mysql-2.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 30.342940635s logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:01:56 +0000 UTC Normal Pod service-per-pod-mysql-2.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:01:56 +0000 UTC Normal Pod service-per-pod-mysql-2.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:01:56 +0000 UTC Normal Pod service-per-pod-mysql-2.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:01:56 +0000 UTC Normal Pod service-per-pod-mysql-2.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 383.040743ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:01:56 +0000 UTC Normal Pod service-per-pod-mysql-2.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:01:56 +0000 UTC Normal Pod service-per-pod-mysql-2.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:02:03 +0000 UTC Warning Pod service-per-pod-mysql-2.spec.containers{mysql} Unhealthy Startup probe failed: logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:02:03 +0000 UTC Normal Pod service-per-pod-mysql-2.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:02:07 +0000 UTC Normal Pod service-per-pod-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 368.749191ms logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:03:42 +0000 UTC Normal Service service-per-pod-mysql-0 Type ClusterIP -> LoadBalancer logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:03:42 +0000 UTC Normal Service service-per-pod-mysql-0 EnsuringLoadBalancer Ensuring load balancer logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:03:42 +0000 UTC Normal Service service-per-pod-mysql-1 Type ClusterIP -> LoadBalancer logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:03:42 +0000 UTC Normal Service service-per-pod-mysql-1 EnsuringLoadBalancer Ensuring load balancer logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:03:42 +0000 UTC Normal Service service-per-pod-mysql-2 Type ClusterIP -> LoadBalancer logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:03:42 +0000 UTC Normal Service service-per-pod-mysql-2 EnsuringLoadBalancer Ensuring load balancer logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:04:24 +0000 UTC Normal Service service-per-pod-mysql-1 EnsuredLoadBalancer Ensured load balancer logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:04:29 +0000 UTC Normal Service service-per-pod-mysql-2 EnsuredLoadBalancer Ensured load balancer logger.go:42: 13:05:29 | service-per-pod | 2023-08-09 13:04:33 +0000 UTC Normal Service service-per-pod-mysql-0 EnsuredLoadBalancer Ensured load balancer logger.go:42: 13:05:29 | service-per-pod | Deleting namespace: kuttl-test-moving-shepherd === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (465.33s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/service-per-pod (461.62s) PASS