=== RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. I0208 07:27:46.135032 12667 request.go:682] Waited for 1.013499722s due to client-side throttling, not priority and fairness, request: GET:https://35.225.58.111/apis/networking.gke.io/v1beta1?timeout=32s harness.go:275: Successful connection to cluster at: https://35.225.58.111 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 180 seconds for each step harness.go:372: testsuite: e2e-tests/tests has 30 tests === RUN kuttl/harness === RUN kuttl/harness/gr-self-healing === PAUSE kuttl/harness/gr-self-healing === CONT kuttl/harness/gr-self-healing logger.go:42: 07:27:52 | gr-self-healing | Creating namespace: kuttl-test-emerging-raptor logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | starting test step 0-deploy-operator logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | running command: [sh -c set -o errexit set -o xtrace source ../../functions init_temp_dir # do this only in the first TestStep deploy_operator deploy_non_tls_cluster_secrets deploy_tls_cluster_secrets deploy_client] logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | + source ../../functions logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ realpath ../../.. logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | ++++ pwd logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-self-healing logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | ++ test_name=gr-self-healing logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ export GIT_BRANCH=PR-523 logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ GIT_BRANCH=PR-523 logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ export VERSION=PR-523-f00253e logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ VERSION=PR-523-f00253e logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | ++++ which gdate logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | ++++ which date logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ date=/usr/bin/date logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ command -v oc logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ kubectl get nodes logger.go:42: 07:27:52 | gr-self-healing/0-deploy-operator | +++ grep '^minikube' logger.go:42: 07:27:53 | gr-self-healing/0-deploy-operator | E0208 07:27:53.792828 12784 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:27:54 | gr-self-healing/0-deploy-operator | E0208 07:27:54.004015 12784 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:27:54 | gr-self-healing/0-deploy-operator | E0208 07:27:54.111951 12784 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:27:54 | gr-self-healing/0-deploy-operator | E0208 07:27:54.218267 12784 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:27:54 | gr-self-healing/0-deploy-operator | + init_temp_dir logger.go:42: 07:27:54 | gr-self-healing/0-deploy-operator | + rm -rf /tmp/kuttl/ps/gr-self-healing logger.go:42: 07:27:54 | gr-self-healing/0-deploy-operator | + mkdir -p /tmp/kuttl/ps/gr-self-healing logger.go:42: 07:27:54 | gr-self-healing/0-deploy-operator | + deploy_operator logger.go:42: 07:27:54 | gr-self-healing/0-deploy-operator | + kubectl -n kuttl-test-emerging-raptor apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy/crd.yaml logger.go:42: 07:27:55 | gr-self-healing/0-deploy-operator | E0208 07:27:55.457244 12824 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:27:55 | gr-self-healing/0-deploy-operator | E0208 07:27:55.564128 12824 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:27:55 | gr-self-healing/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlbackups.ps.percona.com serverside-applied logger.go:42: 07:27:55 | gr-self-healing/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlrestores.ps.percona.com serverside-applied logger.go:42: 07:27:56 | gr-self-healing/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqls.ps.percona.com serverside-applied logger.go:42: 07:27:56 | gr-self-healing/0-deploy-operator | + kubectl -n kuttl-test-emerging-raptor apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy/rbac.yaml logger.go:42: 07:27:58 | gr-self-healing/0-deploy-operator | E0208 07:27:58.745536 12866 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:27:58 | gr-self-healing/0-deploy-operator | E0208 07:27:58.946493 12866 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:27:59 | gr-self-healing/0-deploy-operator | serviceaccount/percona-server-mysql-operator created logger.go:42: 07:27:59 | gr-self-healing/0-deploy-operator | serviceaccount/percona-server-mysql-operator-orchestrator created logger.go:42: 07:28:00 | gr-self-healing/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 07:28:00 | gr-self-healing/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 07:28:00 | gr-self-healing/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 07:28:01 | gr-self-healing/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 07:28:01 | gr-self-healing/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 07:28:01 | gr-self-healing/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 07:28:01 | gr-self-healing/0-deploy-operator | + yq eval '(select(documentIndex==1).spec.template.spec.containers[] | select(.name=="manager").env[] | select(.name=="DISABLE_TELEMETRY").value) = "true"' logger.go:42: 07:28:01 | gr-self-healing/0-deploy-operator | + yq eval '(select(documentIndex==1).spec.template.spec.containers[] | select(.name=="manager").env[] | select(.name=="LOG_LEVEL").value) = "DEBUG"' logger.go:42: 07:28:01 | gr-self-healing/0-deploy-operator | ++ printf 'select(documentIndex==1).spec.template.spec.containers[0].image="%s"' perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:28:01 | gr-self-healing/0-deploy-operator | + kubectl -n kuttl-test-emerging-raptor apply -f - logger.go:42: 07:28:01 | gr-self-healing/0-deploy-operator | + yq eval 'select(documentIndex==1).spec.template.spec.containers[0].image="perconalab/percona-server-mysql-operator:PR-523-f00253e"' /mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy/operator.yaml logger.go:42: 07:28:02 | gr-self-healing/0-deploy-operator | E0208 07:28:02.586228 12935 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:02 | gr-self-healing/0-deploy-operator | E0208 07:28:02.799995 12935 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:03 | gr-self-healing/0-deploy-operator | configmap/percona-server-mysql-operator-config created logger.go:42: 07:28:04 | gr-self-healing/0-deploy-operator | deployment.apps/percona-server-mysql-operator created logger.go:42: 07:28:04 | gr-self-healing/0-deploy-operator | + deploy_non_tls_cluster_secrets logger.go:42: 07:28:04 | gr-self-healing/0-deploy-operator | + kubectl -n kuttl-test-emerging-raptor apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf/secrets.yaml logger.go:42: 07:28:05 | gr-self-healing/0-deploy-operator | E0208 07:28:05.002958 12985 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:05 | gr-self-healing/0-deploy-operator | E0208 07:28:05.224924 12985 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:05 | gr-self-healing/0-deploy-operator | secret/test-secrets created logger.go:42: 07:28:05 | gr-self-healing/0-deploy-operator | + deploy_tls_cluster_secrets logger.go:42: 07:28:05 | gr-self-healing/0-deploy-operator | + kubectl -n kuttl-test-emerging-raptor apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf/ssl-secret.yaml logger.go:42: 07:28:06 | gr-self-healing/0-deploy-operator | E0208 07:28:06.627656 13008 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:06 | gr-self-healing/0-deploy-operator | E0208 07:28:06.816780 13008 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:07 | gr-self-healing/0-deploy-operator | secret/test-ssl created logger.go:42: 07:28:07 | gr-self-healing/0-deploy-operator | + deploy_client logger.go:42: 07:28:07 | gr-self-healing/0-deploy-operator | + kubectl -n kuttl-test-emerging-raptor apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf/client.yaml logger.go:42: 07:28:08 | gr-self-healing/0-deploy-operator | E0208 07:28:08.093431 13038 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:08 | gr-self-healing/0-deploy-operator | E0208 07:28:08.310113 13038 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:08 | gr-self-healing/0-deploy-operator | pod/mysql-client created I0208 07:28:10.667639 12667 request.go:682] Waited for 1.0303395s due to client-side throttling, not priority and fairness, request: GET:https://35.225.58.111/apis/discovery.k8s.io/v1?timeout=32s logger.go:42: 07:28:11 | gr-self-healing/0-deploy-operator | running command: [sh -c kubectl assert exist-enhanced deployment percona-server-mysql-operator -n ${OPERATOR_NS:-$NAMESPACE} --field-selector status.readyReplicas=1] logger.go:42: 07:28:11 | gr-self-healing/0-deploy-operator | ASSERT deployment percona-server-mysql-operator matching field criteria 'status.readyReplicas=1' should exist. logger.go:42: 07:28:12 | gr-self-healing/0-deploy-operator | E0208 07:28:12.569180 13109 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:12 | gr-self-healing/0-deploy-operator | E0208 07:28:12.692205 13109 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:12 | gr-self-healing/0-deploy-operator | E0208 07:28:12.797483 13109 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:12 | gr-self-healing/0-deploy-operator | E0208 07:28:12.902774 13109 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:13 | gr-self-healing/0-deploy-operator | ASSERT FAIL Resource(s) not found. logger.go:42: 07:28:14 | gr-self-healing/0-deploy-operator | running command: [sh -c kubectl assert exist-enhanced deployment percona-server-mysql-operator -n ${OPERATOR_NS:-$NAMESPACE} --field-selector status.readyReplicas=1] logger.go:42: 07:28:14 | gr-self-healing/0-deploy-operator | ASSERT deployment percona-server-mysql-operator matching field criteria 'status.readyReplicas=1' should exist. logger.go:42: 07:28:15 | gr-self-healing/0-deploy-operator | E0208 07:28:15.055931 13152 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:15 | gr-self-healing/0-deploy-operator | E0208 07:28:15.161817 13152 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:15 | gr-self-healing/0-deploy-operator | E0208 07:28:15.266473 13152 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:15 | gr-self-healing/0-deploy-operator | E0208 07:28:15.371540 13152 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:15 | gr-self-healing/0-deploy-operator | ASSERT FAIL Resource(s) not found. logger.go:42: 07:28:16 | gr-self-healing/0-deploy-operator | running command: [sh -c kubectl assert exist-enhanced deployment percona-server-mysql-operator -n ${OPERATOR_NS:-$NAMESPACE} --field-selector status.readyReplicas=1] logger.go:42: 07:28:16 | gr-self-healing/0-deploy-operator | ASSERT deployment percona-server-mysql-operator matching field criteria 'status.readyReplicas=1' should exist. logger.go:42: 07:28:17 | gr-self-healing/0-deploy-operator | E0208 07:28:17.383027 13202 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:17 | gr-self-healing/0-deploy-operator | E0208 07:28:17.612511 13202 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:17 | gr-self-healing/0-deploy-operator | E0208 07:28:17.719035 13202 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:17 | gr-self-healing/0-deploy-operator | E0208 07:28:17.825782 13202 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:17 | gr-self-healing/0-deploy-operator | ASSERT FAIL Resource(s) not found. logger.go:42: 07:28:19 | gr-self-healing/0-deploy-operator | running command: [sh -c kubectl assert exist-enhanced deployment percona-server-mysql-operator -n ${OPERATOR_NS:-$NAMESPACE} --field-selector status.readyReplicas=1] logger.go:42: 07:28:19 | gr-self-healing/0-deploy-operator | ASSERT deployment percona-server-mysql-operator matching field criteria 'status.readyReplicas=1' should exist. logger.go:42: 07:28:19 | gr-self-healing/0-deploy-operator | E0208 07:28:19.828626 13245 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:20 | gr-self-healing/0-deploy-operator | E0208 07:28:20.039627 13245 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:20 | gr-self-healing/0-deploy-operator | E0208 07:28:20.145704 13245 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:20 | gr-self-healing/0-deploy-operator | E0208 07:28:20.251587 13245 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:20 | gr-self-healing/0-deploy-operator | ASSERT FAIL Resource(s) not found. logger.go:42: 07:28:21 | gr-self-healing/0-deploy-operator | running command: [sh -c kubectl assert exist-enhanced deployment percona-server-mysql-operator -n ${OPERATOR_NS:-$NAMESPACE} --field-selector status.readyReplicas=1] logger.go:42: 07:28:21 | gr-self-healing/0-deploy-operator | ASSERT deployment percona-server-mysql-operator matching field criteria 'status.readyReplicas=1' should exist. logger.go:42: 07:28:22 | gr-self-healing/0-deploy-operator | E0208 07:28:22.649455 13297 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:22 | gr-self-healing/0-deploy-operator | E0208 07:28:22.762638 13297 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:22 | gr-self-healing/0-deploy-operator | E0208 07:28:22.870272 13297 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:22 | gr-self-healing/0-deploy-operator | E0208 07:28:22.977768 13297 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:23 | gr-self-healing/0-deploy-operator | ASSERT FAIL Resource(s) not found. logger.go:42: 07:28:24 | gr-self-healing/0-deploy-operator | running command: [sh -c kubectl assert exist-enhanced deployment percona-server-mysql-operator -n ${OPERATOR_NS:-$NAMESPACE} --field-selector status.readyReplicas=1] logger.go:42: 07:28:24 | gr-self-healing/0-deploy-operator | ASSERT deployment percona-server-mysql-operator matching field criteria 'status.readyReplicas=1' should exist. logger.go:42: 07:28:25 | gr-self-healing/0-deploy-operator | E0208 07:28:25.037733 13340 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:25 | gr-self-healing/0-deploy-operator | E0208 07:28:25.253560 13340 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:25 | gr-self-healing/0-deploy-operator | E0208 07:28:25.360586 13340 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:25 | gr-self-healing/0-deploy-operator | E0208 07:28:25.467598 13340 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:25 | gr-self-healing/0-deploy-operator | ASSERT FAIL Resource(s) not found. logger.go:42: 07:28:26 | gr-self-healing/0-deploy-operator | running command: [sh -c kubectl assert exist-enhanced deployment percona-server-mysql-operator -n ${OPERATOR_NS:-$NAMESPACE} --field-selector status.readyReplicas=1] logger.go:42: 07:28:26 | gr-self-healing/0-deploy-operator | ASSERT deployment percona-server-mysql-operator matching field criteria 'status.readyReplicas=1' should exist. logger.go:42: 07:28:27 | gr-self-healing/0-deploy-operator | E0208 07:28:27.568560 13392 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:27 | gr-self-healing/0-deploy-operator | E0208 07:28:27.683368 13392 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:27 | gr-self-healing/0-deploy-operator | E0208 07:28:27.789826 13392 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:27 | gr-self-healing/0-deploy-operator | E0208 07:28:27.896273 13392 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:28 | gr-self-healing/0-deploy-operator | ASSERT FAIL Resource(s) not found. logger.go:42: 07:28:29 | gr-self-healing/0-deploy-operator | running command: [sh -c kubectl assert exist-enhanced deployment percona-server-mysql-operator -n ${OPERATOR_NS:-$NAMESPACE} --field-selector status.readyReplicas=1] logger.go:42: 07:28:29 | gr-self-healing/0-deploy-operator | ASSERT deployment percona-server-mysql-operator matching field criteria 'status.readyReplicas=1' should exist. logger.go:42: 07:28:29 | gr-self-healing/0-deploy-operator | E0208 07:28:29.993305 13435 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:30 | gr-self-healing/0-deploy-operator | E0208 07:28:30.108736 13435 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:30 | gr-self-healing/0-deploy-operator | E0208 07:28:30.214746 13435 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:30 | gr-self-healing/0-deploy-operator | E0208 07:28:30.320837 13435 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:30 | gr-self-healing/0-deploy-operator | ASSERT FAIL Resource(s) not found. logger.go:42: 07:28:31 | gr-self-healing/0-deploy-operator | running command: [sh -c kubectl assert exist-enhanced deployment percona-server-mysql-operator -n ${OPERATOR_NS:-$NAMESPACE} --field-selector status.readyReplicas=1] logger.go:42: 07:28:31 | gr-self-healing/0-deploy-operator | ASSERT deployment percona-server-mysql-operator matching field criteria 'status.readyReplicas=1' should exist. logger.go:42: 07:28:32 | gr-self-healing/0-deploy-operator | E0208 07:28:32.309764 13477 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:32 | gr-self-healing/0-deploy-operator | E0208 07:28:32.521554 13477 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:32 | gr-self-healing/0-deploy-operator | E0208 07:28:32.627171 13477 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:32 | gr-self-healing/0-deploy-operator | E0208 07:28:32.733232 13477 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request logger.go:42: 07:28:32 | gr-self-healing/0-deploy-operator | ASSERT FAIL Resource(s) not found. logger.go:42: 07:28:34 | gr-self-healing/0-deploy-operator | running command: [sh -c kubectl assert exist-enhanced deployment percona-server-mysql-operator -n ${OPERATOR_NS:-$NAMESPACE} --field-selector status.readyReplicas=1] logger.go:42: 07:28:34 | gr-self-healing/0-deploy-operator | ASSERT deployment percona-server-mysql-operator matching field criteria 'status.readyReplicas=1' should exist. logger.go:42: 07:28:35 | gr-self-healing/0-deploy-operator | INFO Found 1 resource(s). logger.go:42: 07:28:35 | gr-self-healing/0-deploy-operator | NAME NAMESPACE COL0 logger.go:42: 07:28:35 | gr-self-healing/0-deploy-operator | percona-server-mysql-operator kuttl-test-emerging-raptor 1 logger.go:42: 07:28:35 | gr-self-healing/0-deploy-operator | ASSERT PASS logger.go:42: 07:28:35 | gr-self-healing/0-deploy-operator | test step completed 0-deploy-operator logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | starting test step 1-deploy-chaos-mesh logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | running command: [sh -c set -o errexit set -o xtrace source ../../functions deploy_chaos_mesh] logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | + source ../../functions logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ realpath ../../.. logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | ++++ pwd logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-self-healing logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | ++ test_name=gr-self-healing logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ export GIT_BRANCH=PR-523 logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ GIT_BRANCH=PR-523 logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ export VERSION=PR-523-f00253e logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ VERSION=PR-523-f00253e logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | ++++ which gdate logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | ++++ which date logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ date=/usr/bin/date logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ command -v oc logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ kubectl get nodes logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | +++ grep '^minikube' logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | + deploy_chaos_mesh logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | + destroy_chaos_mesh logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | ++ helm list --all-namespaces --filter chaos-mesh logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | ++ tail -n1 logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | ++ sed s/NAMESPACE// logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | ++ awk '-F ' '{print $2}' logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 07:28:35 | gr-self-healing/1-deploy-chaos-mesh | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 07:28:36 | gr-self-healing/1-deploy-chaos-mesh | + local chaos_mesh_ns= logger.go:42: 07:28:36 | gr-self-healing/1-deploy-chaos-mesh | ++ kubectl api-resources logger.go:42: 07:28:36 | gr-self-healing/1-deploy-chaos-mesh | ++ grep chaos-mesh logger.go:42: 07:28:36 | gr-self-healing/1-deploy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 07:28:36 | gr-self-healing/1-deploy-chaos-mesh | + '[' -n '' ']' logger.go:42: 07:28:36 | gr-self-healing/1-deploy-chaos-mesh | ++ kubectl get crd logger.go:42: 07:28:36 | gr-self-healing/1-deploy-chaos-mesh | ++ grep chaos-mesh.org logger.go:42: 07:28:36 | gr-self-healing/1-deploy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 07:28:37 | gr-self-healing/1-deploy-chaos-mesh | + timeout 30 kubectl delete crd logger.go:42: 07:28:37 | gr-self-healing/1-deploy-chaos-mesh | error: resource(s) were provided, but no name was specified logger.go:42: 07:28:37 | gr-self-healing/1-deploy-chaos-mesh | + : logger.go:42: 07:28:37 | gr-self-healing/1-deploy-chaos-mesh | ++ kubectl get clusterrolebinding logger.go:42: 07:28:37 | gr-self-healing/1-deploy-chaos-mesh | ++ grep chaos-mesh logger.go:42: 07:28:37 | gr-self-healing/1-deploy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 07:28:38 | gr-self-healing/1-deploy-chaos-mesh | + timeout 30 kubectl delete clusterrolebinding logger.go:42: 07:28:38 | gr-self-healing/1-deploy-chaos-mesh | error: resource(s) were provided, but no name was specified logger.go:42: 07:28:38 | gr-self-healing/1-deploy-chaos-mesh | + : logger.go:42: 07:28:38 | gr-self-healing/1-deploy-chaos-mesh | ++ kubectl get clusterrole logger.go:42: 07:28:38 | gr-self-healing/1-deploy-chaos-mesh | ++ grep chaos-mesh logger.go:42: 07:28:38 | gr-self-healing/1-deploy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 07:28:38 | gr-self-healing/1-deploy-chaos-mesh | + timeout 30 kubectl delete clusterrole logger.go:42: 07:28:38 | gr-self-healing/1-deploy-chaos-mesh | error: resource(s) were provided, but no name was specified logger.go:42: 07:28:38 | gr-self-healing/1-deploy-chaos-mesh | + : logger.go:42: 07:28:38 | gr-self-healing/1-deploy-chaos-mesh | ++ kubectl get MutatingWebhookConfiguration logger.go:42: 07:28:38 | gr-self-healing/1-deploy-chaos-mesh | ++ grep chaos-mesh logger.go:42: 07:28:38 | gr-self-healing/1-deploy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 07:28:39 | gr-self-healing/1-deploy-chaos-mesh | + timeout 30 kubectl delete MutatingWebhookConfiguration logger.go:42: 07:28:39 | gr-self-healing/1-deploy-chaos-mesh | error: resource(s) were provided, but no name was specified logger.go:42: 07:28:39 | gr-self-healing/1-deploy-chaos-mesh | + : logger.go:42: 07:28:39 | gr-self-healing/1-deploy-chaos-mesh | ++ kubectl get ValidatingWebhookConfiguration logger.go:42: 07:28:39 | gr-self-healing/1-deploy-chaos-mesh | ++ grep chaos-mesh logger.go:42: 07:28:39 | gr-self-healing/1-deploy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 07:28:39 | gr-self-healing/1-deploy-chaos-mesh | + timeout 30 kubectl delete ValidatingWebhookConfiguration logger.go:42: 07:28:39 | gr-self-healing/1-deploy-chaos-mesh | error: resource(s) were provided, but no name was specified logger.go:42: 07:28:39 | gr-self-healing/1-deploy-chaos-mesh | + : logger.go:42: 07:28:39 | gr-self-healing/1-deploy-chaos-mesh | ++ kubectl get ValidatingWebhookConfiguration logger.go:42: 07:28:39 | gr-self-healing/1-deploy-chaos-mesh | ++ grep validate-auth logger.go:42: 07:28:39 | gr-self-healing/1-deploy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 07:28:39 | gr-self-healing/1-deploy-chaos-mesh | + timeout 30 kubectl delete ValidatingWebhookConfiguration logger.go:42: 07:28:40 | gr-self-healing/1-deploy-chaos-mesh | error: resource(s) were provided, but no name was specified logger.go:42: 07:28:40 | gr-self-healing/1-deploy-chaos-mesh | + : logger.go:42: 07:28:40 | gr-self-healing/1-deploy-chaos-mesh | + helm repo add chaos-mesh https://charts.chaos-mesh.org logger.go:42: 07:28:40 | gr-self-healing/1-deploy-chaos-mesh | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 07:28:40 | gr-self-healing/1-deploy-chaos-mesh | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 07:28:40 | gr-self-healing/1-deploy-chaos-mesh | "chaos-mesh" has been added to your repositories logger.go:42: 07:28:40 | gr-self-healing/1-deploy-chaos-mesh | + '[' -n '' ']' logger.go:42: 07:28:40 | gr-self-healing/1-deploy-chaos-mesh | + helm install chaos-mesh chaos-mesh/chaos-mesh --namespace=kuttl-test-emerging-raptor --set chaosDaemon.runtime=containerd --set chaosDaemon.socketPath=/run/containerd/containerd.sock --set dashboard.create=false --version 2.5.1 logger.go:42: 07:28:40 | gr-self-healing/1-deploy-chaos-mesh | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 07:28:40 | gr-self-healing/1-deploy-chaos-mesh | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 07:29:09 | gr-self-healing/1-deploy-chaos-mesh | NAME: chaos-mesh logger.go:42: 07:29:09 | gr-self-healing/1-deploy-chaos-mesh | LAST DEPLOYED: Thu Feb 8 07:28:56 2024 logger.go:42: 07:29:09 | gr-self-healing/1-deploy-chaos-mesh | NAMESPACE: kuttl-test-emerging-raptor logger.go:42: 07:29:09 | gr-self-healing/1-deploy-chaos-mesh | STATUS: deployed logger.go:42: 07:29:09 | gr-self-healing/1-deploy-chaos-mesh | REVISION: 1 logger.go:42: 07:29:09 | gr-self-healing/1-deploy-chaos-mesh | TEST SUITE: None logger.go:42: 07:29:09 | gr-self-healing/1-deploy-chaos-mesh | NOTES: logger.go:42: 07:29:09 | gr-self-healing/1-deploy-chaos-mesh | 1. Make sure chaos-mesh components are running logger.go:42: 07:29:09 | gr-self-healing/1-deploy-chaos-mesh | kubectl get pods --namespace kuttl-test-emerging-raptor -l app.kubernetes.io/instance=chaos-mesh logger.go:42: 07:29:09 | gr-self-healing/1-deploy-chaos-mesh | + sleep 10 logger.go:42: 07:29:27 | gr-self-healing/1-deploy-chaos-mesh | test step completed 1-deploy-chaos-mesh logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | starting test step 2-create-cluster logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | running command: [sh -c set -o errexit set -o xtrace source ../../functions get_cr \ | yq eval '.spec.mysql.clusterType="group-replication"' - \ | yq eval '.spec.mysql.size=3' - \ | yq eval '.spec.mysql.affinity.antiAffinityTopologyKey="none"' - \ | yq eval '.spec.proxy.haproxy.enabled=false' - \ | yq eval '.spec.proxy.router.enabled=true' - \ | yq eval '.spec.proxy.router.size=3' - \ | yq eval '.spec.proxy.router.affinity.antiAffinityTopologyKey="none"' - \ | yq eval '.spec.orchestrator.enabled=false' - \ | kubectl -n "${NAMESPACE}" apply -f -] logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + source ../../functions logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ realpath ../../.. logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | ++++ pwd logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-self-healing logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | ++ test_name=gr-self-healing logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ export GIT_BRANCH=PR-523 logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ GIT_BRANCH=PR-523 logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ export VERSION=PR-523-f00253e logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ VERSION=PR-523-f00253e logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | ++++ which gdate logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | ++++ which date logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ date=/usr/bin/date logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ command -v oc logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ kubectl get nodes logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | +++ grep '^minikube' logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + yq eval '.spec.mysql.clusterType="group-replication"' - logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + yq eval .spec.mysql.size=3 - logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + yq eval .spec.proxy.haproxy.enabled=false - logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + get_cr logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + local name_suffix= logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | ++ printf '.spec.proxy.haproxy.image="%s"' perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + yq eval .spec.proxy.router.enabled=true - logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + yq eval '.spec.mysql.affinity.antiAffinityTopologyKey="none"' - logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + yq eval .spec.proxy.router.size=3 - logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + yq eval .spec.orchestrator.enabled=false - logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + yq eval '.spec.proxy.router.affinity.antiAffinityTopologyKey="none"' - logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + yq eval '.spec.secretsName="test-secrets"' - logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + kubectl -n kuttl-test-emerging-raptor apply -f - logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + yq eval '.spec.sslSecretName="test-ssl"' - logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + yq eval '.spec.upgradeOptions.apply="disabled"' - logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + '[' -n '' ']' logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + yq eval - logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | ++ printf '.spec.toolkit.image="%s"' perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | ++ printf '.metadata.name="%s"' gr-self-healing logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | ++ printf '.spec.orchestrator.image="%s"' perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | ++ printf '.spec.pmm.image="%s"' perconalab/pmm-client:dev-latest logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | ++ printf '.spec.proxy.router.image="%s"' perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + yq eval '.metadata.name="gr-self-healing"' /mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy/cr.yaml logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | ++ printf '.spec.initImage="%s"' perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + yq eval '.spec.initImage="perconalab/percona-server-mysql-operator:PR-523-f00253e"' - logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + yq eval '.spec.proxy.router.image="perconalab/percona-server-mysql-operator:main-router"' - logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | ++ printf '.spec.mysql.image="%s"' perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + yq eval '.spec.pmm.image="perconalab/pmm-client:dev-latest"' - logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + yq eval '.spec.mysql.image="perconalab/percona-server-mysql-operator:main-psmysql"' - logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + yq eval '.spec.proxy.haproxy.image="perconalab/percona-server-mysql-operator:main-haproxy"' - logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + yq eval '.spec.orchestrator.image="perconalab/percona-server-mysql-operator:main-orchestrator"' - logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | ++ printf '.spec.backup.image="%s"' perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + yq eval '.spec.backup.image="perconalab/percona-server-mysql-operator:main-backup"' - logger.go:42: 07:29:27 | gr-self-healing/2-create-cluster | + yq eval '.spec.toolkit.image="perconalab/percona-server-mysql-operator:main-toolkit"' - logger.go:42: 07:29:29 | gr-self-healing/2-create-cluster | perconaservermysql.ps.percona.com/gr-self-healing created logger.go:42: 07:35:43 | gr-self-healing/2-create-cluster | test step completed 2-create-cluster logger.go:42: 07:35:43 | gr-self-healing/3-write-data | starting test step 3-write-data logger.go:42: 07:35:43 | gr-self-healing/3-write-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" \ "-h $(get_mysql_router_service $(get_cluster_name)) -P 6446 -uroot -proot_password" run_mysql \ "INSERT myDB.myTable (id) VALUES (100500)" \ "-h $(get_mysql_router_service $(get_cluster_name)) -P 6446 -uroot -proot_password"] logger.go:42: 07:35:43 | gr-self-healing/3-write-data | + source ../../functions logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ realpath ../../.. logger.go:42: 07:35:43 | gr-self-healing/3-write-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:35:43 | gr-self-healing/3-write-data | ++++ pwd logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-self-healing logger.go:42: 07:35:43 | gr-self-healing/3-write-data | ++ test_name=gr-self-healing logger.go:42: 07:35:43 | gr-self-healing/3-write-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:35:43 | gr-self-healing/3-write-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ export GIT_BRANCH=PR-523 logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ GIT_BRANCH=PR-523 logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ export VERSION=PR-523-f00253e logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ VERSION=PR-523-f00253e logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:35:43 | gr-self-healing/3-write-data | ++++ which gdate logger.go:42: 07:35:43 | gr-self-healing/3-write-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 07:35:43 | gr-self-healing/3-write-data | ++++ which date logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ date=/usr/bin/date logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ command -v oc logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ kubectl get nodes logger.go:42: 07:35:43 | gr-self-healing/3-write-data | +++ grep '^minikube' logger.go:42: 07:35:44 | gr-self-healing/3-write-data | +++ get_cluster_name logger.go:42: 07:35:44 | gr-self-healing/3-write-data | +++ kubectl -n kuttl-test-emerging-raptor get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 07:35:44 | gr-self-healing/3-write-data | ++ get_mysql_router_service gr-self-healing logger.go:42: 07:35:44 | gr-self-healing/3-write-data | ++ local cluster=gr-self-healing logger.go:42: 07:35:44 | gr-self-healing/3-write-data | ++ echo gr-self-healing-router logger.go:42: 07:35:44 | gr-self-healing/3-write-data | + run_mysql 'CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' '-h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:35:44 | gr-self-healing/3-write-data | + local 'command=CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' logger.go:42: 07:35:44 | gr-self-healing/3-write-data | + local 'uri=-h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:35:44 | gr-self-healing/3-write-data | + local pod= logger.go:42: 07:35:44 | gr-self-healing/3-write-data | ++ get_client_pod logger.go:42: 07:35:44 | gr-self-healing/3-write-data | ++ kubectl -n kuttl-test-emerging-raptor get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 07:35:45 | gr-self-healing/3-write-data | + client_pod=mysql-client logger.go:42: 07:35:45 | gr-self-healing/3-write-data | + wait_pod mysql-client logger.go:42: 07:35:45 | gr-self-healing/3-write-data | + local pod=mysql-client logger.go:42: 07:35:45 | gr-self-healing/3-write-data | + set +o xtrace logger.go:42: 07:35:45 | gr-self-healing/3-write-data | mysql-clienttrue logger.go:42: 07:35:45 | gr-self-healing/3-write-data | + kubectl -n kuttl-test-emerging-raptor exec mysql-client -- bash -c 'printf '\''%s\n'\'' "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" | mysql -sN -h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:35:45 | gr-self-healing/3-write-data | + sed -e 's/mysql: //' logger.go:42: 07:35:45 | gr-self-healing/3-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 07:35:47 | gr-self-healing/3-write-data | + : logger.go:42: 07:35:47 | gr-self-healing/3-write-data | +++ get_cluster_name logger.go:42: 07:35:47 | gr-self-healing/3-write-data | +++ kubectl -n kuttl-test-emerging-raptor get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 07:35:47 | gr-self-healing/3-write-data | ++ get_mysql_router_service gr-self-healing logger.go:42: 07:35:47 | gr-self-healing/3-write-data | ++ local cluster=gr-self-healing logger.go:42: 07:35:47 | gr-self-healing/3-write-data | ++ echo gr-self-healing-router logger.go:42: 07:35:47 | gr-self-healing/3-write-data | + run_mysql 'INSERT myDB.myTable (id) VALUES (100500)' '-h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:35:47 | gr-self-healing/3-write-data | + local 'command=INSERT myDB.myTable (id) VALUES (100500)' logger.go:42: 07:35:47 | gr-self-healing/3-write-data | + local 'uri=-h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:35:47 | gr-self-healing/3-write-data | + local pod= logger.go:42: 07:35:47 | gr-self-healing/3-write-data | ++ get_client_pod logger.go:42: 07:35:47 | gr-self-healing/3-write-data | ++ kubectl -n kuttl-test-emerging-raptor get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 07:35:48 | gr-self-healing/3-write-data | + client_pod=mysql-client logger.go:42: 07:35:48 | gr-self-healing/3-write-data | + wait_pod mysql-client logger.go:42: 07:35:48 | gr-self-healing/3-write-data | + local pod=mysql-client logger.go:42: 07:35:48 | gr-self-healing/3-write-data | + set +o xtrace logger.go:42: 07:35:48 | gr-self-healing/3-write-data | mysql-clienttrue logger.go:42: 07:35:48 | gr-self-healing/3-write-data | + kubectl -n kuttl-test-emerging-raptor exec mysql-client -- bash -c 'printf '\''%s\n'\'' "INSERT myDB.myTable (id) VALUES (100500)" | mysql -sN -h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:35:48 | gr-self-healing/3-write-data | + sed -e 's/mysql: //' logger.go:42: 07:35:48 | gr-self-healing/3-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 07:35:49 | gr-self-healing/3-write-data | + : logger.go:42: 07:35:50 | gr-self-healing/3-write-data | test step completed 3-write-data logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | starting test step 4-read-from-primary logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | running command: [sh -c set -o errexit set -o xtrace source ../../functions data=$(run_mysql "SELECT * FROM myDB.myTable" "-h $(get_mysql_router_service $(get_cluster_name)) -P 6446 -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 04-read-from-primary --from-literal=data="${data}"] logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | + source ../../functions logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ realpath ../../.. logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | ++++ pwd logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-self-healing logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | ++ test_name=gr-self-healing logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ export GIT_BRANCH=PR-523 logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ GIT_BRANCH=PR-523 logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ export VERSION=PR-523-f00253e logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ VERSION=PR-523-f00253e logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | ++++ which gdate logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | ++++ which date logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ date=/usr/bin/date logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ command -v oc logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ kubectl get nodes logger.go:42: 07:35:50 | gr-self-healing/4-read-from-primary | +++ grep '^minikube' logger.go:42: 07:35:51 | gr-self-healing/4-read-from-primary | ++++ get_cluster_name logger.go:42: 07:35:51 | gr-self-healing/4-read-from-primary | ++++ kubectl -n kuttl-test-emerging-raptor get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 07:35:51 | gr-self-healing/4-read-from-primary | +++ get_mysql_router_service gr-self-healing logger.go:42: 07:35:51 | gr-self-healing/4-read-from-primary | +++ local cluster=gr-self-healing logger.go:42: 07:35:51 | gr-self-healing/4-read-from-primary | +++ echo gr-self-healing-router logger.go:42: 07:35:51 | gr-self-healing/4-read-from-primary | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:35:51 | gr-self-healing/4-read-from-primary | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 07:35:51 | gr-self-healing/4-read-from-primary | ++ local 'uri=-h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:35:51 | gr-self-healing/4-read-from-primary | ++ local pod= logger.go:42: 07:35:51 | gr-self-healing/4-read-from-primary | +++ get_client_pod logger.go:42: 07:35:51 | gr-self-healing/4-read-from-primary | +++ kubectl -n kuttl-test-emerging-raptor get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 07:35:52 | gr-self-healing/4-read-from-primary | ++ client_pod=mysql-client logger.go:42: 07:35:52 | gr-self-healing/4-read-from-primary | ++ wait_pod mysql-client logger.go:42: 07:35:52 | gr-self-healing/4-read-from-primary | ++ local pod=mysql-client logger.go:42: 07:35:52 | gr-self-healing/4-read-from-primary | ++ set +o xtrace logger.go:42: 07:35:52 | gr-self-healing/4-read-from-primary | mysql-clienttrue logger.go:42: 07:35:52 | gr-self-healing/4-read-from-primary | ++ kubectl -n kuttl-test-emerging-raptor exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:35:52 | gr-self-healing/4-read-from-primary | ++ sed -e 's/mysql: //' logger.go:42: 07:35:52 | gr-self-healing/4-read-from-primary | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 07:35:53 | gr-self-healing/4-read-from-primary | + data=100500 logger.go:42: 07:35:53 | gr-self-healing/4-read-from-primary | + kubectl create configmap -n kuttl-test-emerging-raptor 04-read-from-primary --from-literal=data=100500 logger.go:42: 07:35:54 | gr-self-healing/4-read-from-primary | configmap/04-read-from-primary created logger.go:42: 07:35:55 | gr-self-healing/4-read-from-primary | test step completed 4-read-from-primary logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | starting test step 5-kill-primary logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | running command: [sh -c set -o errexit set -o xtrace source ../../functions init_pod="$(get_primary_from_group_replication)" kill_pods "${NAMESPACE}" "pod" "$init_pod" "" "primary" sleep 10 # wait a bit for pod to be killed if [ "$init_pod" == "$(get_primary_from_group_replication)" ]; then echo "primary pod was not killed! something went wrong." exit 1 fi] logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | + source ../../functions logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ realpath ../../.. logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | ++++ pwd logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-self-healing logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | ++ test_name=gr-self-healing logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ export GIT_BRANCH=PR-523 logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ GIT_BRANCH=PR-523 logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ export VERSION=PR-523-f00253e logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ VERSION=PR-523-f00253e logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | ++++ which gdate logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | ++++ which date logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ date=/usr/bin/date logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ command -v oc logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ kubectl get nodes logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ grep '^minikube' logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | ++ get_primary_from_group_replication logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | ++ cut -d. -f1 logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | ++++ get_cluster_name logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | ++++ kubectl -n kuttl-test-emerging-raptor get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ get_mysql_router_service gr-self-healing logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ local cluster=gr-self-healing logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ echo gr-self-healing-router logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | ++ run_mysql 'SELECT MEMBER_HOST FROM performance_schema.replication_group_members where MEMBER_ROLE='\''PRIMARY'\'';' '-h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | ++ local 'command=SELECT MEMBER_HOST FROM performance_schema.replication_group_members where MEMBER_ROLE='\''PRIMARY'\'';' logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | ++ local 'uri=-h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | ++ local pod= logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ get_client_pod logger.go:42: 07:35:55 | gr-self-healing/5-kill-primary | +++ kubectl -n kuttl-test-emerging-raptor get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 07:35:56 | gr-self-healing/5-kill-primary | ++ client_pod=mysql-client logger.go:42: 07:35:56 | gr-self-healing/5-kill-primary | ++ wait_pod mysql-client logger.go:42: 07:35:56 | gr-self-healing/5-kill-primary | ++ local pod=mysql-client logger.go:42: 07:35:56 | gr-self-healing/5-kill-primary | ++ set +o xtrace logger.go:42: 07:35:56 | gr-self-healing/5-kill-primary | mysql-clienttrue logger.go:42: 07:35:56 | gr-self-healing/5-kill-primary | ++ kubectl -n kuttl-test-emerging-raptor exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT MEMBER_HOST FROM performance_schema.replication_group_members where MEMBER_ROLE='\''PRIMARY'\'';" | mysql -sN -h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:35:56 | gr-self-healing/5-kill-primary | ++ sed -e 's/mysql: //' logger.go:42: 07:35:56 | gr-self-healing/5-kill-primary | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 07:35:57 | gr-self-healing/5-kill-primary | + init_pod=gr-self-healing-mysql-0 logger.go:42: 07:35:57 | gr-self-healing/5-kill-primary | + kill_pods kuttl-test-emerging-raptor pod gr-self-healing-mysql-0 '' primary logger.go:42: 07:35:57 | gr-self-healing/5-kill-primary | + local ns=kuttl-test-emerging-raptor logger.go:42: 07:35:57 | gr-self-healing/5-kill-primary | + local selector=pod logger.go:42: 07:35:57 | gr-self-healing/5-kill-primary | + local pod_label=gr-self-healing-mysql-0 logger.go:42: 07:35:57 | gr-self-healing/5-kill-primary | + local label_value= logger.go:42: 07:35:57 | gr-self-healing/5-kill-primary | + local chaos_suffix=primary logger.go:42: 07:35:57 | gr-self-healing/5-kill-primary | + '[' pod == pod ']' logger.go:42: 07:35:57 | gr-self-healing/5-kill-primary | + kubectl apply --namespace kuttl-test-emerging-raptor -f - logger.go:42: 07:35:57 | gr-self-healing/5-kill-primary | + yq eval ' logger.go:42: 07:35:57 | gr-self-healing/5-kill-primary | .metadata.name = "chaos-pod-kill-primary" | logger.go:42: 07:35:57 | gr-self-healing/5-kill-primary | del(.spec.selector.pods.test-namespace) | logger.go:42: 07:35:57 | gr-self-healing/5-kill-primary | .spec.selector.pods.kuttl-test-emerging-raptor[0] = "gr-self-healing-mysql-0"' /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf/chaos-pod-kill.yml logger.go:42: 07:35:59 | gr-self-healing/5-kill-primary | podchaos.chaos-mesh.org/chaos-pod-kill-primary created logger.go:42: 07:35:59 | gr-self-healing/5-kill-primary | + sleep 5 logger.go:42: 07:36:04 | gr-self-healing/5-kill-primary | + sleep 10 logger.go:42: 07:36:14 | gr-self-healing/5-kill-primary | ++ get_primary_from_group_replication logger.go:42: 07:36:14 | gr-self-healing/5-kill-primary | ++ cut -d. -f1 logger.go:42: 07:36:14 | gr-self-healing/5-kill-primary | ++++ get_cluster_name logger.go:42: 07:36:14 | gr-self-healing/5-kill-primary | ++++ kubectl -n kuttl-test-emerging-raptor get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 07:36:15 | gr-self-healing/5-kill-primary | +++ get_mysql_router_service gr-self-healing logger.go:42: 07:36:15 | gr-self-healing/5-kill-primary | +++ local cluster=gr-self-healing logger.go:42: 07:36:15 | gr-self-healing/5-kill-primary | +++ echo gr-self-healing-router logger.go:42: 07:36:15 | gr-self-healing/5-kill-primary | ++ run_mysql 'SELECT MEMBER_HOST FROM performance_schema.replication_group_members where MEMBER_ROLE='\''PRIMARY'\'';' '-h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:36:15 | gr-self-healing/5-kill-primary | ++ local 'command=SELECT MEMBER_HOST FROM performance_schema.replication_group_members where MEMBER_ROLE='\''PRIMARY'\'';' logger.go:42: 07:36:15 | gr-self-healing/5-kill-primary | ++ local 'uri=-h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:36:15 | gr-self-healing/5-kill-primary | ++ local pod= logger.go:42: 07:36:15 | gr-self-healing/5-kill-primary | +++ get_client_pod logger.go:42: 07:36:15 | gr-self-healing/5-kill-primary | +++ kubectl -n kuttl-test-emerging-raptor get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 07:36:15 | gr-self-healing/5-kill-primary | ++ client_pod=mysql-client logger.go:42: 07:36:15 | gr-self-healing/5-kill-primary | ++ wait_pod mysql-client logger.go:42: 07:36:15 | gr-self-healing/5-kill-primary | ++ local pod=mysql-client logger.go:42: 07:36:15 | gr-self-healing/5-kill-primary | ++ set +o xtrace logger.go:42: 07:36:15 | gr-self-healing/5-kill-primary | mysql-clienttrue logger.go:42: 07:36:15 | gr-self-healing/5-kill-primary | ++ sed -e 's/mysql: //' logger.go:42: 07:36:15 | gr-self-healing/5-kill-primary | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 07:36:15 | gr-self-healing/5-kill-primary | ++ kubectl -n kuttl-test-emerging-raptor exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT MEMBER_HOST FROM performance_schema.replication_group_members where MEMBER_ROLE='\''PRIMARY'\'';" | mysql -sN -h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:36:17 | gr-self-healing/5-kill-primary | + '[' gr-self-healing-mysql-0 == gr-self-healing-mysql-1 ']' logger.go:42: 07:36:50 | gr-self-healing/5-kill-primary | test step completed 5-kill-primary logger.go:42: 07:36:50 | gr-self-healing/6-write-data | starting test step 6-write-data logger.go:42: 07:36:50 | gr-self-healing/6-write-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "INSERT myDB.myTable (id) VALUES (100501)" \ "-h $(get_mysql_router_service $(get_cluster_name)) -P 6446 -uroot -proot_password"] logger.go:42: 07:36:50 | gr-self-healing/6-write-data | + source ../../functions logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ realpath ../../.. logger.go:42: 07:36:50 | gr-self-healing/6-write-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:36:50 | gr-self-healing/6-write-data | ++++ pwd logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-self-healing logger.go:42: 07:36:50 | gr-self-healing/6-write-data | ++ test_name=gr-self-healing logger.go:42: 07:36:50 | gr-self-healing/6-write-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:36:50 | gr-self-healing/6-write-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ export GIT_BRANCH=PR-523 logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ GIT_BRANCH=PR-523 logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ export VERSION=PR-523-f00253e logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ VERSION=PR-523-f00253e logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:36:50 | gr-self-healing/6-write-data | ++++ which gdate logger.go:42: 07:36:50 | gr-self-healing/6-write-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 07:36:50 | gr-self-healing/6-write-data | ++++ which date logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ date=/usr/bin/date logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ command -v oc logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ kubectl get nodes logger.go:42: 07:36:50 | gr-self-healing/6-write-data | +++ grep '^minikube' logger.go:42: 07:36:51 | gr-self-healing/6-write-data | +++ get_cluster_name logger.go:42: 07:36:51 | gr-self-healing/6-write-data | +++ kubectl -n kuttl-test-emerging-raptor get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 07:36:51 | gr-self-healing/6-write-data | ++ get_mysql_router_service gr-self-healing logger.go:42: 07:36:51 | gr-self-healing/6-write-data | ++ local cluster=gr-self-healing logger.go:42: 07:36:51 | gr-self-healing/6-write-data | ++ echo gr-self-healing-router logger.go:42: 07:36:51 | gr-self-healing/6-write-data | + run_mysql 'INSERT myDB.myTable (id) VALUES (100501)' '-h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:36:51 | gr-self-healing/6-write-data | + local 'command=INSERT myDB.myTable (id) VALUES (100501)' logger.go:42: 07:36:51 | gr-self-healing/6-write-data | + local 'uri=-h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:36:51 | gr-self-healing/6-write-data | + local pod= logger.go:42: 07:36:51 | gr-self-healing/6-write-data | ++ get_client_pod logger.go:42: 07:36:51 | gr-self-healing/6-write-data | ++ kubectl -n kuttl-test-emerging-raptor get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 07:36:52 | gr-self-healing/6-write-data | + client_pod=mysql-client logger.go:42: 07:36:52 | gr-self-healing/6-write-data | + wait_pod mysql-client logger.go:42: 07:36:52 | gr-self-healing/6-write-data | + local pod=mysql-client logger.go:42: 07:36:52 | gr-self-healing/6-write-data | + set +o xtrace logger.go:42: 07:36:52 | gr-self-healing/6-write-data | mysql-clienttrue logger.go:42: 07:36:52 | gr-self-healing/6-write-data | + kubectl -n kuttl-test-emerging-raptor exec mysql-client -- bash -c 'printf '\''%s\n'\'' "INSERT myDB.myTable (id) VALUES (100501)" | mysql -sN -h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:36:52 | gr-self-healing/6-write-data | + sed -e 's/mysql: //' logger.go:42: 07:36:52 | gr-self-healing/6-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 07:36:53 | gr-self-healing/6-write-data | + : logger.go:42: 07:36:54 | gr-self-healing/6-write-data | test step completed 6-write-data logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | starting test step 7-read-from-replicas logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | running command: [sh -c set -o errexit set -o xtrace source ../../functions for i in 0 1 2; do host=$(get_mysql_headless_fqdn $(get_cluster_name) $i) data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${host} -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 07-read-from-replicas-${i} --from-literal=data="${data}" done] logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | + source ../../functions logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ realpath ../../.. logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | ++++ pwd logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-self-healing logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | ++ test_name=gr-self-healing logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ export GIT_BRANCH=PR-523 logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ GIT_BRANCH=PR-523 logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ export VERSION=PR-523-f00253e logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ VERSION=PR-523-f00253e logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | ++++ which gdate logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | ++++ which date logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ date=/usr/bin/date logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ command -v oc logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ kubectl get nodes logger.go:42: 07:36:54 | gr-self-healing/7-read-from-replicas | +++ grep '^minikube' logger.go:42: 07:36:55 | gr-self-healing/7-read-from-replicas | + for i in 0 1 2 logger.go:42: 07:36:55 | gr-self-healing/7-read-from-replicas | +++ get_cluster_name logger.go:42: 07:36:55 | gr-self-healing/7-read-from-replicas | +++ kubectl -n kuttl-test-emerging-raptor get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 07:36:55 | gr-self-healing/7-read-from-replicas | ++ get_mysql_headless_fqdn gr-self-healing 0 logger.go:42: 07:36:55 | gr-self-healing/7-read-from-replicas | ++ local cluster=gr-self-healing logger.go:42: 07:36:55 | gr-self-healing/7-read-from-replicas | ++ local index=0 logger.go:42: 07:36:55 | gr-self-healing/7-read-from-replicas | ++ echo gr-self-healing-mysql-0.gr-self-healing-mysql logger.go:42: 07:36:55 | gr-self-healing/7-read-from-replicas | + host=gr-self-healing-mysql-0.gr-self-healing-mysql logger.go:42: 07:36:55 | gr-self-healing/7-read-from-replicas | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-self-healing-mysql-0.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:36:55 | gr-self-healing/7-read-from-replicas | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 07:36:55 | gr-self-healing/7-read-from-replicas | ++ local 'uri=-h gr-self-healing-mysql-0.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:36:55 | gr-self-healing/7-read-from-replicas | ++ local pod= logger.go:42: 07:36:55 | gr-self-healing/7-read-from-replicas | +++ get_client_pod logger.go:42: 07:36:55 | gr-self-healing/7-read-from-replicas | +++ kubectl -n kuttl-test-emerging-raptor get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 07:36:55 | gr-self-healing/7-read-from-replicas | ++ client_pod=mysql-client logger.go:42: 07:36:55 | gr-self-healing/7-read-from-replicas | ++ wait_pod mysql-client logger.go:42: 07:36:55 | gr-self-healing/7-read-from-replicas | ++ local pod=mysql-client logger.go:42: 07:36:55 | gr-self-healing/7-read-from-replicas | ++ set +o xtrace logger.go:42: 07:36:56 | gr-self-healing/7-read-from-replicas | mysql-clienttrue logger.go:42: 07:36:56 | gr-self-healing/7-read-from-replicas | ++ kubectl -n kuttl-test-emerging-raptor exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-self-healing-mysql-0.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:36:56 | gr-self-healing/7-read-from-replicas | ++ sed -e 's/mysql: //' logger.go:42: 07:36:56 | gr-self-healing/7-read-from-replicas | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 07:36:57 | gr-self-healing/7-read-from-replicas | + data='100500 logger.go:42: 07:36:57 | gr-self-healing/7-read-from-replicas | 100501' logger.go:42: 07:36:57 | gr-self-healing/7-read-from-replicas | + kubectl create configmap -n kuttl-test-emerging-raptor 07-read-from-replicas-0 '--from-literal=data=100500 logger.go:42: 07:36:57 | gr-self-healing/7-read-from-replicas | 100501' logger.go:42: 07:36:57 | gr-self-healing/7-read-from-replicas | configmap/07-read-from-replicas-0 created logger.go:42: 07:36:57 | gr-self-healing/7-read-from-replicas | + for i in 0 1 2 logger.go:42: 07:36:57 | gr-self-healing/7-read-from-replicas | +++ get_cluster_name logger.go:42: 07:36:57 | gr-self-healing/7-read-from-replicas | +++ kubectl -n kuttl-test-emerging-raptor get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 07:36:58 | gr-self-healing/7-read-from-replicas | ++ get_mysql_headless_fqdn gr-self-healing 1 logger.go:42: 07:36:58 | gr-self-healing/7-read-from-replicas | ++ local cluster=gr-self-healing logger.go:42: 07:36:58 | gr-self-healing/7-read-from-replicas | ++ local index=1 logger.go:42: 07:36:58 | gr-self-healing/7-read-from-replicas | ++ echo gr-self-healing-mysql-1.gr-self-healing-mysql logger.go:42: 07:36:58 | gr-self-healing/7-read-from-replicas | + host=gr-self-healing-mysql-1.gr-self-healing-mysql logger.go:42: 07:36:58 | gr-self-healing/7-read-from-replicas | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-self-healing-mysql-1.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:36:58 | gr-self-healing/7-read-from-replicas | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 07:36:58 | gr-self-healing/7-read-from-replicas | ++ local 'uri=-h gr-self-healing-mysql-1.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:36:58 | gr-self-healing/7-read-from-replicas | ++ local pod= logger.go:42: 07:36:58 | gr-self-healing/7-read-from-replicas | +++ get_client_pod logger.go:42: 07:36:58 | gr-self-healing/7-read-from-replicas | +++ kubectl -n kuttl-test-emerging-raptor get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 07:36:58 | gr-self-healing/7-read-from-replicas | ++ client_pod=mysql-client logger.go:42: 07:36:58 | gr-self-healing/7-read-from-replicas | ++ wait_pod mysql-client logger.go:42: 07:36:58 | gr-self-healing/7-read-from-replicas | ++ local pod=mysql-client logger.go:42: 07:36:58 | gr-self-healing/7-read-from-replicas | ++ set +o xtrace logger.go:42: 07:36:58 | gr-self-healing/7-read-from-replicas | mysql-clienttrue logger.go:42: 07:36:58 | gr-self-healing/7-read-from-replicas | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 07:36:58 | gr-self-healing/7-read-from-replicas | ++ kubectl -n kuttl-test-emerging-raptor exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-self-healing-mysql-1.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:36:58 | gr-self-healing/7-read-from-replicas | ++ sed -e 's/mysql: //' logger.go:42: 07:37:00 | gr-self-healing/7-read-from-replicas | + data='100500 logger.go:42: 07:37:00 | gr-self-healing/7-read-from-replicas | 100501' logger.go:42: 07:37:00 | gr-self-healing/7-read-from-replicas | + kubectl create configmap -n kuttl-test-emerging-raptor 07-read-from-replicas-1 '--from-literal=data=100500 logger.go:42: 07:37:00 | gr-self-healing/7-read-from-replicas | 100501' logger.go:42: 07:37:00 | gr-self-healing/7-read-from-replicas | configmap/07-read-from-replicas-1 created logger.go:42: 07:37:00 | gr-self-healing/7-read-from-replicas | + for i in 0 1 2 logger.go:42: 07:37:00 | gr-self-healing/7-read-from-replicas | +++ get_cluster_name logger.go:42: 07:37:00 | gr-self-healing/7-read-from-replicas | +++ kubectl -n kuttl-test-emerging-raptor get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 07:37:00 | gr-self-healing/7-read-from-replicas | ++ get_mysql_headless_fqdn gr-self-healing 2 logger.go:42: 07:37:00 | gr-self-healing/7-read-from-replicas | ++ local cluster=gr-self-healing logger.go:42: 07:37:00 | gr-self-healing/7-read-from-replicas | ++ local index=2 logger.go:42: 07:37:00 | gr-self-healing/7-read-from-replicas | ++ echo gr-self-healing-mysql-2.gr-self-healing-mysql logger.go:42: 07:37:00 | gr-self-healing/7-read-from-replicas | + host=gr-self-healing-mysql-2.gr-self-healing-mysql logger.go:42: 07:37:00 | gr-self-healing/7-read-from-replicas | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-self-healing-mysql-2.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:37:00 | gr-self-healing/7-read-from-replicas | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 07:37:00 | gr-self-healing/7-read-from-replicas | ++ local 'uri=-h gr-self-healing-mysql-2.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:37:00 | gr-self-healing/7-read-from-replicas | ++ local pod= logger.go:42: 07:37:00 | gr-self-healing/7-read-from-replicas | +++ get_client_pod logger.go:42: 07:37:00 | gr-self-healing/7-read-from-replicas | +++ kubectl -n kuttl-test-emerging-raptor get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 07:37:01 | gr-self-healing/7-read-from-replicas | ++ client_pod=mysql-client logger.go:42: 07:37:01 | gr-self-healing/7-read-from-replicas | ++ wait_pod mysql-client logger.go:42: 07:37:01 | gr-self-healing/7-read-from-replicas | ++ local pod=mysql-client logger.go:42: 07:37:01 | gr-self-healing/7-read-from-replicas | ++ set +o xtrace logger.go:42: 07:37:01 | gr-self-healing/7-read-from-replicas | mysql-clienttrue logger.go:42: 07:37:01 | gr-self-healing/7-read-from-replicas | ++ sed -e 's/mysql: //' logger.go:42: 07:37:01 | gr-self-healing/7-read-from-replicas | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 07:37:01 | gr-self-healing/7-read-from-replicas | ++ kubectl -n kuttl-test-emerging-raptor exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-self-healing-mysql-2.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:37:02 | gr-self-healing/7-read-from-replicas | + data='100500 logger.go:42: 07:37:02 | gr-self-healing/7-read-from-replicas | 100501' logger.go:42: 07:37:02 | gr-self-healing/7-read-from-replicas | + kubectl create configmap -n kuttl-test-emerging-raptor 07-read-from-replicas-2 '--from-literal=data=100500 logger.go:42: 07:37:02 | gr-self-healing/7-read-from-replicas | 100501' logger.go:42: 07:37:03 | gr-self-healing/7-read-from-replicas | configmap/07-read-from-replicas-2 created logger.go:42: 07:37:04 | gr-self-healing/7-read-from-replicas | test step completed 7-read-from-replicas logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | starting test step 8-failure-primary logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | running command: [sh -c set -o errexit set -o xtrace source ../../functions failure_pod "${NAMESPACE}" "$(get_primary_from_group_replication)" "primary" sleep 10 # wait a bit for pod to be killed] logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | + source ../../functions logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ realpath ../../.. logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | ++++ pwd logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-self-healing logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | ++ test_name=gr-self-healing logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ export GIT_BRANCH=PR-523 logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ GIT_BRANCH=PR-523 logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ export VERSION=PR-523-f00253e logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ VERSION=PR-523-f00253e logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | ++++ which gdate logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | ++++ which date logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ date=/usr/bin/date logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ command -v oc logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ kubectl get nodes logger.go:42: 07:37:04 | gr-self-healing/8-failure-primary | +++ grep '^minikube' logger.go:42: 07:37:05 | gr-self-healing/8-failure-primary | ++ get_primary_from_group_replication logger.go:42: 07:37:05 | gr-self-healing/8-failure-primary | ++ cut -d. -f1 logger.go:42: 07:37:05 | gr-self-healing/8-failure-primary | ++++ get_cluster_name logger.go:42: 07:37:05 | gr-self-healing/8-failure-primary | ++++ kubectl -n kuttl-test-emerging-raptor get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 07:37:05 | gr-self-healing/8-failure-primary | +++ get_mysql_router_service gr-self-healing logger.go:42: 07:37:05 | gr-self-healing/8-failure-primary | +++ local cluster=gr-self-healing logger.go:42: 07:37:05 | gr-self-healing/8-failure-primary | +++ echo gr-self-healing-router logger.go:42: 07:37:05 | gr-self-healing/8-failure-primary | ++ run_mysql 'SELECT MEMBER_HOST FROM performance_schema.replication_group_members where MEMBER_ROLE='\''PRIMARY'\'';' '-h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:37:05 | gr-self-healing/8-failure-primary | ++ local 'command=SELECT MEMBER_HOST FROM performance_schema.replication_group_members where MEMBER_ROLE='\''PRIMARY'\'';' logger.go:42: 07:37:05 | gr-self-healing/8-failure-primary | ++ local 'uri=-h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:37:05 | gr-self-healing/8-failure-primary | ++ local pod= logger.go:42: 07:37:05 | gr-self-healing/8-failure-primary | +++ get_client_pod logger.go:42: 07:37:05 | gr-self-healing/8-failure-primary | +++ kubectl -n kuttl-test-emerging-raptor get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 07:37:05 | gr-self-healing/8-failure-primary | ++ client_pod=mysql-client logger.go:42: 07:37:05 | gr-self-healing/8-failure-primary | ++ wait_pod mysql-client logger.go:42: 07:37:05 | gr-self-healing/8-failure-primary | ++ local pod=mysql-client logger.go:42: 07:37:05 | gr-self-healing/8-failure-primary | ++ set +o xtrace logger.go:42: 07:37:06 | gr-self-healing/8-failure-primary | mysql-clienttrue logger.go:42: 07:37:06 | gr-self-healing/8-failure-primary | ++ kubectl -n kuttl-test-emerging-raptor exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT MEMBER_HOST FROM performance_schema.replication_group_members where MEMBER_ROLE='\''PRIMARY'\'';" | mysql -sN -h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:37:06 | gr-self-healing/8-failure-primary | ++ sed -e 's/mysql: //' logger.go:42: 07:37:06 | gr-self-healing/8-failure-primary | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 07:37:07 | gr-self-healing/8-failure-primary | + failure_pod kuttl-test-emerging-raptor gr-self-healing-mysql-1 primary logger.go:42: 07:37:07 | gr-self-healing/8-failure-primary | + local ns=kuttl-test-emerging-raptor logger.go:42: 07:37:07 | gr-self-healing/8-failure-primary | + local pod=gr-self-healing-mysql-1 logger.go:42: 07:37:07 | gr-self-healing/8-failure-primary | + local chaos_suffix=primary logger.go:42: 07:37:07 | gr-self-healing/8-failure-primary | + yq eval ' logger.go:42: 07:37:07 | gr-self-healing/8-failure-primary | .metadata.name = "chaos-pod-failure-primary" | logger.go:42: 07:37:07 | gr-self-healing/8-failure-primary | del(.spec.selector.pods.test-namespace) | logger.go:42: 07:37:07 | gr-self-healing/8-failure-primary | .spec.selector.pods.kuttl-test-emerging-raptor[0] = "gr-self-healing-mysql-1"' /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf/chaos-pod-failure.yml logger.go:42: 07:37:07 | gr-self-healing/8-failure-primary | + kubectl apply --namespace kuttl-test-emerging-raptor -f - logger.go:42: 07:37:08 | gr-self-healing/8-failure-primary | podchaos.chaos-mesh.org/chaos-pod-failure-primary created logger.go:42: 07:37:08 | gr-self-healing/8-failure-primary | + sleep 5 logger.go:42: 07:37:13 | gr-self-healing/8-failure-primary | + sleep 10 logger.go:42: 07:38:44 | gr-self-healing/8-failure-primary | test step completed 8-failure-primary logger.go:42: 07:38:44 | gr-self-healing/9-write-data | starting test step 9-write-data logger.go:42: 07:38:44 | gr-self-healing/9-write-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "INSERT myDB.myTable (id) VALUES (100502)" \ "-h $(get_mysql_router_service $(get_cluster_name)) -P 6446 -uroot -proot_password"] logger.go:42: 07:38:44 | gr-self-healing/9-write-data | + source ../../functions logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ realpath ../../.. logger.go:42: 07:38:44 | gr-self-healing/9-write-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:38:44 | gr-self-healing/9-write-data | ++++ pwd logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-self-healing logger.go:42: 07:38:44 | gr-self-healing/9-write-data | ++ test_name=gr-self-healing logger.go:42: 07:38:44 | gr-self-healing/9-write-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:38:44 | gr-self-healing/9-write-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ export GIT_BRANCH=PR-523 logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ GIT_BRANCH=PR-523 logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ export VERSION=PR-523-f00253e logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ VERSION=PR-523-f00253e logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:38:44 | gr-self-healing/9-write-data | ++++ which gdate logger.go:42: 07:38:44 | gr-self-healing/9-write-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 07:38:44 | gr-self-healing/9-write-data | ++++ which date logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ date=/usr/bin/date logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ command -v oc logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ kubectl get nodes logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ grep '^minikube' logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ get_cluster_name logger.go:42: 07:38:44 | gr-self-healing/9-write-data | +++ kubectl -n kuttl-test-emerging-raptor get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 07:38:45 | gr-self-healing/9-write-data | ++ get_mysql_router_service gr-self-healing logger.go:42: 07:38:45 | gr-self-healing/9-write-data | ++ local cluster=gr-self-healing logger.go:42: 07:38:45 | gr-self-healing/9-write-data | ++ echo gr-self-healing-router logger.go:42: 07:38:45 | gr-self-healing/9-write-data | + run_mysql 'INSERT myDB.myTable (id) VALUES (100502)' '-h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:38:45 | gr-self-healing/9-write-data | + local 'command=INSERT myDB.myTable (id) VALUES (100502)' logger.go:42: 07:38:45 | gr-self-healing/9-write-data | + local 'uri=-h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:38:45 | gr-self-healing/9-write-data | + local pod= logger.go:42: 07:38:45 | gr-self-healing/9-write-data | ++ get_client_pod logger.go:42: 07:38:45 | gr-self-healing/9-write-data | ++ kubectl -n kuttl-test-emerging-raptor get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 07:38:45 | gr-self-healing/9-write-data | + client_pod=mysql-client logger.go:42: 07:38:45 | gr-self-healing/9-write-data | + wait_pod mysql-client logger.go:42: 07:38:45 | gr-self-healing/9-write-data | + local pod=mysql-client logger.go:42: 07:38:45 | gr-self-healing/9-write-data | + set +o xtrace logger.go:42: 07:38:46 | gr-self-healing/9-write-data | mysql-clienttrue logger.go:42: 07:38:46 | gr-self-healing/9-write-data | + kubectl -n kuttl-test-emerging-raptor exec mysql-client -- bash -c 'printf '\''%s\n'\'' "INSERT myDB.myTable (id) VALUES (100502)" | mysql -sN -h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:38:46 | gr-self-healing/9-write-data | + sed -e 's/mysql: //' logger.go:42: 07:38:46 | gr-self-healing/9-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 07:38:47 | gr-self-healing/9-write-data | + : logger.go:42: 07:38:48 | gr-self-healing/9-write-data | test step completed 9-write-data logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | starting test step 10-read-from-replicas logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | running command: [sh -c set -o errexit set -o xtrace source ../../functions for i in 0 1 2; do host=$(get_mysql_headless_fqdn $(get_cluster_name) $i) data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${host} -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 10-read-from-replicas-${i} --from-literal=data="${data}" done] logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | + source ../../functions logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ realpath ../../.. logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | ++++ pwd logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-self-healing logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | ++ test_name=gr-self-healing logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ export GIT_BRANCH=PR-523 logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ GIT_BRANCH=PR-523 logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ export VERSION=PR-523-f00253e logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ VERSION=PR-523-f00253e logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | ++++ which gdate logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | ++++ which date logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ date=/usr/bin/date logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ command -v oc logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ kubectl get nodes logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ grep '^minikube' logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | + for i in 0 1 2 logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ get_cluster_name logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ kubectl -n kuttl-test-emerging-raptor get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | ++ get_mysql_headless_fqdn gr-self-healing 0 logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | ++ local cluster=gr-self-healing logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | ++ local index=0 logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | ++ echo gr-self-healing-mysql-0.gr-self-healing-mysql logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | + host=gr-self-healing-mysql-0.gr-self-healing-mysql logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-self-healing-mysql-0.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | ++ local 'uri=-h gr-self-healing-mysql-0.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | ++ local pod= logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ get_client_pod logger.go:42: 07:38:48 | gr-self-healing/10-read-from-replicas | +++ kubectl -n kuttl-test-emerging-raptor get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 07:38:49 | gr-self-healing/10-read-from-replicas | ++ client_pod=mysql-client logger.go:42: 07:38:49 | gr-self-healing/10-read-from-replicas | ++ wait_pod mysql-client logger.go:42: 07:38:49 | gr-self-healing/10-read-from-replicas | ++ local pod=mysql-client logger.go:42: 07:38:49 | gr-self-healing/10-read-from-replicas | ++ set +o xtrace logger.go:42: 07:38:49 | gr-self-healing/10-read-from-replicas | mysql-clienttrue logger.go:42: 07:38:49 | gr-self-healing/10-read-from-replicas | ++ kubectl -n kuttl-test-emerging-raptor exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-self-healing-mysql-0.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:38:49 | gr-self-healing/10-read-from-replicas | ++ sed -e 's/mysql: //' logger.go:42: 07:38:49 | gr-self-healing/10-read-from-replicas | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 07:38:50 | gr-self-healing/10-read-from-replicas | + data='100500 logger.go:42: 07:38:50 | gr-self-healing/10-read-from-replicas | 100501 logger.go:42: 07:38:50 | gr-self-healing/10-read-from-replicas | 100502' logger.go:42: 07:38:50 | gr-self-healing/10-read-from-replicas | + kubectl create configmap -n kuttl-test-emerging-raptor 10-read-from-replicas-0 '--from-literal=data=100500 logger.go:42: 07:38:50 | gr-self-healing/10-read-from-replicas | 100501 logger.go:42: 07:38:50 | gr-self-healing/10-read-from-replicas | 100502' logger.go:42: 07:38:51 | gr-self-healing/10-read-from-replicas | configmap/10-read-from-replicas-0 created logger.go:42: 07:38:51 | gr-self-healing/10-read-from-replicas | + for i in 0 1 2 logger.go:42: 07:38:51 | gr-self-healing/10-read-from-replicas | +++ get_cluster_name logger.go:42: 07:38:51 | gr-self-healing/10-read-from-replicas | +++ kubectl -n kuttl-test-emerging-raptor get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 07:38:51 | gr-self-healing/10-read-from-replicas | ++ get_mysql_headless_fqdn gr-self-healing 1 logger.go:42: 07:38:51 | gr-self-healing/10-read-from-replicas | ++ local cluster=gr-self-healing logger.go:42: 07:38:51 | gr-self-healing/10-read-from-replicas | ++ local index=1 logger.go:42: 07:38:51 | gr-self-healing/10-read-from-replicas | ++ echo gr-self-healing-mysql-1.gr-self-healing-mysql logger.go:42: 07:38:51 | gr-self-healing/10-read-from-replicas | + host=gr-self-healing-mysql-1.gr-self-healing-mysql logger.go:42: 07:38:51 | gr-self-healing/10-read-from-replicas | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-self-healing-mysql-1.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:38:51 | gr-self-healing/10-read-from-replicas | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 07:38:51 | gr-self-healing/10-read-from-replicas | ++ local 'uri=-h gr-self-healing-mysql-1.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:38:51 | gr-self-healing/10-read-from-replicas | ++ local pod= logger.go:42: 07:38:51 | gr-self-healing/10-read-from-replicas | +++ get_client_pod logger.go:42: 07:38:51 | gr-self-healing/10-read-from-replicas | +++ kubectl -n kuttl-test-emerging-raptor get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 07:38:52 | gr-self-healing/10-read-from-replicas | ++ client_pod=mysql-client logger.go:42: 07:38:52 | gr-self-healing/10-read-from-replicas | ++ wait_pod mysql-client logger.go:42: 07:38:52 | gr-self-healing/10-read-from-replicas | ++ local pod=mysql-client logger.go:42: 07:38:52 | gr-self-healing/10-read-from-replicas | ++ set +o xtrace logger.go:42: 07:38:52 | gr-self-healing/10-read-from-replicas | mysql-clienttrue logger.go:42: 07:38:52 | gr-self-healing/10-read-from-replicas | ++ kubectl -n kuttl-test-emerging-raptor exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-self-healing-mysql-1.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:38:52 | gr-self-healing/10-read-from-replicas | ++ sed -e 's/mysql: //' logger.go:42: 07:38:52 | gr-self-healing/10-read-from-replicas | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 07:38:53 | gr-self-healing/10-read-from-replicas | + data='100500 logger.go:42: 07:38:53 | gr-self-healing/10-read-from-replicas | 100501 logger.go:42: 07:38:53 | gr-self-healing/10-read-from-replicas | 100502' logger.go:42: 07:38:53 | gr-self-healing/10-read-from-replicas | + kubectl create configmap -n kuttl-test-emerging-raptor 10-read-from-replicas-1 '--from-literal=data=100500 logger.go:42: 07:38:53 | gr-self-healing/10-read-from-replicas | 100501 logger.go:42: 07:38:53 | gr-self-healing/10-read-from-replicas | 100502' logger.go:42: 07:38:53 | gr-self-healing/10-read-from-replicas | configmap/10-read-from-replicas-1 created logger.go:42: 07:38:53 | gr-self-healing/10-read-from-replicas | + for i in 0 1 2 logger.go:42: 07:38:53 | gr-self-healing/10-read-from-replicas | +++ get_cluster_name logger.go:42: 07:38:53 | gr-self-healing/10-read-from-replicas | +++ kubectl -n kuttl-test-emerging-raptor get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 07:38:54 | gr-self-healing/10-read-from-replicas | ++ get_mysql_headless_fqdn gr-self-healing 2 logger.go:42: 07:38:54 | gr-self-healing/10-read-from-replicas | ++ local cluster=gr-self-healing logger.go:42: 07:38:54 | gr-self-healing/10-read-from-replicas | ++ local index=2 logger.go:42: 07:38:54 | gr-self-healing/10-read-from-replicas | ++ echo gr-self-healing-mysql-2.gr-self-healing-mysql logger.go:42: 07:38:54 | gr-self-healing/10-read-from-replicas | + host=gr-self-healing-mysql-2.gr-self-healing-mysql logger.go:42: 07:38:54 | gr-self-healing/10-read-from-replicas | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-self-healing-mysql-2.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:38:54 | gr-self-healing/10-read-from-replicas | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 07:38:54 | gr-self-healing/10-read-from-replicas | ++ local 'uri=-h gr-self-healing-mysql-2.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:38:54 | gr-self-healing/10-read-from-replicas | ++ local pod= logger.go:42: 07:38:54 | gr-self-healing/10-read-from-replicas | +++ get_client_pod logger.go:42: 07:38:54 | gr-self-healing/10-read-from-replicas | +++ kubectl -n kuttl-test-emerging-raptor get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 07:38:54 | gr-self-healing/10-read-from-replicas | ++ client_pod=mysql-client logger.go:42: 07:38:54 | gr-self-healing/10-read-from-replicas | ++ wait_pod mysql-client logger.go:42: 07:38:54 | gr-self-healing/10-read-from-replicas | ++ local pod=mysql-client logger.go:42: 07:38:54 | gr-self-healing/10-read-from-replicas | ++ set +o xtrace logger.go:42: 07:38:55 | gr-self-healing/10-read-from-replicas | mysql-clienttrue logger.go:42: 07:38:55 | gr-self-healing/10-read-from-replicas | ++ sed -e 's/mysql: //' logger.go:42: 07:38:55 | gr-self-healing/10-read-from-replicas | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 07:38:55 | gr-self-healing/10-read-from-replicas | ++ kubectl -n kuttl-test-emerging-raptor exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-self-healing-mysql-2.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:38:56 | gr-self-healing/10-read-from-replicas | + data='100500 logger.go:42: 07:38:56 | gr-self-healing/10-read-from-replicas | 100501 logger.go:42: 07:38:56 | gr-self-healing/10-read-from-replicas | 100502' logger.go:42: 07:38:56 | gr-self-healing/10-read-from-replicas | + kubectl create configmap -n kuttl-test-emerging-raptor 10-read-from-replicas-2 '--from-literal=data=100500 logger.go:42: 07:38:56 | gr-self-healing/10-read-from-replicas | 100501 logger.go:42: 07:38:56 | gr-self-healing/10-read-from-replicas | 100502' logger.go:42: 07:38:56 | gr-self-healing/10-read-from-replicas | configmap/10-read-from-replicas-2 created logger.go:42: 07:38:58 | gr-self-healing/10-read-from-replicas | test step completed 10-read-from-replicas logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | starting test step 11-network-loss-primary logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | running command: [sh -c set -o errexit set -o xtrace source ../../functions network_loss "${NAMESPACE}" "$(get_primary_from_group_replication)" "primary" sleep 30 # wait for new master to get elected] logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | + source ../../functions logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ realpath ../../.. logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | ++++ pwd logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-self-healing logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | ++ test_name=gr-self-healing logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ export GIT_BRANCH=PR-523 logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ GIT_BRANCH=PR-523 logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ export VERSION=PR-523-f00253e logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ VERSION=PR-523-f00253e logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | ++++ which gdate logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | ++++ which date logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ date=/usr/bin/date logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ command -v oc logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ kubectl get nodes logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ grep '^minikube' logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | ++ get_primary_from_group_replication logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | ++ cut -d. -f1 logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | ++++ get_cluster_name logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | ++++ kubectl -n kuttl-test-emerging-raptor get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ get_mysql_router_service gr-self-healing logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ local cluster=gr-self-healing logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ echo gr-self-healing-router logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | ++ run_mysql 'SELECT MEMBER_HOST FROM performance_schema.replication_group_members where MEMBER_ROLE='\''PRIMARY'\'';' '-h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | ++ local 'command=SELECT MEMBER_HOST FROM performance_schema.replication_group_members where MEMBER_ROLE='\''PRIMARY'\'';' logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | ++ local 'uri=-h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | ++ local pod= logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ get_client_pod logger.go:42: 07:38:58 | gr-self-healing/11-network-loss-primary | +++ kubectl -n kuttl-test-emerging-raptor get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 07:38:59 | gr-self-healing/11-network-loss-primary | ++ client_pod=mysql-client logger.go:42: 07:38:59 | gr-self-healing/11-network-loss-primary | ++ wait_pod mysql-client logger.go:42: 07:38:59 | gr-self-healing/11-network-loss-primary | ++ local pod=mysql-client logger.go:42: 07:38:59 | gr-self-healing/11-network-loss-primary | ++ set +o xtrace logger.go:42: 07:38:59 | gr-self-healing/11-network-loss-primary | mysql-clienttrue logger.go:42: 07:38:59 | gr-self-healing/11-network-loss-primary | ++ kubectl -n kuttl-test-emerging-raptor exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT MEMBER_HOST FROM performance_schema.replication_group_members where MEMBER_ROLE='\''PRIMARY'\'';" | mysql -sN -h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:38:59 | gr-self-healing/11-network-loss-primary | ++ sed -e 's/mysql: //' logger.go:42: 07:38:59 | gr-self-healing/11-network-loss-primary | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 07:39:00 | gr-self-healing/11-network-loss-primary | + network_loss kuttl-test-emerging-raptor gr-self-healing-mysql-2 primary logger.go:42: 07:39:00 | gr-self-healing/11-network-loss-primary | + local ns=kuttl-test-emerging-raptor logger.go:42: 07:39:00 | gr-self-healing/11-network-loss-primary | + local pod=gr-self-healing-mysql-2 logger.go:42: 07:39:00 | gr-self-healing/11-network-loss-primary | + local chaos_suffix=primary logger.go:42: 07:39:00 | gr-self-healing/11-network-loss-primary | + kubectl apply --namespace kuttl-test-emerging-raptor -f - logger.go:42: 07:39:00 | gr-self-healing/11-network-loss-primary | + yq eval ' logger.go:42: 07:39:00 | gr-self-healing/11-network-loss-primary | .metadata.name = "chaos-pod-network-loss-primary" | logger.go:42: 07:39:00 | gr-self-healing/11-network-loss-primary | del(.spec.selector.pods.test-namespace) | logger.go:42: 07:39:00 | gr-self-healing/11-network-loss-primary | .spec.selector.pods.kuttl-test-emerging-raptor[0] = "gr-self-healing-mysql-2"' /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf/chaos-network-loss.yml logger.go:42: 07:39:01 | gr-self-healing/11-network-loss-primary | networkchaos.chaos-mesh.org/chaos-pod-network-loss-primary created logger.go:42: 07:39:01 | gr-self-healing/11-network-loss-primary | + sleep 5 logger.go:42: 07:39:06 | gr-self-healing/11-network-loss-primary | + sleep 30 logger.go:42: 07:40:44 | gr-self-healing/11-network-loss-primary | test step completed 11-network-loss-primary logger.go:42: 07:40:44 | gr-self-healing/12-write-data | starting test step 12-write-data logger.go:42: 07:40:44 | gr-self-healing/12-write-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "INSERT myDB.myTable (id) VALUES (100503)" \ "-h $(get_mysql_router_service $(get_cluster_name)) -P 6446 -uroot -proot_password"] logger.go:42: 07:40:44 | gr-self-healing/12-write-data | + source ../../functions logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ realpath ../../.. logger.go:42: 07:40:44 | gr-self-healing/12-write-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:40:44 | gr-self-healing/12-write-data | ++++ pwd logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-self-healing logger.go:42: 07:40:44 | gr-self-healing/12-write-data | ++ test_name=gr-self-healing logger.go:42: 07:40:44 | gr-self-healing/12-write-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:40:44 | gr-self-healing/12-write-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ export GIT_BRANCH=PR-523 logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ GIT_BRANCH=PR-523 logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ export VERSION=PR-523-f00253e logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ VERSION=PR-523-f00253e logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:40:44 | gr-self-healing/12-write-data | ++++ which gdate logger.go:42: 07:40:44 | gr-self-healing/12-write-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 07:40:44 | gr-self-healing/12-write-data | ++++ which date logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ date=/usr/bin/date logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ command -v oc logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ kubectl get nodes logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ grep '^minikube' logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ get_cluster_name logger.go:42: 07:40:44 | gr-self-healing/12-write-data | +++ kubectl -n kuttl-test-emerging-raptor get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 07:40:45 | gr-self-healing/12-write-data | ++ get_mysql_router_service gr-self-healing logger.go:42: 07:40:45 | gr-self-healing/12-write-data | ++ local cluster=gr-self-healing logger.go:42: 07:40:45 | gr-self-healing/12-write-data | ++ echo gr-self-healing-router logger.go:42: 07:40:45 | gr-self-healing/12-write-data | + run_mysql 'INSERT myDB.myTable (id) VALUES (100503)' '-h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:40:45 | gr-self-healing/12-write-data | + local 'command=INSERT myDB.myTable (id) VALUES (100503)' logger.go:42: 07:40:45 | gr-self-healing/12-write-data | + local 'uri=-h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:40:45 | gr-self-healing/12-write-data | + local pod= logger.go:42: 07:40:45 | gr-self-healing/12-write-data | ++ get_client_pod logger.go:42: 07:40:45 | gr-self-healing/12-write-data | ++ kubectl -n kuttl-test-emerging-raptor get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 07:40:45 | gr-self-healing/12-write-data | + client_pod=mysql-client logger.go:42: 07:40:45 | gr-self-healing/12-write-data | + wait_pod mysql-client logger.go:42: 07:40:45 | gr-self-healing/12-write-data | + local pod=mysql-client logger.go:42: 07:40:45 | gr-self-healing/12-write-data | + set +o xtrace logger.go:42: 07:40:45 | gr-self-healing/12-write-data | mysql-clienttrue logger.go:42: 07:40:45 | gr-self-healing/12-write-data | + kubectl -n kuttl-test-emerging-raptor exec mysql-client -- bash -c 'printf '\''%s\n'\'' "INSERT myDB.myTable (id) VALUES (100503)" | mysql -sN -h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:40:45 | gr-self-healing/12-write-data | + sed -e 's/mysql: //' logger.go:42: 07:40:45 | gr-self-healing/12-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 07:40:47 | gr-self-healing/12-write-data | + : logger.go:42: 07:40:47 | gr-self-healing/12-write-data | test step completed 12-write-data logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | starting test step 13-read-from-replicas logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | running command: [sh -c set -o errexit set -o xtrace source ../../functions for i in 0 1 2; do host=$(get_mysql_headless_fqdn $(get_cluster_name) $i) data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${host} -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 13-read-from-replicas-${i} --from-literal=data="${data}" done] logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | + source ../../functions logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ realpath ../../.. logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | ++++ pwd logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-self-healing logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | ++ test_name=gr-self-healing logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ export GIT_BRANCH=PR-523 logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ GIT_BRANCH=PR-523 logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ export VERSION=PR-523-f00253e logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ VERSION=PR-523-f00253e logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | ++++ which gdate logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | ++++ which date logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ date=/usr/bin/date logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ command -v oc logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ kubectl get nodes logger.go:42: 07:40:47 | gr-self-healing/13-read-from-replicas | +++ grep '^minikube' logger.go:42: 07:40:48 | gr-self-healing/13-read-from-replicas | + for i in 0 1 2 logger.go:42: 07:40:48 | gr-self-healing/13-read-from-replicas | +++ get_cluster_name logger.go:42: 07:40:48 | gr-self-healing/13-read-from-replicas | +++ kubectl -n kuttl-test-emerging-raptor get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 07:40:48 | gr-self-healing/13-read-from-replicas | ++ get_mysql_headless_fqdn gr-self-healing 0 logger.go:42: 07:40:48 | gr-self-healing/13-read-from-replicas | ++ local cluster=gr-self-healing logger.go:42: 07:40:48 | gr-self-healing/13-read-from-replicas | ++ local index=0 logger.go:42: 07:40:48 | gr-self-healing/13-read-from-replicas | ++ echo gr-self-healing-mysql-0.gr-self-healing-mysql logger.go:42: 07:40:48 | gr-self-healing/13-read-from-replicas | + host=gr-self-healing-mysql-0.gr-self-healing-mysql logger.go:42: 07:40:48 | gr-self-healing/13-read-from-replicas | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-self-healing-mysql-0.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:40:48 | gr-self-healing/13-read-from-replicas | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 07:40:48 | gr-self-healing/13-read-from-replicas | ++ local 'uri=-h gr-self-healing-mysql-0.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:40:48 | gr-self-healing/13-read-from-replicas | ++ local pod= logger.go:42: 07:40:48 | gr-self-healing/13-read-from-replicas | +++ get_client_pod logger.go:42: 07:40:48 | gr-self-healing/13-read-from-replicas | +++ kubectl -n kuttl-test-emerging-raptor get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 07:40:49 | gr-self-healing/13-read-from-replicas | ++ client_pod=mysql-client logger.go:42: 07:40:49 | gr-self-healing/13-read-from-replicas | ++ wait_pod mysql-client logger.go:42: 07:40:49 | gr-self-healing/13-read-from-replicas | ++ local pod=mysql-client logger.go:42: 07:40:49 | gr-self-healing/13-read-from-replicas | ++ set +o xtrace logger.go:42: 07:40:49 | gr-self-healing/13-read-from-replicas | mysql-clienttrue logger.go:42: 07:40:49 | gr-self-healing/13-read-from-replicas | ++ kubectl -n kuttl-test-emerging-raptor exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-self-healing-mysql-0.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:40:49 | gr-self-healing/13-read-from-replicas | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 07:40:49 | gr-self-healing/13-read-from-replicas | ++ sed -e 's/mysql: //' logger.go:42: 07:40:50 | gr-self-healing/13-read-from-replicas | + data='100500 logger.go:42: 07:40:50 | gr-self-healing/13-read-from-replicas | 100501 logger.go:42: 07:40:50 | gr-self-healing/13-read-from-replicas | 100502 logger.go:42: 07:40:50 | gr-self-healing/13-read-from-replicas | 100503' logger.go:42: 07:40:50 | gr-self-healing/13-read-from-replicas | + kubectl create configmap -n kuttl-test-emerging-raptor 13-read-from-replicas-0 '--from-literal=data=100500 logger.go:42: 07:40:50 | gr-self-healing/13-read-from-replicas | 100501 logger.go:42: 07:40:50 | gr-self-healing/13-read-from-replicas | 100502 logger.go:42: 07:40:50 | gr-self-healing/13-read-from-replicas | 100503' logger.go:42: 07:40:51 | gr-self-healing/13-read-from-replicas | configmap/13-read-from-replicas-0 created logger.go:42: 07:40:51 | gr-self-healing/13-read-from-replicas | + for i in 0 1 2 logger.go:42: 07:40:51 | gr-self-healing/13-read-from-replicas | +++ get_cluster_name logger.go:42: 07:40:51 | gr-self-healing/13-read-from-replicas | +++ kubectl -n kuttl-test-emerging-raptor get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 07:40:51 | gr-self-healing/13-read-from-replicas | ++ get_mysql_headless_fqdn gr-self-healing 1 logger.go:42: 07:40:51 | gr-self-healing/13-read-from-replicas | ++ local cluster=gr-self-healing logger.go:42: 07:40:51 | gr-self-healing/13-read-from-replicas | ++ local index=1 logger.go:42: 07:40:51 | gr-self-healing/13-read-from-replicas | ++ echo gr-self-healing-mysql-1.gr-self-healing-mysql logger.go:42: 07:40:51 | gr-self-healing/13-read-from-replicas | + host=gr-self-healing-mysql-1.gr-self-healing-mysql logger.go:42: 07:40:51 | gr-self-healing/13-read-from-replicas | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-self-healing-mysql-1.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:40:51 | gr-self-healing/13-read-from-replicas | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 07:40:51 | gr-self-healing/13-read-from-replicas | ++ local 'uri=-h gr-self-healing-mysql-1.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:40:51 | gr-self-healing/13-read-from-replicas | ++ local pod= logger.go:42: 07:40:51 | gr-self-healing/13-read-from-replicas | +++ get_client_pod logger.go:42: 07:40:51 | gr-self-healing/13-read-from-replicas | +++ kubectl -n kuttl-test-emerging-raptor get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 07:40:51 | gr-self-healing/13-read-from-replicas | ++ client_pod=mysql-client logger.go:42: 07:40:51 | gr-self-healing/13-read-from-replicas | ++ wait_pod mysql-client logger.go:42: 07:40:51 | gr-self-healing/13-read-from-replicas | ++ local pod=mysql-client logger.go:42: 07:40:51 | gr-self-healing/13-read-from-replicas | ++ set +o xtrace logger.go:42: 07:40:52 | gr-self-healing/13-read-from-replicas | mysql-clienttrue logger.go:42: 07:40:52 | gr-self-healing/13-read-from-replicas | ++ kubectl -n kuttl-test-emerging-raptor exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-self-healing-mysql-1.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:40:52 | gr-self-healing/13-read-from-replicas | ++ sed -e 's/mysql: //' logger.go:42: 07:40:52 | gr-self-healing/13-read-from-replicas | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 07:40:53 | gr-self-healing/13-read-from-replicas | + data='100500 logger.go:42: 07:40:53 | gr-self-healing/13-read-from-replicas | 100501 logger.go:42: 07:40:53 | gr-self-healing/13-read-from-replicas | 100502 logger.go:42: 07:40:53 | gr-self-healing/13-read-from-replicas | 100503' logger.go:42: 07:40:53 | gr-self-healing/13-read-from-replicas | + kubectl create configmap -n kuttl-test-emerging-raptor 13-read-from-replicas-1 '--from-literal=data=100500 logger.go:42: 07:40:53 | gr-self-healing/13-read-from-replicas | 100501 logger.go:42: 07:40:53 | gr-self-healing/13-read-from-replicas | 100502 logger.go:42: 07:40:53 | gr-self-healing/13-read-from-replicas | 100503' logger.go:42: 07:40:53 | gr-self-healing/13-read-from-replicas | configmap/13-read-from-replicas-1 created logger.go:42: 07:40:53 | gr-self-healing/13-read-from-replicas | + for i in 0 1 2 logger.go:42: 07:40:53 | gr-self-healing/13-read-from-replicas | +++ get_cluster_name logger.go:42: 07:40:53 | gr-self-healing/13-read-from-replicas | +++ kubectl -n kuttl-test-emerging-raptor get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 07:40:54 | gr-self-healing/13-read-from-replicas | ++ get_mysql_headless_fqdn gr-self-healing 2 logger.go:42: 07:40:54 | gr-self-healing/13-read-from-replicas | ++ local cluster=gr-self-healing logger.go:42: 07:40:54 | gr-self-healing/13-read-from-replicas | ++ local index=2 logger.go:42: 07:40:54 | gr-self-healing/13-read-from-replicas | ++ echo gr-self-healing-mysql-2.gr-self-healing-mysql logger.go:42: 07:40:54 | gr-self-healing/13-read-from-replicas | + host=gr-self-healing-mysql-2.gr-self-healing-mysql logger.go:42: 07:40:54 | gr-self-healing/13-read-from-replicas | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-self-healing-mysql-2.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:40:54 | gr-self-healing/13-read-from-replicas | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 07:40:54 | gr-self-healing/13-read-from-replicas | ++ local 'uri=-h gr-self-healing-mysql-2.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:40:54 | gr-self-healing/13-read-from-replicas | ++ local pod= logger.go:42: 07:40:54 | gr-self-healing/13-read-from-replicas | +++ get_client_pod logger.go:42: 07:40:54 | gr-self-healing/13-read-from-replicas | +++ kubectl -n kuttl-test-emerging-raptor get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 07:40:54 | gr-self-healing/13-read-from-replicas | ++ client_pod=mysql-client logger.go:42: 07:40:54 | gr-self-healing/13-read-from-replicas | ++ wait_pod mysql-client logger.go:42: 07:40:54 | gr-self-healing/13-read-from-replicas | ++ local pod=mysql-client logger.go:42: 07:40:54 | gr-self-healing/13-read-from-replicas | ++ set +o xtrace logger.go:42: 07:40:54 | gr-self-healing/13-read-from-replicas | mysql-clienttrue logger.go:42: 07:40:54 | gr-self-healing/13-read-from-replicas | ++ kubectl -n kuttl-test-emerging-raptor exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-self-healing-mysql-2.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:40:54 | gr-self-healing/13-read-from-replicas | ++ sed -e 's/mysql: //' logger.go:42: 07:40:54 | gr-self-healing/13-read-from-replicas | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 07:40:56 | gr-self-healing/13-read-from-replicas | + data='100500 logger.go:42: 07:40:56 | gr-self-healing/13-read-from-replicas | 100501 logger.go:42: 07:40:56 | gr-self-healing/13-read-from-replicas | 100502 logger.go:42: 07:40:56 | gr-self-healing/13-read-from-replicas | 100503' logger.go:42: 07:40:56 | gr-self-healing/13-read-from-replicas | + kubectl create configmap -n kuttl-test-emerging-raptor 13-read-from-replicas-2 '--from-literal=data=100500 logger.go:42: 07:40:56 | gr-self-healing/13-read-from-replicas | 100501 logger.go:42: 07:40:56 | gr-self-healing/13-read-from-replicas | 100502 logger.go:42: 07:40:56 | gr-self-healing/13-read-from-replicas | 100503' logger.go:42: 07:40:56 | gr-self-healing/13-read-from-replicas | configmap/13-read-from-replicas-2 created logger.go:42: 07:40:58 | gr-self-healing/13-read-from-replicas | test step completed 13-read-from-replicas logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | starting test step 14-cluster-crash logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | running command: [sh -c set -o errexit set -o xtrace source ../../functions kill_pods "${NAMESPACE}" "label" "app.kubernetes.io/instance" "gr-self-healing" "cluster-crash" sleep 30 # wait for crash] logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | + source ../../functions logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ realpath ../../.. logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | ++++ pwd logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-self-healing logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | ++ test_name=gr-self-healing logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ export GIT_BRANCH=PR-523 logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ GIT_BRANCH=PR-523 logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ export VERSION=PR-523-f00253e logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ VERSION=PR-523-f00253e logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | ++++ which gdate logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | ++++ which date logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ date=/usr/bin/date logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ command -v oc logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ kubectl get nodes logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | +++ grep '^minikube' logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | + kill_pods kuttl-test-emerging-raptor label app.kubernetes.io/instance gr-self-healing cluster-crash logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | + local ns=kuttl-test-emerging-raptor logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | + local selector=label logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | + local pod_label=app.kubernetes.io/instance logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | + local label_value=gr-self-healing logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | + local chaos_suffix=cluster-crash logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | + '[' label == pod ']' logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | + '[' label == label ']' logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | + kubectl apply --namespace kuttl-test-emerging-raptor -f - logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | + yq eval ' logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | .metadata.name = "chaos-kill-label-cluster-crash" | logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | .spec.mode = "all" | logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | del(.spec.selector.pods) | logger.go:42: 07:40:58 | gr-self-healing/14-cluster-crash | .spec.selector.labelSelectors."app.kubernetes.io/instance" = "gr-self-healing"' /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf/chaos-pod-kill.yml logger.go:42: 07:40:59 | gr-self-healing/14-cluster-crash | podchaos.chaos-mesh.org/chaos-kill-label-cluster-crash created logger.go:42: 07:40:59 | gr-self-healing/14-cluster-crash | + sleep 5 logger.go:42: 07:41:04 | gr-self-healing/14-cluster-crash | + sleep 30 logger.go:42: 07:47:09 | gr-self-healing/14-cluster-crash | test step completed 14-cluster-crash logger.go:42: 07:47:09 | gr-self-healing/15-write-data | starting test step 15-write-data logger.go:42: 07:47:09 | gr-self-healing/15-write-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "INSERT myDB.myTable (id) VALUES (100504)" \ "-h $(get_mysql_router_service $(get_cluster_name)) -P 6446 -uroot -proot_password"] logger.go:42: 07:47:09 | gr-self-healing/15-write-data | + source ../../functions logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ realpath ../../.. logger.go:42: 07:47:09 | gr-self-healing/15-write-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:47:09 | gr-self-healing/15-write-data | ++++ pwd logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-self-healing logger.go:42: 07:47:09 | gr-self-healing/15-write-data | ++ test_name=gr-self-healing logger.go:42: 07:47:09 | gr-self-healing/15-write-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:47:09 | gr-self-healing/15-write-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ export GIT_BRANCH=PR-523 logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ GIT_BRANCH=PR-523 logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ export VERSION=PR-523-f00253e logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ VERSION=PR-523-f00253e logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:47:09 | gr-self-healing/15-write-data | ++++ which gdate logger.go:42: 07:47:09 | gr-self-healing/15-write-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 07:47:09 | gr-self-healing/15-write-data | ++++ which date logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ date=/usr/bin/date logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ command -v oc logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ kubectl get nodes logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ grep '^minikube' logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ get_cluster_name logger.go:42: 07:47:09 | gr-self-healing/15-write-data | +++ kubectl -n kuttl-test-emerging-raptor get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 07:47:10 | gr-self-healing/15-write-data | ++ get_mysql_router_service gr-self-healing logger.go:42: 07:47:10 | gr-self-healing/15-write-data | ++ local cluster=gr-self-healing logger.go:42: 07:47:10 | gr-self-healing/15-write-data | ++ echo gr-self-healing-router logger.go:42: 07:47:10 | gr-self-healing/15-write-data | + run_mysql 'INSERT myDB.myTable (id) VALUES (100504)' '-h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:47:10 | gr-self-healing/15-write-data | + local 'command=INSERT myDB.myTable (id) VALUES (100504)' logger.go:42: 07:47:10 | gr-self-healing/15-write-data | + local 'uri=-h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:47:10 | gr-self-healing/15-write-data | + local pod= logger.go:42: 07:47:10 | gr-self-healing/15-write-data | ++ get_client_pod logger.go:42: 07:47:10 | gr-self-healing/15-write-data | ++ kubectl -n kuttl-test-emerging-raptor get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 07:47:10 | gr-self-healing/15-write-data | + client_pod=mysql-client logger.go:42: 07:47:10 | gr-self-healing/15-write-data | + wait_pod mysql-client logger.go:42: 07:47:10 | gr-self-healing/15-write-data | + local pod=mysql-client logger.go:42: 07:47:10 | gr-self-healing/15-write-data | + set +o xtrace logger.go:42: 07:47:10 | gr-self-healing/15-write-data | mysql-clienttrue logger.go:42: 07:47:10 | gr-self-healing/15-write-data | + kubectl -n kuttl-test-emerging-raptor exec mysql-client -- bash -c 'printf '\''%s\n'\'' "INSERT myDB.myTable (id) VALUES (100504)" | mysql -sN -h gr-self-healing-router -P 6446 -uroot -proot_password' logger.go:42: 07:47:10 | gr-self-healing/15-write-data | + sed -e 's/mysql: //' logger.go:42: 07:47:10 | gr-self-healing/15-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 07:47:12 | gr-self-healing/15-write-data | + : logger.go:42: 07:47:12 | gr-self-healing/15-write-data | test step completed 15-write-data logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | starting test step 16-read-from-replicas logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | running command: [sh -c set -o errexit set -o xtrace source ../../functions for i in 0 1 2; do host=$(get_mysql_headless_fqdn $(get_cluster_name) $i) data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${host} -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 16-read-from-replicas-${i} --from-literal=data="${data}" done] logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | + source ../../functions logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ realpath ../../.. logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | ++++ pwd logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-self-healing logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | ++ test_name=gr-self-healing logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ export GIT_BRANCH=PR-523 logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ GIT_BRANCH=PR-523 logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ export VERSION=PR-523-f00253e logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ VERSION=PR-523-f00253e logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | ++++ which gdate logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | ++++ which date logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ date=/usr/bin/date logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ command -v oc logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ kubectl get nodes logger.go:42: 07:47:12 | gr-self-healing/16-read-from-replicas | +++ grep '^minikube' logger.go:42: 07:47:13 | gr-self-healing/16-read-from-replicas | + for i in 0 1 2 logger.go:42: 07:47:13 | gr-self-healing/16-read-from-replicas | +++ get_cluster_name logger.go:42: 07:47:13 | gr-self-healing/16-read-from-replicas | +++ kubectl -n kuttl-test-emerging-raptor get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 07:47:13 | gr-self-healing/16-read-from-replicas | ++ get_mysql_headless_fqdn gr-self-healing 0 logger.go:42: 07:47:13 | gr-self-healing/16-read-from-replicas | ++ local cluster=gr-self-healing logger.go:42: 07:47:13 | gr-self-healing/16-read-from-replicas | ++ local index=0 logger.go:42: 07:47:13 | gr-self-healing/16-read-from-replicas | ++ echo gr-self-healing-mysql-0.gr-self-healing-mysql logger.go:42: 07:47:13 | gr-self-healing/16-read-from-replicas | + host=gr-self-healing-mysql-0.gr-self-healing-mysql logger.go:42: 07:47:13 | gr-self-healing/16-read-from-replicas | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-self-healing-mysql-0.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:47:13 | gr-self-healing/16-read-from-replicas | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 07:47:13 | gr-self-healing/16-read-from-replicas | ++ local 'uri=-h gr-self-healing-mysql-0.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:47:13 | gr-self-healing/16-read-from-replicas | ++ local pod= logger.go:42: 07:47:13 | gr-self-healing/16-read-from-replicas | +++ get_client_pod logger.go:42: 07:47:13 | gr-self-healing/16-read-from-replicas | +++ kubectl -n kuttl-test-emerging-raptor get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 07:47:14 | gr-self-healing/16-read-from-replicas | ++ client_pod=mysql-client logger.go:42: 07:47:14 | gr-self-healing/16-read-from-replicas | ++ wait_pod mysql-client logger.go:42: 07:47:14 | gr-self-healing/16-read-from-replicas | ++ local pod=mysql-client logger.go:42: 07:47:14 | gr-self-healing/16-read-from-replicas | ++ set +o xtrace logger.go:42: 07:47:14 | gr-self-healing/16-read-from-replicas | mysql-clienttrue logger.go:42: 07:47:14 | gr-self-healing/16-read-from-replicas | ++ kubectl -n kuttl-test-emerging-raptor exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-self-healing-mysql-0.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:47:14 | gr-self-healing/16-read-from-replicas | ++ sed -e 's/mysql: //' logger.go:42: 07:47:14 | gr-self-healing/16-read-from-replicas | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 07:47:15 | gr-self-healing/16-read-from-replicas | + data='100500 logger.go:42: 07:47:15 | gr-self-healing/16-read-from-replicas | 100501 logger.go:42: 07:47:15 | gr-self-healing/16-read-from-replicas | 100502 logger.go:42: 07:47:15 | gr-self-healing/16-read-from-replicas | 100503 logger.go:42: 07:47:15 | gr-self-healing/16-read-from-replicas | 100504' logger.go:42: 07:47:15 | gr-self-healing/16-read-from-replicas | + kubectl create configmap -n kuttl-test-emerging-raptor 16-read-from-replicas-0 '--from-literal=data=100500 logger.go:42: 07:47:15 | gr-self-healing/16-read-from-replicas | 100501 logger.go:42: 07:47:15 | gr-self-healing/16-read-from-replicas | 100502 logger.go:42: 07:47:15 | gr-self-healing/16-read-from-replicas | 100503 logger.go:42: 07:47:15 | gr-self-healing/16-read-from-replicas | 100504' logger.go:42: 07:47:16 | gr-self-healing/16-read-from-replicas | configmap/16-read-from-replicas-0 created logger.go:42: 07:47:16 | gr-self-healing/16-read-from-replicas | + for i in 0 1 2 logger.go:42: 07:47:16 | gr-self-healing/16-read-from-replicas | +++ get_cluster_name logger.go:42: 07:47:16 | gr-self-healing/16-read-from-replicas | +++ kubectl -n kuttl-test-emerging-raptor get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 07:47:16 | gr-self-healing/16-read-from-replicas | ++ get_mysql_headless_fqdn gr-self-healing 1 logger.go:42: 07:47:16 | gr-self-healing/16-read-from-replicas | ++ local cluster=gr-self-healing logger.go:42: 07:47:16 | gr-self-healing/16-read-from-replicas | ++ local index=1 logger.go:42: 07:47:16 | gr-self-healing/16-read-from-replicas | ++ echo gr-self-healing-mysql-1.gr-self-healing-mysql logger.go:42: 07:47:16 | gr-self-healing/16-read-from-replicas | + host=gr-self-healing-mysql-1.gr-self-healing-mysql logger.go:42: 07:47:16 | gr-self-healing/16-read-from-replicas | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-self-healing-mysql-1.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:47:16 | gr-self-healing/16-read-from-replicas | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 07:47:16 | gr-self-healing/16-read-from-replicas | ++ local 'uri=-h gr-self-healing-mysql-1.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:47:16 | gr-self-healing/16-read-from-replicas | ++ local pod= logger.go:42: 07:47:16 | gr-self-healing/16-read-from-replicas | +++ get_client_pod logger.go:42: 07:47:16 | gr-self-healing/16-read-from-replicas | +++ kubectl -n kuttl-test-emerging-raptor get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 07:47:16 | gr-self-healing/16-read-from-replicas | ++ client_pod=mysql-client logger.go:42: 07:47:16 | gr-self-healing/16-read-from-replicas | ++ wait_pod mysql-client logger.go:42: 07:47:16 | gr-self-healing/16-read-from-replicas | ++ local pod=mysql-client logger.go:42: 07:47:16 | gr-self-healing/16-read-from-replicas | ++ set +o xtrace logger.go:42: 07:47:17 | gr-self-healing/16-read-from-replicas | mysql-clienttrue logger.go:42: 07:47:17 | gr-self-healing/16-read-from-replicas | ++ kubectl -n kuttl-test-emerging-raptor exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-self-healing-mysql-1.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:47:17 | gr-self-healing/16-read-from-replicas | ++ sed -e 's/mysql: //' logger.go:42: 07:47:17 | gr-self-healing/16-read-from-replicas | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 07:47:18 | gr-self-healing/16-read-from-replicas | + data='100500 logger.go:42: 07:47:18 | gr-self-healing/16-read-from-replicas | 100501 logger.go:42: 07:47:18 | gr-self-healing/16-read-from-replicas | 100502 logger.go:42: 07:47:18 | gr-self-healing/16-read-from-replicas | 100503 logger.go:42: 07:47:18 | gr-self-healing/16-read-from-replicas | 100504' logger.go:42: 07:47:18 | gr-self-healing/16-read-from-replicas | + kubectl create configmap -n kuttl-test-emerging-raptor 16-read-from-replicas-1 '--from-literal=data=100500 logger.go:42: 07:47:18 | gr-self-healing/16-read-from-replicas | 100501 logger.go:42: 07:47:18 | gr-self-healing/16-read-from-replicas | 100502 logger.go:42: 07:47:18 | gr-self-healing/16-read-from-replicas | 100503 logger.go:42: 07:47:18 | gr-self-healing/16-read-from-replicas | 100504' logger.go:42: 07:47:18 | gr-self-healing/16-read-from-replicas | configmap/16-read-from-replicas-1 created logger.go:42: 07:47:18 | gr-self-healing/16-read-from-replicas | + for i in 0 1 2 logger.go:42: 07:47:18 | gr-self-healing/16-read-from-replicas | +++ get_cluster_name logger.go:42: 07:47:18 | gr-self-healing/16-read-from-replicas | +++ kubectl -n kuttl-test-emerging-raptor get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 07:47:19 | gr-self-healing/16-read-from-replicas | ++ get_mysql_headless_fqdn gr-self-healing 2 logger.go:42: 07:47:19 | gr-self-healing/16-read-from-replicas | ++ local cluster=gr-self-healing logger.go:42: 07:47:19 | gr-self-healing/16-read-from-replicas | ++ local index=2 logger.go:42: 07:47:19 | gr-self-healing/16-read-from-replicas | ++ echo gr-self-healing-mysql-2.gr-self-healing-mysql logger.go:42: 07:47:19 | gr-self-healing/16-read-from-replicas | + host=gr-self-healing-mysql-2.gr-self-healing-mysql logger.go:42: 07:47:19 | gr-self-healing/16-read-from-replicas | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h gr-self-healing-mysql-2.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:47:19 | gr-self-healing/16-read-from-replicas | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 07:47:19 | gr-self-healing/16-read-from-replicas | ++ local 'uri=-h gr-self-healing-mysql-2.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:47:19 | gr-self-healing/16-read-from-replicas | ++ local pod= logger.go:42: 07:47:19 | gr-self-healing/16-read-from-replicas | +++ get_client_pod logger.go:42: 07:47:19 | gr-self-healing/16-read-from-replicas | +++ kubectl -n kuttl-test-emerging-raptor get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 07:47:19 | gr-self-healing/16-read-from-replicas | ++ client_pod=mysql-client logger.go:42: 07:47:19 | gr-self-healing/16-read-from-replicas | ++ wait_pod mysql-client logger.go:42: 07:47:19 | gr-self-healing/16-read-from-replicas | ++ local pod=mysql-client logger.go:42: 07:47:19 | gr-self-healing/16-read-from-replicas | ++ set +o xtrace logger.go:42: 07:47:19 | gr-self-healing/16-read-from-replicas | mysql-clienttrue logger.go:42: 07:47:19 | gr-self-healing/16-read-from-replicas | ++ kubectl -n kuttl-test-emerging-raptor exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h gr-self-healing-mysql-2.gr-self-healing-mysql -uroot -proot_password' logger.go:42: 07:47:19 | gr-self-healing/16-read-from-replicas | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 07:47:19 | gr-self-healing/16-read-from-replicas | ++ sed -e 's/mysql: //' logger.go:42: 07:47:21 | gr-self-healing/16-read-from-replicas | + data='100500 logger.go:42: 07:47:21 | gr-self-healing/16-read-from-replicas | 100501 logger.go:42: 07:47:21 | gr-self-healing/16-read-from-replicas | 100502 logger.go:42: 07:47:21 | gr-self-healing/16-read-from-replicas | 100503 logger.go:42: 07:47:21 | gr-self-healing/16-read-from-replicas | 100504' logger.go:42: 07:47:21 | gr-self-healing/16-read-from-replicas | + kubectl create configmap -n kuttl-test-emerging-raptor 16-read-from-replicas-2 '--from-literal=data=100500 logger.go:42: 07:47:21 | gr-self-healing/16-read-from-replicas | 100501 logger.go:42: 07:47:21 | gr-self-healing/16-read-from-replicas | 100502 logger.go:42: 07:47:21 | gr-self-healing/16-read-from-replicas | 100503 logger.go:42: 07:47:21 | gr-self-healing/16-read-from-replicas | 100504' logger.go:42: 07:47:21 | gr-self-healing/16-read-from-replicas | configmap/16-read-from-replicas-2 created logger.go:42: 07:47:23 | gr-self-healing/16-read-from-replicas | test step completed 16-read-from-replicas logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | starting test step 17-destroy-chaos-mesh logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | running command: [sh -c set -o errexit set -o xtrace source ../../functions destroy_chaos_mesh] logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | + source ../../functions logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ realpath ../../.. logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | ++++ pwd logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/tests/gr-self-healing logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | ++ test_name=gr-self-healing logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/vars.sh logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-523 logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/deploy logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-523/e2e-tests/conf logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ export TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ TEMP_DIR=/tmp/kuttl/ps/gr-self-healing logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ export GIT_BRANCH=PR-523 logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ GIT_BRANCH=PR-523 logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ export VERSION=PR-523-f00253e logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ VERSION=PR-523-f00253e logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-523-f00253e logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | ++++ which gdate logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-523/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | ++++ which date logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ date=/usr/bin/date logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ command -v oc logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ kubectl get nodes logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | +++ grep '^minikube' logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | + destroy_chaos_mesh logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | ++ helm list --all-namespaces --filter chaos-mesh logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | ++ tail -n1 logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | ++ sed s/NAMESPACE// logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | ++ awk '-F ' '{print $2}' logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 07:47:23 | gr-self-healing/17-destroy-chaos-mesh | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 07:47:24 | gr-self-healing/17-destroy-chaos-mesh | + local chaos_mesh_ns=kuttl-test-emerging-raptor logger.go:42: 07:47:24 | gr-self-healing/17-destroy-chaos-mesh | ++ kubectl api-resources logger.go:42: 07:47:24 | gr-self-healing/17-destroy-chaos-mesh | ++ grep chaos-mesh logger.go:42: 07:47:24 | gr-self-healing/17-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 07:47:25 | gr-self-healing/17-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 07:47:25 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete awschaos --all --all-namespaces logger.go:42: 07:47:25 | gr-self-healing/17-destroy-chaos-mesh | No resources found logger.go:42: 07:47:25 | gr-self-healing/17-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 07:47:25 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete azurechaos --all --all-namespaces logger.go:42: 07:47:26 | gr-self-healing/17-destroy-chaos-mesh | No resources found logger.go:42: 07:47:26 | gr-self-healing/17-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 07:47:26 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete blockchaos --all --all-namespaces logger.go:42: 07:47:26 | gr-self-healing/17-destroy-chaos-mesh | No resources found logger.go:42: 07:47:26 | gr-self-healing/17-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 07:47:26 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete dnschaos --all --all-namespaces logger.go:42: 07:47:27 | gr-self-healing/17-destroy-chaos-mesh | No resources found logger.go:42: 07:47:27 | gr-self-healing/17-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 07:47:27 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete gcpchaos --all --all-namespaces logger.go:42: 07:47:27 | gr-self-healing/17-destroy-chaos-mesh | No resources found logger.go:42: 07:47:27 | gr-self-healing/17-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 07:47:27 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete httpchaos --all --all-namespaces logger.go:42: 07:47:27 | gr-self-healing/17-destroy-chaos-mesh | No resources found logger.go:42: 07:47:27 | gr-self-healing/17-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 07:47:27 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete iochaos --all --all-namespaces logger.go:42: 07:47:28 | gr-self-healing/17-destroy-chaos-mesh | No resources found logger.go:42: 07:47:28 | gr-self-healing/17-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 07:47:28 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete jvmchaos --all --all-namespaces logger.go:42: 07:47:28 | gr-self-healing/17-destroy-chaos-mesh | No resources found logger.go:42: 07:47:28 | gr-self-healing/17-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 07:47:28 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete kernelchaos --all --all-namespaces logger.go:42: 07:47:28 | gr-self-healing/17-destroy-chaos-mesh | No resources found logger.go:42: 07:47:28 | gr-self-healing/17-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 07:47:28 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete networkchaos --all --all-namespaces logger.go:42: 07:47:29 | gr-self-healing/17-destroy-chaos-mesh | networkchaos.chaos-mesh.org "chaos-pod-network-loss-primary" deleted logger.go:42: 07:47:29 | gr-self-healing/17-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 07:47:29 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete physicalmachinechaos --all --all-namespaces logger.go:42: 07:47:29 | gr-self-healing/17-destroy-chaos-mesh | No resources found logger.go:42: 07:47:29 | gr-self-healing/17-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 07:47:29 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete physicalmachines --all --all-namespaces logger.go:42: 07:47:30 | gr-self-healing/17-destroy-chaos-mesh | No resources found logger.go:42: 07:47:30 | gr-self-healing/17-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 07:47:30 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete podchaos --all --all-namespaces logger.go:42: 07:47:30 | gr-self-healing/17-destroy-chaos-mesh | podchaos.chaos-mesh.org "chaos-kill-label-cluster-crash" deleted logger.go:42: 07:47:30 | gr-self-healing/17-destroy-chaos-mesh | podchaos.chaos-mesh.org "chaos-pod-failure-primary" deleted logger.go:42: 07:47:30 | gr-self-healing/17-destroy-chaos-mesh | podchaos.chaos-mesh.org "chaos-pod-kill-primary" deleted logger.go:42: 07:47:31 | gr-self-healing/17-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 07:47:31 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete podhttpchaos --all --all-namespaces logger.go:42: 07:47:31 | gr-self-healing/17-destroy-chaos-mesh | No resources found logger.go:42: 07:47:31 | gr-self-healing/17-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 07:47:31 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete podiochaos --all --all-namespaces logger.go:42: 07:47:31 | gr-self-healing/17-destroy-chaos-mesh | No resources found logger.go:42: 07:47:31 | gr-self-healing/17-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 07:47:31 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete podnetworkchaos --all --all-namespaces logger.go:42: 07:47:32 | gr-self-healing/17-destroy-chaos-mesh | No resources found logger.go:42: 07:47:32 | gr-self-healing/17-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 07:47:32 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete remoteclusters --all --all-namespaces logger.go:42: 07:47:32 | gr-self-healing/17-destroy-chaos-mesh | No resources found logger.go:42: 07:47:32 | gr-self-healing/17-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 07:47:32 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete schedules --all --all-namespaces logger.go:42: 07:47:33 | gr-self-healing/17-destroy-chaos-mesh | No resources found logger.go:42: 07:47:33 | gr-self-healing/17-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 07:47:33 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete statuschecks --all --all-namespaces logger.go:42: 07:47:33 | gr-self-healing/17-destroy-chaos-mesh | No resources found logger.go:42: 07:47:33 | gr-self-healing/17-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 07:47:33 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete stresschaos --all --all-namespaces logger.go:42: 07:47:33 | gr-self-healing/17-destroy-chaos-mesh | No resources found logger.go:42: 07:47:33 | gr-self-healing/17-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 07:47:33 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete timechaos --all --all-namespaces logger.go:42: 07:47:34 | gr-self-healing/17-destroy-chaos-mesh | No resources found logger.go:42: 07:47:34 | gr-self-healing/17-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 07:47:34 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete workflownodes --all --all-namespaces logger.go:42: 07:47:34 | gr-self-healing/17-destroy-chaos-mesh | No resources found logger.go:42: 07:47:34 | gr-self-healing/17-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 07:47:34 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete workflows --all --all-namespaces logger.go:42: 07:47:34 | gr-self-healing/17-destroy-chaos-mesh | No resources found logger.go:42: 07:47:34 | gr-self-healing/17-destroy-chaos-mesh | + '[' -n kuttl-test-emerging-raptor ']' logger.go:42: 07:47:34 | gr-self-healing/17-destroy-chaos-mesh | + helm uninstall chaos-mesh --namespace kuttl-test-emerging-raptor logger.go:42: 07:47:34 | gr-self-healing/17-destroy-chaos-mesh | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 07:47:34 | gr-self-healing/17-destroy-chaos-mesh | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-523/kubeconfig logger.go:42: 07:47:39 | gr-self-healing/17-destroy-chaos-mesh | release "chaos-mesh" uninstalled logger.go:42: 07:47:39 | gr-self-healing/17-destroy-chaos-mesh | ++ kubectl get crd logger.go:42: 07:47:39 | gr-self-healing/17-destroy-chaos-mesh | ++ grep chaos-mesh.org logger.go:42: 07:47:39 | gr-self-healing/17-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 07:47:40 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete crd awschaos.chaos-mesh.org azurechaos.chaos-mesh.org blockchaos.chaos-mesh.org dnschaos.chaos-mesh.org gcpchaos.chaos-mesh.org httpchaos.chaos-mesh.org iochaos.chaos-mesh.org jvmchaos.chaos-mesh.org kernelchaos.chaos-mesh.org networkchaos.chaos-mesh.org physicalmachinechaos.chaos-mesh.org physicalmachines.chaos-mesh.org podchaos.chaos-mesh.org podhttpchaos.chaos-mesh.org podiochaos.chaos-mesh.org podnetworkchaos.chaos-mesh.org remoteclusters.chaos-mesh.org schedules.chaos-mesh.org statuschecks.chaos-mesh.org stresschaos.chaos-mesh.org timechaos.chaos-mesh.org workflownodes.chaos-mesh.org workflows.chaos-mesh.org logger.go:42: 07:47:41 | gr-self-healing/17-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "awschaos.chaos-mesh.org" deleted logger.go:42: 07:47:41 | gr-self-healing/17-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "azurechaos.chaos-mesh.org" deleted logger.go:42: 07:47:41 | gr-self-healing/17-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "blockchaos.chaos-mesh.org" deleted logger.go:42: 07:47:41 | gr-self-healing/17-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "dnschaos.chaos-mesh.org" deleted logger.go:42: 07:47:41 | gr-self-healing/17-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "gcpchaos.chaos-mesh.org" deleted logger.go:42: 07:47:41 | gr-self-healing/17-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "httpchaos.chaos-mesh.org" deleted logger.go:42: 07:47:41 | gr-self-healing/17-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "iochaos.chaos-mesh.org" deleted logger.go:42: 07:47:42 | gr-self-healing/17-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "jvmchaos.chaos-mesh.org" deleted logger.go:42: 07:47:42 | gr-self-healing/17-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "kernelchaos.chaos-mesh.org" deleted logger.go:42: 07:47:42 | gr-self-healing/17-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "networkchaos.chaos-mesh.org" deleted logger.go:42: 07:47:42 | gr-self-healing/17-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "physicalmachinechaos.chaos-mesh.org" deleted logger.go:42: 07:47:42 | gr-self-healing/17-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "physicalmachines.chaos-mesh.org" deleted logger.go:42: 07:47:43 | gr-self-healing/17-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "podchaos.chaos-mesh.org" deleted logger.go:42: 07:47:43 | gr-self-healing/17-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "podhttpchaos.chaos-mesh.org" deleted logger.go:42: 07:47:43 | gr-self-healing/17-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "podiochaos.chaos-mesh.org" deleted logger.go:42: 07:47:43 | gr-self-healing/17-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "podnetworkchaos.chaos-mesh.org" deleted logger.go:42: 07:47:43 | gr-self-healing/17-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "remoteclusters.chaos-mesh.org" deleted logger.go:42: 07:47:45 | gr-self-healing/17-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "schedules.chaos-mesh.org" deleted logger.go:42: 07:47:45 | gr-self-healing/17-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "statuschecks.chaos-mesh.org" deleted logger.go:42: 07:47:45 | gr-self-healing/17-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "stresschaos.chaos-mesh.org" deleted logger.go:42: 07:47:46 | gr-self-healing/17-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "timechaos.chaos-mesh.org" deleted logger.go:42: 07:47:48 | gr-self-healing/17-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "workflownodes.chaos-mesh.org" deleted logger.go:42: 07:47:50 | gr-self-healing/17-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "workflows.chaos-mesh.org" deleted logger.go:42: 07:47:56 | gr-self-healing/17-destroy-chaos-mesh | ++ kubectl get clusterrolebinding logger.go:42: 07:47:56 | gr-self-healing/17-destroy-chaos-mesh | ++ grep chaos-mesh logger.go:42: 07:47:56 | gr-self-healing/17-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 07:47:56 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete clusterrolebinding logger.go:42: 07:47:56 | gr-self-healing/17-destroy-chaos-mesh | error: resource(s) were provided, but no name was specified logger.go:42: 07:47:56 | gr-self-healing/17-destroy-chaos-mesh | + : logger.go:42: 07:47:56 | gr-self-healing/17-destroy-chaos-mesh | ++ kubectl get clusterrole logger.go:42: 07:47:56 | gr-self-healing/17-destroy-chaos-mesh | ++ grep chaos-mesh logger.go:42: 07:47:56 | gr-self-healing/17-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 07:47:57 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete clusterrole logger.go:42: 07:47:57 | gr-self-healing/17-destroy-chaos-mesh | error: resource(s) were provided, but no name was specified logger.go:42: 07:47:57 | gr-self-healing/17-destroy-chaos-mesh | + : logger.go:42: 07:47:57 | gr-self-healing/17-destroy-chaos-mesh | ++ kubectl get MutatingWebhookConfiguration logger.go:42: 07:47:57 | gr-self-healing/17-destroy-chaos-mesh | ++ grep chaos-mesh logger.go:42: 07:47:57 | gr-self-healing/17-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 07:47:57 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete MutatingWebhookConfiguration logger.go:42: 07:47:58 | gr-self-healing/17-destroy-chaos-mesh | error: resource(s) were provided, but no name was specified logger.go:42: 07:47:58 | gr-self-healing/17-destroy-chaos-mesh | + : logger.go:42: 07:47:58 | gr-self-healing/17-destroy-chaos-mesh | ++ kubectl get ValidatingWebhookConfiguration logger.go:42: 07:47:58 | gr-self-healing/17-destroy-chaos-mesh | ++ grep chaos-mesh logger.go:42: 07:47:58 | gr-self-healing/17-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 07:47:58 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete ValidatingWebhookConfiguration logger.go:42: 07:47:58 | gr-self-healing/17-destroy-chaos-mesh | error: resource(s) were provided, but no name was specified logger.go:42: 07:47:58 | gr-self-healing/17-destroy-chaos-mesh | + : logger.go:42: 07:47:58 | gr-self-healing/17-destroy-chaos-mesh | ++ kubectl get ValidatingWebhookConfiguration logger.go:42: 07:47:58 | gr-self-healing/17-destroy-chaos-mesh | ++ grep validate-auth logger.go:42: 07:47:58 | gr-self-healing/17-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 07:47:58 | gr-self-healing/17-destroy-chaos-mesh | + timeout 30 kubectl delete ValidatingWebhookConfiguration logger.go:42: 07:47:58 | gr-self-healing/17-destroy-chaos-mesh | error: resource(s) were provided, but no name was specified logger.go:42: 07:47:58 | gr-self-healing/17-destroy-chaos-mesh | + : logger.go:42: 07:47:59 | gr-self-healing/17-destroy-chaos-mesh | test step completed 17-destroy-chaos-mesh logger.go:42: 07:47:59 | gr-self-healing/99-drop-finalizer | starting test step 99-drop-finalizer logger.go:42: 07:48:00 | gr-self-healing/99-drop-finalizer | PerconaServerMySQL:kuttl-test-emerging-raptor/gr-self-healing updated logger.go:42: 07:48:00 | gr-self-healing/99-drop-finalizer | test step completed 99-drop-finalizer logger.go:42: 07:48:00 | gr-self-healing | gr-self-healing events from ns kuttl-test-emerging-raptor: logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:28:04 +0000 UTC Normal Pod percona-server-mysql-operator-b4c599bbb-8wlqs Scheduled Successfully assigned kuttl-test-emerging-raptor/percona-server-mysql-operator-b4c599bbb-8wlqs to gke-jen-ps-523-f00253e-7-default-pool-bcfb7a6f-9sgd default-scheduler logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:28:04 +0000 UTC Normal ReplicaSet.apps percona-server-mysql-operator-b4c599bbb SuccessfulCreate Created pod: percona-server-mysql-operator-b4c599bbb-8wlqs replicaset-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:28:04 +0000 UTC Normal Deployment.apps percona-server-mysql-operator ScalingReplicaSet Scaled up replica set percona-server-mysql-operator-b4c599bbb to 1 deployment-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:28:06 +0000 UTC Normal Pod percona-server-mysql-operator-b4c599bbb-8wlqs.spec.containers{manager} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:28:08 +0000 UTC Normal Pod mysql-client Scheduled Successfully assigned kuttl-test-emerging-raptor/mysql-client to gke-jen-ps-523-f00253e-7-default-pool-bcfb7a6f-9sgd default-scheduler logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:28:09 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Pulling Pulling image "percona/percona-server:8.0.33" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:28:25 +0000 UTC Normal Pod percona-server-mysql-operator-b4c599bbb-8wlqs.spec.containers{manager} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 19.267004245s (19.267074888s including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:28:25 +0000 UTC Normal Pod percona-server-mysql-operator-b4c599bbb-8wlqs.spec.containers{manager} Created Created container manager kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:28:25 +0000 UTC Normal Pod percona-server-mysql-operator-b4c599bbb-8wlqs.spec.containers{manager} Started Started container manager kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:28:26 +0000 UTC Normal Lease.coordination.k8s.io 08db2feb.percona.com LeaderElection percona-server-mysql-operator-b4c599bbb-8wlqs_a6e39204-57d2-46af-b2ed-ee531aff8fb2 became leader percona-server-mysql-operator-b4c599bbb-8wlqs_a6e39204-57d2-46af-b2ed-ee531aff8fb2 logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:28:51 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Pulled Successfully pulled image "percona/percona-server:8.0.33" in 25.266715364s (41.366685191s including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:28:51 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Created Created container mysql-client kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:28:51 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Started Started container mysql-client kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:08 +0000 UTC Normal Pod chaos-controller-manager-79c896cbb8-5l4x6 Scheduled Successfully assigned kuttl-test-emerging-raptor/chaos-controller-manager-79c896cbb8-5l4x6 to gke-jen-ps-523-f00253e-7-default-pool-bcfb7a6f-7ntv default-scheduler logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:08 +0000 UTC Normal Pod chaos-controller-manager-79c896cbb8-knt2h Scheduled Successfully assigned kuttl-test-emerging-raptor/chaos-controller-manager-79c896cbb8-knt2h to gke-jen-ps-523-f00253e-7-default-pool-bcfb7a6f-tk4k default-scheduler logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:08 +0000 UTC Normal Pod chaos-controller-manager-79c896cbb8-tg794 Scheduled Successfully assigned kuttl-test-emerging-raptor/chaos-controller-manager-79c896cbb8-tg794 to gke-jen-ps-523-f00253e-7-default-pool-bcfb7a6f-9sgd default-scheduler logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:08 +0000 UTC Normal ReplicaSet.apps chaos-controller-manager-79c896cbb8 SuccessfulCreate Created pod: chaos-controller-manager-79c896cbb8-tg794 replicaset-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:08 +0000 UTC Normal ReplicaSet.apps chaos-controller-manager-79c896cbb8 SuccessfulCreate Created pod: chaos-controller-manager-79c896cbb8-knt2h replicaset-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:08 +0000 UTC Normal ReplicaSet.apps chaos-controller-manager-79c896cbb8 SuccessfulCreate Created pod: chaos-controller-manager-79c896cbb8-5l4x6 replicaset-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:08 +0000 UTC Normal Deployment.apps chaos-controller-manager ScalingReplicaSet Scaled up replica set chaos-controller-manager-79c896cbb8 to 3 deployment-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:08 +0000 UTC Normal Pod chaos-daemon-5h7zt Scheduled Successfully assigned kuttl-test-emerging-raptor/chaos-daemon-5h7zt to gke-jen-ps-523-f00253e-7-default-pool-bcfb7a6f-tk4k default-scheduler logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:08 +0000 UTC Normal Pod chaos-daemon-5h7zt.spec.containers{chaos-daemon} Pulling Pulling image "ghcr.io/chaos-mesh/chaos-daemon:v2.5.1" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:08 +0000 UTC Normal Pod chaos-daemon-6rlxh Scheduled Successfully assigned kuttl-test-emerging-raptor/chaos-daemon-6rlxh to gke-jen-ps-523-f00253e-7-default-pool-bcfb7a6f-7ntv default-scheduler logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:08 +0000 UTC Normal Pod chaos-daemon-6rlxh.spec.containers{chaos-daemon} Pulling Pulling image "ghcr.io/chaos-mesh/chaos-daemon:v2.5.1" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:08 +0000 UTC Normal Pod chaos-daemon-xpr87 Scheduled Successfully assigned kuttl-test-emerging-raptor/chaos-daemon-xpr87 to gke-jen-ps-523-f00253e-7-default-pool-bcfb7a6f-9sgd default-scheduler logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:08 +0000 UTC Normal DaemonSet.apps chaos-daemon SuccessfulCreate Created pod: chaos-daemon-6rlxh daemonset-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:08 +0000 UTC Normal DaemonSet.apps chaos-daemon SuccessfulCreate Created pod: chaos-daemon-5h7zt daemonset-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:08 +0000 UTC Normal DaemonSet.apps chaos-daemon SuccessfulCreate Created pod: chaos-daemon-xpr87 daemonset-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:09 +0000 UTC Normal Pod chaos-controller-manager-79c896cbb8-5l4x6.spec.containers{chaos-mesh} Pulling Pulling image "ghcr.io/chaos-mesh/chaos-mesh:v2.5.1" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:09 +0000 UTC Warning Pod chaos-controller-manager-79c896cbb8-tg794 FailedMount MountVolume.SetUp failed for volume "chaosd-client-cert" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:09 +0000 UTC Warning Pod chaos-controller-manager-79c896cbb8-tg794 FailedMount MountVolume.SetUp failed for volume "webhook-certs" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:09 +0000 UTC Normal Pod chaos-daemon-xpr87.spec.containers{chaos-daemon} Pulling Pulling image "ghcr.io/chaos-mesh/chaos-daemon:v2.5.1" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:10 +0000 UTC Warning Pod chaos-controller-manager-79c896cbb8-knt2h FailedMount MountVolume.SetUp failed for volume "chaos-daemon-client-cert" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:10 +0000 UTC Warning Pod chaos-controller-manager-79c896cbb8-knt2h FailedMount MountVolume.SetUp failed for volume "chaosd-client-cert" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:11 +0000 UTC Normal Pod chaos-controller-manager-79c896cbb8-knt2h.spec.containers{chaos-mesh} Pulling Pulling image "ghcr.io/chaos-mesh/chaos-mesh:v2.5.1" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:11 +0000 UTC Normal Pod chaos-controller-manager-79c896cbb8-tg794.spec.containers{chaos-mesh} Pulling Pulling image "ghcr.io/chaos-mesh/chaos-mesh:v2.5.1" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:19 +0000 UTC Normal Pod chaos-daemon-5h7zt.spec.containers{chaos-daemon} Pulled Successfully pulled image "ghcr.io/chaos-mesh/chaos-daemon:v2.5.1" in 10.857595392s (10.857728402s including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:19 +0000 UTC Normal Pod chaos-daemon-5h7zt.spec.containers{chaos-daemon} Created Created container chaos-daemon kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:19 +0000 UTC Normal Pod chaos-daemon-5h7zt.spec.containers{chaos-daemon} Started Started container chaos-daemon kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:19 +0000 UTC Normal Pod chaos-daemon-6rlxh.spec.containers{chaos-daemon} Pulled Successfully pulled image "ghcr.io/chaos-mesh/chaos-daemon:v2.5.1" in 10.41769293s (10.417794871s including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:19 +0000 UTC Normal Pod chaos-daemon-6rlxh.spec.containers{chaos-daemon} Created Created container chaos-daemon kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:19 +0000 UTC Normal Pod chaos-daemon-6rlxh.spec.containers{chaos-daemon} Started Started container chaos-daemon kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:21 +0000 UTC Normal Pod chaos-controller-manager-79c896cbb8-5l4x6.spec.containers{chaos-mesh} Pulled Successfully pulled image "ghcr.io/chaos-mesh/chaos-mesh:v2.5.1" in 1.961055317s (11.603380426s including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:21 +0000 UTC Normal Pod chaos-controller-manager-79c896cbb8-5l4x6.spec.containers{chaos-mesh} Created Created container chaos-mesh kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:21 +0000 UTC Normal Pod chaos-controller-manager-79c896cbb8-5l4x6.spec.containers{chaos-mesh} Started Started container chaos-mesh kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:21 +0000 UTC Normal Pod chaos-controller-manager-79c896cbb8-knt2h.spec.containers{chaos-mesh} Pulled Successfully pulled image "ghcr.io/chaos-mesh/chaos-mesh:v2.5.1" in 1.927957372s (10.031414221s including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:21 +0000 UTC Normal Pod chaos-controller-manager-79c896cbb8-knt2h.spec.containers{chaos-mesh} Created Created container chaos-mesh kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:21 +0000 UTC Normal Pod chaos-controller-manager-79c896cbb8-knt2h.spec.containers{chaos-mesh} Started Started container chaos-mesh kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:22 +0000 UTC Warning Endpoints chaos-mesh-controller-manager FailedToUpdateEndpoint Failed to update endpoint kuttl-test-emerging-raptor/chaos-mesh-controller-manager: Operation cannot be fulfilled on endpoints "chaos-mesh-controller-manager": the object has been modified; please apply your changes to the latest version and try again endpoint-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:22 +0000 UTC Normal ConfigMap chaos-mesh LeaderElection chaos-controller-manager-79c896cbb8-5l4x6_b1898b1d-b21d-4076-9ecb-6a71584c79a9 became leader logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:22 +0000 UTC Normal Lease.coordination.k8s.io chaos-mesh LeaderElection chaos-controller-manager-79c896cbb8-5l4x6_b1898b1d-b21d-4076-9ecb-6a71584c79a9 became leader logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:23 +0000 UTC Normal Pod chaos-daemon-xpr87.spec.containers{chaos-daemon} Pulled Successfully pulled image "ghcr.io/chaos-mesh/chaos-daemon:v2.5.1" in 13.80760061s (13.807701638s including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:23 +0000 UTC Normal Pod chaos-daemon-xpr87.spec.containers{chaos-daemon} Created Created container chaos-daemon kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:23 +0000 UTC Normal Pod chaos-daemon-xpr87.spec.containers{chaos-daemon} Started Started container chaos-daemon kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:25 +0000 UTC Normal Pod chaos-controller-manager-79c896cbb8-tg794.spec.containers{chaos-mesh} Pulled Successfully pulled image "ghcr.io/chaos-mesh/chaos-mesh:v2.5.1" in 1.87868997s (14.16414057s including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:25 +0000 UTC Normal Pod chaos-controller-manager-79c896cbb8-tg794.spec.containers{chaos-mesh} Created Created container chaos-mesh kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:25 +0000 UTC Normal Pod chaos-controller-manager-79c896cbb8-tg794.spec.containers{chaos-mesh} Started Started container chaos-mesh kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:32 +0000 UTC Normal PersistentVolumeClaim datadir-gr-self-healing-mysql-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:32 +0000 UTC Normal PersistentVolumeClaim datadir-gr-self-healing-mysql-0 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:32 +0000 UTC Normal StatefulSet.apps gr-self-healing-mysql SuccessfulCreate create Claim datadir-gr-self-healing-mysql-0 Pod gr-self-healing-mysql-0 in StatefulSet gr-self-healing-mysql success statefulset-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:32 +0000 UTC Normal StatefulSet.apps gr-self-healing-mysql SuccessfulCreate create Pod gr-self-healing-mysql-0 in StatefulSet gr-self-healing-mysql successful statefulset-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:33 +0000 UTC Normal PersistentVolumeClaim datadir-gr-self-healing-mysql-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-emerging-raptor/datadir-gr-self-healing-mysql-0" pd.csi.storage.gke.io_gke-2b3665bdb5444e358fad-0db2-5fc3-vm_91cf7654-81a4-401d-a58f-1e36f2a9506e logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:37 +0000 UTC Normal PersistentVolumeClaim datadir-gr-self-healing-mysql-0 ProvisioningSucceeded Successfully provisioned volume pvc-e8a8b6dd-7162-48e7-86f1-48da1c21f2e6 pd.csi.storage.gke.io_gke-2b3665bdb5444e358fad-0db2-5fc3-vm_91cf7654-81a4-401d-a58f-1e36f2a9506e logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:37 +0000 UTC Normal Pod gr-self-healing-mysql-0 Scheduled Successfully assigned kuttl-test-emerging-raptor/gr-self-healing-mysql-0 to gke-jen-ps-523-f00253e-7-default-pool-bcfb7a6f-9sgd default-scheduler logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:43 +0000 UTC Normal Pod gr-self-healing-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-e8a8b6dd-7162-48e7-86f1-48da1c21f2e6" attachdetach-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:47 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:47 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 266.386193ms (266.401406ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:47 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:47 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:29:53 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:30:16 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 23.401435281s (23.401451712s including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:30:16 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:30:16 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:30:16 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:30:29 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 12.992121702s (12.992136897s including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:30:30 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:30:30 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:30:46 +0000 UTC Normal PersistentVolumeClaim datadir-gr-self-healing-mysql-1 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:30:46 +0000 UTC Normal PersistentVolumeClaim datadir-gr-self-healing-mysql-1 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:30:46 +0000 UTC Normal PersistentVolumeClaim datadir-gr-self-healing-mysql-1 Provisioning External provisioner is provisioning volume for claim "kuttl-test-emerging-raptor/datadir-gr-self-healing-mysql-1" pd.csi.storage.gke.io_gke-2b3665bdb5444e358fad-0db2-5fc3-vm_91cf7654-81a4-401d-a58f-1e36f2a9506e logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:30:46 +0000 UTC Normal StatefulSet.apps gr-self-healing-mysql SuccessfulCreate create Claim datadir-gr-self-healing-mysql-1 Pod gr-self-healing-mysql-1 in StatefulSet gr-self-healing-mysql success statefulset-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:30:46 +0000 UTC Normal StatefulSet.apps gr-self-healing-mysql SuccessfulCreate create Pod gr-self-healing-mysql-1 in StatefulSet gr-self-healing-mysql successful statefulset-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:30:50 +0000 UTC Normal PersistentVolumeClaim datadir-gr-self-healing-mysql-1 ProvisioningSucceeded Successfully provisioned volume pvc-10aabf4c-4270-4245-8228-1f128dd1e51b pd.csi.storage.gke.io_gke-2b3665bdb5444e358fad-0db2-5fc3-vm_91cf7654-81a4-401d-a58f-1e36f2a9506e logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:30:50 +0000 UTC Normal Pod gr-self-healing-mysql-1 Scheduled Successfully assigned kuttl-test-emerging-raptor/gr-self-healing-mysql-1 to gke-jen-ps-523-f00253e-7-default-pool-bcfb7a6f-tk4k default-scheduler logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:30:55 +0000 UTC Normal Pod gr-self-healing-mysql-1 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-10aabf4c-4270-4245-8228-1f128dd1e51b" attachdetach-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:30:59 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:31:16 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 16.99112108s (16.991132524s including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:31:16 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:31:16 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:31:22 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:31:44 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 22.555269848s (22.555278034s including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:31:44 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:31:44 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:31:44 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:31:56 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 11.28943687s (11.289477793s including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:31:56 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:31:56 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:32:21 +0000 UTC Warning Pod gr-self-healing-mysql-1.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:32:21 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:32:24 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 189.03574ms (189.04582ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:32:58 +0000 UTC Normal PersistentVolumeClaim datadir-gr-self-healing-mysql-2 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:32:58 +0000 UTC Normal StatefulSet.apps gr-self-healing-mysql SuccessfulCreate create Claim datadir-gr-self-healing-mysql-2 Pod gr-self-healing-mysql-2 in StatefulSet gr-self-healing-mysql success statefulset-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:32:58 +0000 UTC Normal StatefulSet.apps gr-self-healing-mysql SuccessfulCreate create Pod gr-self-healing-mysql-2 in StatefulSet gr-self-healing-mysql successful statefulset-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:32:59 +0000 UTC Normal PersistentVolumeClaim datadir-gr-self-healing-mysql-2 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:32:59 +0000 UTC Normal PersistentVolumeClaim datadir-gr-self-healing-mysql-2 Provisioning External provisioner is provisioning volume for claim "kuttl-test-emerging-raptor/datadir-gr-self-healing-mysql-2" pd.csi.storage.gke.io_gke-2b3665bdb5444e358fad-0db2-5fc3-vm_91cf7654-81a4-401d-a58f-1e36f2a9506e logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:33:02 +0000 UTC Normal PersistentVolumeClaim datadir-gr-self-healing-mysql-2 ProvisioningSucceeded Successfully provisioned volume pvc-56ee2d45-9e11-4685-9fbb-0d5cdb8a34ed pd.csi.storage.gke.io_gke-2b3665bdb5444e358fad-0db2-5fc3-vm_91cf7654-81a4-401d-a58f-1e36f2a9506e logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:33:02 +0000 UTC Normal Pod gr-self-healing-mysql-2 Scheduled Successfully assigned kuttl-test-emerging-raptor/gr-self-healing-mysql-2 to gke-jen-ps-523-f00253e-7-default-pool-bcfb7a6f-7ntv default-scheduler logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:33:10 +0000 UTC Normal Pod gr-self-healing-mysql-2 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-56ee2d45-9e11-4685-9fbb-0d5cdb8a34ed" attachdetach-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:33:12 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:33:26 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 14.68660073s (14.686617554s including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:33:26 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:33:26 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:33:33 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:33:55 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 21.719087565s (21.719108034s including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:33:55 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:33:55 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:33:55 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:34:07 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 12.062617696s (12.062633087s including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:34:07 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:34:07 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:34:31 +0000 UTC Warning Pod gr-self-healing-mysql-2.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:34:31 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:34:34 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 187.512693ms (187.526779ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:14 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-4mx6l Scheduled Successfully assigned kuttl-test-emerging-raptor/gr-self-healing-router-7b64bcbf5f-4mx6l to gke-jen-ps-523-f00253e-7-default-pool-bcfb7a6f-tk4k default-scheduler logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:14 +0000 UTC Normal ReplicaSet.apps gr-self-healing-router-7b64bcbf5f SuccessfulCreate Created pod: gr-self-healing-router-7b64bcbf5f-4mx6l replicaset-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:14 +0000 UTC Normal Deployment.apps gr-self-healing-router ScalingReplicaSet Scaled up replica set gr-self-healing-router-7b64bcbf5f to 3 deployment-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:15 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-4mx6l.spec.initContainers{router-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:15 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-4mx6l.spec.initContainers{router-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 236.68591ms (236.720172ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:15 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-hwl44 Scheduled Successfully assigned kuttl-test-emerging-raptor/gr-self-healing-router-7b64bcbf5f-hwl44 to gke-jen-ps-523-f00253e-7-default-pool-bcfb7a6f-7ntv default-scheduler logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:15 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-hwl44.spec.initContainers{router-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:15 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-ml8qn Scheduled Successfully assigned kuttl-test-emerging-raptor/gr-self-healing-router-7b64bcbf5f-ml8qn to gke-jen-ps-523-f00253e-7-default-pool-bcfb7a6f-9sgd default-scheduler logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:15 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-ml8qn.spec.initContainers{router-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:15 +0000 UTC Normal ReplicaSet.apps gr-self-healing-router-7b64bcbf5f SuccessfulCreate Created pod: gr-self-healing-router-7b64bcbf5f-ml8qn replicaset-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:15 +0000 UTC Normal ReplicaSet.apps gr-self-healing-router-7b64bcbf5f SuccessfulCreate Created pod: gr-self-healing-router-7b64bcbf5f-hwl44 replicaset-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:16 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-4mx6l.spec.initContainers{router-init} Created Created container router-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:16 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-4mx6l.spec.initContainers{router-init} Started Started container router-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:16 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-hwl44.spec.initContainers{router-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 230.91772ms (230.926397ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:16 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-hwl44.spec.initContainers{router-init} Created Created container router-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:16 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-hwl44.spec.initContainers{router-init} Started Started container router-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:16 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-ml8qn.spec.initContainers{router-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 203.180043ms (203.193948ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:16 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-ml8qn.spec.initContainers{router-init} Created Created container router-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:16 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-ml8qn.spec.initContainers{router-init} Started Started container router-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:17 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-hwl44.spec.containers{router} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-router" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:18 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-4mx6l.spec.containers{router} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-router" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:18 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-ml8qn.spec.containers{router} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-router" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:34 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-hwl44.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 16.69097578s (16.690990661s including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:34 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-hwl44.spec.containers{router} Created Created container router kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:34 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-hwl44.spec.containers{router} Started Started container router kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:35 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-4mx6l.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 17.657694648s (17.657709535s including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:35 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-4mx6l.spec.containers{router} Created Created container router kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:35 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-4mx6l.spec.containers{router} Started Started container router kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:35 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-ml8qn.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 17.060805785s (17.060820707s including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:35 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-ml8qn.spec.containers{router} Created Created container router kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:35 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-ml8qn.spec.containers{router} Started Started container router kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:59 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-kill-primary FinalizerInited Finalizer has been inited logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:59 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-kill-primary Updated Successfully update finalizer of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:59 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-kill-primary Updated Successfully update desiredPhase of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:59 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-kill-primary Applied Successfully apply chaos for kuttl-test-emerging-raptor/gr-self-healing-mysql-0 logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:59 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-kill-primary Updated Successfully update records of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:59 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:35:59 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{xtrabackup} Killing Stopping container xtrabackup kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:36:00 +0000 UTC Normal Pod gr-self-healing-mysql-0 Scheduled Successfully assigned kuttl-test-emerging-raptor/gr-self-healing-mysql-0 to gke-jen-ps-523-f00253e-7-default-pool-bcfb7a6f-9sgd default-scheduler logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:36:00 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:36:01 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 257.925052ms (257.942596ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:36:01 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:36:01 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:36:03 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:36:03 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 217.122181ms (217.136448ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:36:03 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:36:03 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:36:03 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:36:03 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 227.12128ms (227.13589ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:36:03 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:36:03 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:37:08 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-failure-primary FinalizerInited Finalizer has been inited logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:37:08 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-failure-primary Updated Successfully update finalizer of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:37:08 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-failure-primary Started Experiment has started logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:37:08 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-failure-primary Updated Successfully update desiredPhase of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:37:08 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-failure-primary Applied Successfully apply chaos for kuttl-test-emerging-raptor/gr-self-healing-mysql-1 logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:37:08 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-failure-primary Updated Successfully update records of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:37:08 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{mysql} Killing Container mysql definition changed, will be restarted kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:37:13 +0000 UTC Warning Pod gr-self-healing-mysql-1.spec.containers{mysql} Unhealthy Readiness probe failed: 2024/02/08 07:37:13 readiness check failed: connect to db: ping DB: dial tcp 10.130.177.6:33062: connect: connection refused kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:37:15 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{xtrabackup} Killing Container xtrabackup definition changed, will be restarted kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:37:15 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{mysql} Pulling Pulling image "gcr.io/google-containers/pause:latest" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:37:15 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "gcr.io/google-containers/pause:latest" in 335.39046ms (335.409011ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:37:15 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{xtrabackup} Pulling Pulling image "gcr.io/google-containers/pause:latest" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:38:08 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-failure-primary TimeUp Time up according to the duration logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:38:08 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-failure-primary Updated Successfully update desiredPhase of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:38:08 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-failure-primary Recovered Successfully recover chaos for kuttl-test-emerging-raptor/gr-self-healing-mysql-1 logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:38:08 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-failure-primary Updated Successfully update records of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:39:01 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss-primary FinalizerInited Finalizer has been inited logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:39:01 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss-primary Updated Successfully update finalizer of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:39:01 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss-primary Started Experiment has started logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:39:01 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss-primary Updated Successfully update desiredPhase of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:39:02 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss-primary Updated Successfully update records of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:39:02 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss-primary Applied Successfully apply chaos for kuttl-test-emerging-raptor/gr-self-healing-mysql-2 logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:39:02 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss-primary Updated Successfully update records of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:39:02 +0000 UTC Normal PodNetworkChaos.chaos-mesh.org gr-self-healing-mysql-2 Updated Successfully update ObservedGeneration and FailedMessage of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:39:11 +0000 UTC Warning Pod gr-self-healing-mysql-2.spec.containers{mysql} Unhealthy Liveness probe failed: 2024/02/08 07:39:11 in primary partition: false 2024/02/08 07:39:11 liveness check failed: possible split brain! kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:39:15 +0000 UTC Warning Pod gr-self-healing-router-7b64bcbf5f-4mx6l.spec.containers{router} Unhealthy Readiness probe failed: Read-write route is not healthy kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:39:15 +0000 UTC Warning Pod gr-self-healing-router-7b64bcbf5f-hwl44.spec.containers{router} Unhealthy Readiness probe failed: Read-write route is not healthy kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:39:15 +0000 UTC Warning Pod gr-self-healing-router-7b64bcbf5f-ml8qn.spec.containers{router} Unhealthy Readiness probe failed: Read-write route is not healthy kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:39:21 +0000 UTC Warning Pod gr-self-healing-mysql-2.spec.containers{mysql} Unhealthy Liveness probe failed: 2024/02/08 07:39:21 in primary partition: false 2024/02/08 07:39:21 liveness check failed: possible split brain! kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:39:31 +0000 UTC Warning Pod gr-self-healing-mysql-2.spec.containers{mysql} Unhealthy Liveness probe failed: 2024/02/08 07:39:31 in primary partition: false 2024/02/08 07:39:31 liveness check failed: possible split brain! kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:39:31 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.containers{mysql} Killing Container mysql failed liveness probe, will be restarted kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:39:45 +0000 UTC Warning Pod gr-self-healing-mysql-2.spec.containers{mysql} Unhealthy Readiness probe errored: command "/opt/percona/healthcheck readiness" timed out kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:40:01 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss-primary TimeUp Time up according to the duration logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:40:01 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss-primary Updated Successfully update desiredPhase of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:40:01 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss-primary Updated Successfully update records of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:40:01 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss-primary Recovered Successfully recover chaos for kuttl-test-emerging-raptor/gr-self-healing-mysql-2 logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:40:01 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss-primary Updated Successfully update records of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:40:01 +0000 UTC Normal PodNetworkChaos.chaos-mesh.org gr-self-healing-mysql-2 Updated Successfully update ObservedGeneration and FailedMessage of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:40:01 +0000 UTC Warning Pod gr-self-healing-mysql-2.spec.containers{mysql} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:40:01 +0000 UTC Warning Pod gr-self-healing-mysql-2.spec.containers{mysql} Unhealthy Readiness probe errored: rpc error: code = NotFound desc = failed to exec in container: failed to load task: no running task found: task c479c23d0a66cb802cee4e7a02a9ca39668a7a40e2bebb45069a59d198b1e357 not found: not found kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:40:59 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-kill-label-cluster-crash FinalizerInited Finalizer has been inited logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:40:59 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-kill-label-cluster-crash Updated Successfully update finalizer of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:40:59 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-kill-label-cluster-crash Updated Successfully update desiredPhase of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:40:59 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-kill-label-cluster-crash Applied Successfully apply chaos for kuttl-test-emerging-raptor/gr-self-healing-mysql-1 logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:40:59 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-kill-label-cluster-crash Applied Successfully apply chaos for kuttl-test-emerging-raptor/gr-self-healing-router-7b64bcbf5f-4mx6l logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:40:59 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:40:59 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:40:59 +0000 UTC Warning Endpoints gr-self-healing-mysql-proxy FailedToUpdateEndpoint Failed to update endpoint kuttl-test-emerging-raptor/gr-self-healing-mysql-proxy: Operation cannot be fulfilled on endpoints "gr-self-healing-mysql-proxy": the object has been modified; please apply your changes to the latest version and try again endpoint-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:40:59 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-4mx6l.spec.containers{router} Killing Stopping container router kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:00 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-kill-label-cluster-crash Applied Successfully apply chaos for kuttl-test-emerging-raptor/gr-self-healing-mysql-0 logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:00 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-kill-label-cluster-crash Applied Successfully apply chaos for kuttl-test-emerging-raptor/gr-self-healing-router-7b64bcbf5f-hwl44 logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:00 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-kill-label-cluster-crash Applied Successfully apply chaos for kuttl-test-emerging-raptor/gr-self-healing-mysql-2 logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:00 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-kill-label-cluster-crash Applied Successfully apply chaos for kuttl-test-emerging-raptor/gr-self-healing-router-7b64bcbf5f-ml8qn logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:00 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-kill-label-cluster-crash Updated Successfully update records of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:00 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{xtrabackup} Killing Stopping container xtrabackup kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:00 +0000 UTC Normal Pod gr-self-healing-mysql-1 Scheduled Successfully assigned kuttl-test-emerging-raptor/gr-self-healing-mysql-1 to gke-jen-ps-523-f00253e-7-default-pool-bcfb7a6f-9sgd default-scheduler logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:00 +0000 UTC Warning Pod gr-self-healing-mysql-1 FailedAttachVolume Multi-Attach error for volume "pvc-10aabf4c-4270-4245-8228-1f128dd1e51b" Volume is already exclusively attached to one node and can't be attached to another attachdetach-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:00 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-84jvg Scheduled Successfully assigned kuttl-test-emerging-raptor/gr-self-healing-router-7b64bcbf5f-84jvg to gke-jen-ps-523-f00253e-7-default-pool-bcfb7a6f-tk4k default-scheduler logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:00 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-hwl44.spec.containers{router} Killing Stopping container router kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:00 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-ml8qn.spec.containers{router} Killing Stopping container router kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:00 +0000 UTC Normal ReplicaSet.apps gr-self-healing-router-7b64bcbf5f SuccessfulCreate Created pod: gr-self-healing-router-7b64bcbf5f-84jvg replicaset-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:00 +0000 UTC Warning Endpoints gr-self-healing-router FailedToUpdateEndpoint Failed to update endpoint kuttl-test-emerging-raptor/gr-self-healing-router: Operation cannot be fulfilled on endpoints "gr-self-healing-router": the object has been modified; please apply your changes to the latest version and try again endpoint-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:01 +0000 UTC Normal Pod gr-self-healing-mysql-0 Scheduled Successfully assigned kuttl-test-emerging-raptor/gr-self-healing-mysql-0 to gke-jen-ps-523-f00253e-7-default-pool-bcfb7a6f-7ntv default-scheduler logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:01 +0000 UTC Warning Pod gr-self-healing-mysql-0 FailedAttachVolume Multi-Attach error for volume "pvc-e8a8b6dd-7162-48e7-86f1-48da1c21f2e6" Volume is already exclusively attached to one node and can't be attached to another attachdetach-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:01 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-84jvg.spec.initContainers{router-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:01 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-84jvg.spec.initContainers{router-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 238.151813ms (238.172452ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:01 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-84jvg.spec.initContainers{router-init} Created Created container router-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:01 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-84jvg.spec.initContainers{router-init} Started Started container router-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:01 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-9mkdw Scheduled Successfully assigned kuttl-test-emerging-raptor/gr-self-healing-router-7b64bcbf5f-9mkdw to gke-jen-ps-523-f00253e-7-default-pool-bcfb7a6f-7ntv default-scheduler logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:01 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-9mkdw.spec.initContainers{router-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:01 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-hrkbt Scheduled Successfully assigned kuttl-test-emerging-raptor/gr-self-healing-router-7b64bcbf5f-hrkbt to gke-jen-ps-523-f00253e-7-default-pool-bcfb7a6f-9sgd default-scheduler logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:01 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-hrkbt.spec.initContainers{router-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:01 +0000 UTC Normal ReplicaSet.apps gr-self-healing-router-7b64bcbf5f SuccessfulCreate Created pod: gr-self-healing-router-7b64bcbf5f-hrkbt replicaset-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:01 +0000 UTC Normal ReplicaSet.apps gr-self-healing-router-7b64bcbf5f SuccessfulCreate Created pod: gr-self-healing-router-7b64bcbf5f-9mkdw replicaset-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:02 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-9mkdw.spec.initContainers{router-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 225.49135ms (225.505798ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:02 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-9mkdw.spec.initContainers{router-init} Created Created container router-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:02 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-9mkdw.spec.initContainers{router-init} Started Started container router-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:02 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-hrkbt.spec.initContainers{router-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 219.635387ms (219.654425ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:02 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-hrkbt.spec.initContainers{router-init} Created Created container router-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:02 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-hrkbt.spec.initContainers{router-init} Started Started container router-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:02 +0000 UTC Warning Pod gr-self-healing-router-7b64bcbf5f-hwl44.spec.containers{router} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:03 +0000 UTC Warning Pod gr-self-healing-mysql-0.spec.containers{mysql} Unhealthy Readiness probe failed: [mysql] 2024/02/08 07:41:03 packets.go:37: read tcp 10.130.178.10:53484->10.130.178.10:33062: read: connection reset by peer 2024/02/08 07:41:03 readiness check failed: connect to db: ping DB: dial tcp 10.130.178.10:33062: connect: connection refused kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:03 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-84jvg.spec.containers{router} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-router" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:03 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-84jvg.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 218.971921ms (218.984337ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:03 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-84jvg.spec.containers{router} Created Created container router kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:03 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-84jvg.spec.containers{router} Started Started container router kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:03 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-9mkdw.spec.containers{router} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-router" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:04 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-84jvg.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 219.531883ms (219.546893ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:04 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-9mkdw.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 187.818804ms (187.827693ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:04 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-9mkdw.spec.containers{router} Created Created container router kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:04 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-9mkdw.spec.containers{router} Started Started container router kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:04 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-hrkbt.spec.containers{router} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-router" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:04 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-hrkbt.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 241.996367ms (242.003624ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:04 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-hrkbt.spec.containers{router} Created Created container router kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:04 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-hrkbt.spec.containers{router} Started Started container router kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:05 +0000 UTC Warning Pod gr-self-healing-router-7b64bcbf5f-84jvg.spec.containers{router} BackOff Back-off restarting failed container router in pod gr-self-healing-router-7b64bcbf5f-84jvg_kuttl-test-emerging-raptor(9ff675a0-0290-47a9-937f-4c85f4d9e89e) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:05 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-9mkdw.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 211.03562ms (211.055931ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:05 +0000 UTC Warning Pod gr-self-healing-router-7b64bcbf5f-9mkdw.spec.containers{router} BackOff Back-off restarting failed container router in pod gr-self-healing-router-7b64bcbf5f-9mkdw_kuttl-test-emerging-raptor(31733abd-8ddd-4671-bc20-907cb99f6ed0) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:05 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-hrkbt.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 178.826785ms (178.839311ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:06 +0000 UTC Warning Pod gr-self-healing-router-7b64bcbf5f-hrkbt.spec.containers{router} BackOff Back-off restarting failed container router in pod gr-self-healing-router-7b64bcbf5f-hrkbt_kuttl-test-emerging-raptor(2e5b8c49-33e5-4047-aef4-f453a8eb0658) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:19 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-84jvg.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 228.599238ms (228.616889ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:21 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-9mkdw.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 265.090423ms (265.123775ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:22 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-hrkbt.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 242.393121ms (242.411438ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:24 +0000 UTC Normal Pod gr-self-healing-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-e8a8b6dd-7162-48e7-86f1-48da1c21f2e6" attachdetach-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:25 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:25 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 178.729446ms (178.748622ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:25 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:25 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:26 +0000 UTC Normal Pod gr-self-healing-mysql-1 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-10aabf4c-4270-4245-8228-1f128dd1e51b" attachdetach-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:27 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:27 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 172.234815ms (172.245087ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:27 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:27 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:27 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:27 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 220.960618ms (220.974354ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:27 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:27 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:28 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:28 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 226.955569ms (226.963897ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:28 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:28 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:30 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:30 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 221.503306ms (221.520266ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:30 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:30 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:30 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:30 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 280.123452ms (280.131444ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:30 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:30 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:47 +0000 UTC Warning Pod gr-self-healing-mysql-0.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:47 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:49 +0000 UTC Warning Pod gr-self-healing-mysql-1.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:49 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:50 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-9mkdw.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 225.418567ms (225.429142ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:41:56 +0000 UTC Normal Pod gr-self-healing-router-7b64bcbf5f-hrkbt.spec.containers{router} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-router" in 224.404068ms (224.431394ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:42:17 +0000 UTC Normal Pod gr-self-healing-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 241.972898ms (241.987561ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:42:19 +0000 UTC Normal Pod gr-self-healing-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 233.533357ms (233.550984ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:42:52 +0000 UTC Normal Pod gr-self-healing-mysql-2 Scheduled Successfully assigned kuttl-test-emerging-raptor/gr-self-healing-mysql-2 to gke-jen-ps-523-f00253e-7-default-pool-bcfb7a6f-tk4k default-scheduler logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:42:52 +0000 UTC Warning PerconaServerMySQL.ps.percona.com gr-self-healing FullClusterCrashDetected Full cluster crash detected ps-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:43:00 +0000 UTC Normal Pod gr-self-healing-mysql-2 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-56ee2d45-9e11-4685-9fbb-0d5cdb8a34ed" attachdetach-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:43:01 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-523-f00253e" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:43:02 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-523-f00253e" in 252.984443ms (252.991885ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:43:02 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.initContainers{mysql-init} Created Created container mysql-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:43:02 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:43:03 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:43:03 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 270.500665ms (270.515205ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:43:03 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.containers{mysql} Created Created container mysql kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:43:03 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:43:03 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:43:03 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 243.312792ms (243.322116ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:43:03 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.containers{xtrabackup} Created Created container xtrabackup kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:43:04 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:43:23 +0000 UTC Warning Pod gr-self-healing-mysql-2.spec.containers{mysql} Unhealthy Startup probe failed: kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:43:23 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:43:53 +0000 UTC Normal Pod gr-self-healing-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 234.05971ms (234.068068ms including waiting) kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:45:02 +0000 UTC Normal PerconaServerMySQL.ps.percona.com gr-self-healing FullClusterCrashRecovered Cluster recovered from full cluster crash ps-controller logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:29 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss-primary FinalizerInited Finalizer has been removed logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:29 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss-primary Updated Successfully update finalizer of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:30 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-kill-label-cluster-crash Deleted Experiment has been deleted logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:30 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-kill-label-cluster-crash Updated Successfully update desiredPhase of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:30 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-kill-label-cluster-crash Recovered Successfully recover chaos for kuttl-test-emerging-raptor/gr-self-healing-mysql-1 logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:30 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-kill-label-cluster-crash Recovered Successfully recover chaos for kuttl-test-emerging-raptor/gr-self-healing-router-7b64bcbf5f-4mx6l logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:30 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-kill-label-cluster-crash Recovered Successfully recover chaos for kuttl-test-emerging-raptor/gr-self-healing-mysql-0 logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:30 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-kill-label-cluster-crash Recovered Successfully recover chaos for kuttl-test-emerging-raptor/gr-self-healing-router-7b64bcbf5f-hwl44 logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:30 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-kill-label-cluster-crash Recovered Successfully recover chaos for kuttl-test-emerging-raptor/gr-self-healing-mysql-2 logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:30 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-kill-label-cluster-crash Recovered Successfully recover chaos for kuttl-test-emerging-raptor/gr-self-healing-router-7b64bcbf5f-ml8qn logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:30 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-kill-label-cluster-crash Updated Successfully update records of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:30 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-kill-label-cluster-crash FinalizerInited Finalizer has been removed logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:30 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-kill-label-cluster-crash Updated Successfully update finalizer of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:30 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-failure-primary FinalizerInited Finalizer has been removed logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:30 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-failure-primary Updated Successfully update finalizer of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:30 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-kill-primary Deleted Experiment has been deleted logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:30 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-kill-primary Updated Successfully update desiredPhase of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:30 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-kill-primary Recovered Successfully recover chaos for kuttl-test-emerging-raptor/gr-self-healing-mysql-0 logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:30 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-kill-primary Updated Successfully update records of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:30 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-kill-primary FinalizerInited Finalizer has been removed logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:30 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-kill-primary Updated Successfully update finalizer of resource logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:38 +0000 UTC Normal Pod chaos-controller-manager-79c896cbb8-5l4x6.spec.containers{chaos-mesh} Killing Stopping container chaos-mesh kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:38 +0000 UTC Normal Pod chaos-controller-manager-79c896cbb8-knt2h.spec.containers{chaos-mesh} Killing Stopping container chaos-mesh kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:38 +0000 UTC Normal Pod chaos-controller-manager-79c896cbb8-tg794.spec.containers{chaos-mesh} Killing Stopping container chaos-mesh kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:38 +0000 UTC Normal Pod chaos-daemon-5h7zt.spec.containers{chaos-daemon} Killing Stopping container chaos-daemon kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:38 +0000 UTC Normal Pod chaos-daemon-6rlxh.spec.containers{chaos-daemon} Killing Stopping container chaos-daemon kubelet logger.go:42: 07:48:00 | gr-self-healing | 2024-02-08 07:47:38 +0000 UTC Normal Pod chaos-daemon-xpr87.spec.containers{chaos-daemon} Killing Stopping container chaos-daemon kubelet logger.go:42: 07:48:00 | gr-self-healing | Deleting namespace: kuttl-test-emerging-raptor === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (1260.93s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/gr-self-healing (1252.43s) PASS