=== RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://34.70.140.115 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 180 seconds for each step harness.go:372: testsuite: e2e-tests/tests has 26 tests === RUN kuttl/harness === RUN kuttl/harness/self-healing-chaos === PAUSE kuttl/harness/self-healing-chaos === CONT kuttl/harness/self-healing-chaos logger.go:42: 13:09:44 | self-healing-chaos | Creating namespace: kuttl-test-set-bluejay logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | starting test step 0-deploy-operator logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | running command: [sh -c set -o errexit set -o xtrace source ../../functions deploy_operator deploy_non_tls_cluster_secrets deploy_tls_cluster_secrets deploy_client] logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | + source ../../functions logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ realpath ../../.. logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | ++++ mktemp -d logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ export TEMP_DIR=/tmp/tmp.sGspNY1Qz7 logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ TEMP_DIR=/tmp/tmp.sGspNY1Qz7 logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ export GIT_BRANCH=PR-339 logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ GIT_BRANCH=PR-339 logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ export VERSION=PR-339-3a8483b logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ VERSION=PR-339-3a8483b logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | ++++ which gdate logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | ++++ which date logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ date=/usr/bin/date logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ command -v oc logger.go:42: 13:09:44 | self-healing-chaos/0-deploy-operator | +++ oc get projects logger.go:42: 13:09:50 | self-healing-chaos/0-deploy-operator | error: the server doesn't have a resource type "projects" logger.go:42: 13:09:50 | self-healing-chaos/0-deploy-operator | +++ kubectl get nodes logger.go:42: 13:09:50 | self-healing-chaos/0-deploy-operator | +++ grep '^minikube' logger.go:42: 13:09:50 | self-healing-chaos/0-deploy-operator | ++++ pwd logger.go:42: 13:09:50 | self-healing-chaos/0-deploy-operator | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/self-healing-chaos logger.go:42: 13:09:50 | self-healing-chaos/0-deploy-operator | ++ test_name=self-healing-chaos logger.go:42: 13:09:50 | self-healing-chaos/0-deploy-operator | + deploy_operator logger.go:42: 13:09:50 | self-healing-chaos/0-deploy-operator | + kubectl -n kuttl-test-set-bluejay apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/crd.yaml logger.go:42: 13:09:52 | self-healing-chaos/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlbackups.ps.percona.com serverside-applied logger.go:42: 13:09:52 | self-healing-chaos/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlrestores.ps.percona.com serverside-applied logger.go:42: 13:09:53 | self-healing-chaos/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqls.ps.percona.com serverside-applied logger.go:42: 13:09:53 | self-healing-chaos/0-deploy-operator | + kubectl -n kuttl-test-set-bluejay apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/rbac.yaml logger.go:42: 13:09:54 | self-healing-chaos/0-deploy-operator | serviceaccount/percona-server-mysql-operator created logger.go:42: 13:09:54 | self-healing-chaos/0-deploy-operator | serviceaccount/percona-server-mysql-operator-orchestrator created logger.go:42: 13:09:55 | self-healing-chaos/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 13:09:55 | self-healing-chaos/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 13:09:55 | self-healing-chaos/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 13:09:56 | self-healing-chaos/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 13:09:56 | self-healing-chaos/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 13:09:56 | self-healing-chaos/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 13:09:56 | self-healing-chaos/0-deploy-operator | + yq eval '(select(documentIndex==1).spec.template.spec.containers[] | select(.name=="manager").env[] | select(.name=="DISABLE_TELEMETRY").value) = "true"' logger.go:42: 13:09:56 | self-healing-chaos/0-deploy-operator | + kubectl -n kuttl-test-set-bluejay apply -f - logger.go:42: 13:09:56 | self-healing-chaos/0-deploy-operator | ++ printf 'select(documentIndex==1).spec.template.spec.containers[0].image="%s"' perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:09:56 | self-healing-chaos/0-deploy-operator | + yq eval 'select(documentIndex==1).spec.template.spec.containers[0].image="perconalab/percona-server-mysql-operator:PR-339-3a8483b"' /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/operator.yaml logger.go:42: 13:09:58 | self-healing-chaos/0-deploy-operator | configmap/percona-server-mysql-operator-config created logger.go:42: 13:09:58 | self-healing-chaos/0-deploy-operator | deployment.apps/percona-server-mysql-operator created logger.go:42: 13:09:58 | self-healing-chaos/0-deploy-operator | + deploy_non_tls_cluster_secrets logger.go:42: 13:09:58 | self-healing-chaos/0-deploy-operator | + kubectl -n kuttl-test-set-bluejay apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/secrets.yaml logger.go:42: 13:09:59 | self-healing-chaos/0-deploy-operator | secret/test-secrets created logger.go:42: 13:09:59 | self-healing-chaos/0-deploy-operator | + deploy_tls_cluster_secrets logger.go:42: 13:09:59 | self-healing-chaos/0-deploy-operator | + kubectl -n kuttl-test-set-bluejay apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/ssl-secret.yaml logger.go:42: 13:10:01 | self-healing-chaos/0-deploy-operator | secret/test-ssl created logger.go:42: 13:10:01 | self-healing-chaos/0-deploy-operator | + deploy_client logger.go:42: 13:10:01 | self-healing-chaos/0-deploy-operator | + kubectl -n kuttl-test-set-bluejay apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/client.yaml logger.go:42: 13:10:02 | self-healing-chaos/0-deploy-operator | pod/mysql-client created logger.go:42: 13:10:09 | self-healing-chaos/0-deploy-operator | test step completed 0-deploy-operator logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | starting test step 1-create-cluster logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | running command: [sh -c set -o errexit set -o xtrace source ../../functions deploy_chaos_mesh "${NAMESPACE}" get_cr \ | yq eval '.spec.mysql.clusterType="async"' - \ | kubectl -n "${NAMESPACE}" apply -f - sleep 60] logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | + source ../../functions logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ realpath ../../.. logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | ++++ mktemp -d logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ export TEMP_DIR=/tmp/tmp.YSEYNHq52C logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ TEMP_DIR=/tmp/tmp.YSEYNHq52C logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ export GIT_BRANCH=PR-339 logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ GIT_BRANCH=PR-339 logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ export VERSION=PR-339-3a8483b logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ VERSION=PR-339-3a8483b logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | ++++ which gdate logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | ++++ which date logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ date=/usr/bin/date logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ command -v oc logger.go:42: 13:10:09 | self-healing-chaos/1-create-cluster | +++ oc get projects logger.go:42: 13:10:15 | self-healing-chaos/1-create-cluster | error: the server doesn't have a resource type "projects" logger.go:42: 13:10:15 | self-healing-chaos/1-create-cluster | +++ kubectl get nodes logger.go:42: 13:10:15 | self-healing-chaos/1-create-cluster | +++ grep '^minikube' logger.go:42: 13:10:16 | self-healing-chaos/1-create-cluster | ++++ pwd logger.go:42: 13:10:16 | self-healing-chaos/1-create-cluster | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/self-healing-chaos logger.go:42: 13:10:16 | self-healing-chaos/1-create-cluster | ++ test_name=self-healing-chaos logger.go:42: 13:10:16 | self-healing-chaos/1-create-cluster | + deploy_chaos_mesh kuttl-test-set-bluejay logger.go:42: 13:10:16 | self-healing-chaos/1-create-cluster | + local chaos_mesh_ns=kuttl-test-set-bluejay logger.go:42: 13:10:16 | self-healing-chaos/1-create-cluster | + destroy_chaos_mesh logger.go:42: 13:10:16 | self-healing-chaos/1-create-cluster | + local chaos_mesh_ns logger.go:42: 13:10:16 | self-healing-chaos/1-create-cluster | ++ helm list --all-namespaces --filter chaos-mesh logger.go:42: 13:10:16 | self-healing-chaos/1-create-cluster | ++ tail -n1 logger.go:42: 13:10:16 | self-healing-chaos/1-create-cluster | ++ awk '-F ' '{print $2}' logger.go:42: 13:10:16 | self-healing-chaos/1-create-cluster | ++ sed s/NAMESPACE// logger.go:42: 13:10:16 | self-healing-chaos/1-create-cluster | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 13:10:16 | self-healing-chaos/1-create-cluster | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 13:10:17 | self-healing-chaos/1-create-cluster | + chaos_mesh_ns= logger.go:42: 13:10:17 | self-healing-chaos/1-create-cluster | + '[' -n '' ']' logger.go:42: 13:10:17 | self-healing-chaos/1-create-cluster | ++ kubectl api-resources logger.go:42: 13:10:17 | self-healing-chaos/1-create-cluster | ++ grep chaos-mesh logger.go:42: 13:10:17 | self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 13:10:18 | self-healing-chaos/1-create-cluster | ++ kubectl get crd -A logger.go:42: 13:10:18 | self-healing-chaos/1-create-cluster | ++ grep chaos-mesh.org logger.go:42: 13:10:18 | self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 13:10:19 | self-healing-chaos/1-create-cluster | ++ kubectl get clusterrolebinding -A logger.go:42: 13:10:19 | self-healing-chaos/1-create-cluster | ++ grep chaos-mesh logger.go:42: 13:10:19 | self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 13:10:20 | self-healing-chaos/1-create-cluster | ++ kubectl get clusterrole -A logger.go:42: 13:10:20 | self-healing-chaos/1-create-cluster | ++ grep chaos-mesh logger.go:42: 13:10:20 | self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 13:10:21 | self-healing-chaos/1-create-cluster | ++ kubectl get mutatingwebhookconfiguration -A logger.go:42: 13:10:21 | self-healing-chaos/1-create-cluster | ++ grep chaos-mesh logger.go:42: 13:10:21 | self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 13:10:22 | self-healing-chaos/1-create-cluster | ++ kubectl get validatingwebhookconfiguration -A logger.go:42: 13:10:22 | self-healing-chaos/1-create-cluster | ++ grep chaos-mesh logger.go:42: 13:10:22 | self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 13:10:22 | self-healing-chaos/1-create-cluster | ++ kubectl get validatingwebhookconfiguration -A logger.go:42: 13:10:22 | self-healing-chaos/1-create-cluster | ++ grep validate-auth logger.go:42: 13:10:22 | self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 13:10:23 | self-healing-chaos/1-create-cluster | + helm repo add chaos-mesh https://charts.chaos-mesh.org logger.go:42: 13:10:23 | self-healing-chaos/1-create-cluster | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 13:10:23 | self-healing-chaos/1-create-cluster | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 13:10:23 | self-healing-chaos/1-create-cluster | "chaos-mesh" has been added to your repositories logger.go:42: 13:10:23 | self-healing-chaos/1-create-cluster | + helm install chaos-mesh chaos-mesh/chaos-mesh --namespace=kuttl-test-set-bluejay --set chaosDaemon.runtime=containerd --set chaosDaemon.socketPath=/run/containerd/containerd.sock --set dashboard.create=false --version 2.5.1 logger.go:42: 13:10:23 | self-healing-chaos/1-create-cluster | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 13:10:23 | self-healing-chaos/1-create-cluster | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 13:10:54 | self-healing-chaos/1-create-cluster | NAME: chaos-mesh logger.go:42: 13:10:54 | self-healing-chaos/1-create-cluster | LAST DEPLOYED: Wed Aug 9 13:10:39 2023 logger.go:42: 13:10:54 | self-healing-chaos/1-create-cluster | NAMESPACE: kuttl-test-set-bluejay logger.go:42: 13:10:54 | self-healing-chaos/1-create-cluster | STATUS: deployed logger.go:42: 13:10:54 | self-healing-chaos/1-create-cluster | REVISION: 1 logger.go:42: 13:10:54 | self-healing-chaos/1-create-cluster | TEST SUITE: None logger.go:42: 13:10:54 | self-healing-chaos/1-create-cluster | NOTES: logger.go:42: 13:10:54 | self-healing-chaos/1-create-cluster | 1. Make sure chaos-mesh components are running logger.go:42: 13:10:54 | self-healing-chaos/1-create-cluster | kubectl get pods --namespace kuttl-test-set-bluejay -l app.kubernetes.io/instance=chaos-mesh logger.go:42: 13:10:54 | self-healing-chaos/1-create-cluster | + sleep 10 logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | + get_cr logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | + local name_suffix= logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | + kubectl -n kuttl-test-set-bluejay apply -f - logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | ++ printf '.metadata.name="%s"' self-healing-chaos logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | + yq eval '.metadata.name="self-healing-chaos"' /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/cr.yaml logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | + yq eval '.spec.secretsName="test-secrets"' - logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | ++ printf '.spec.initImage="%s"' perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | + yq eval '.spec.initImage="perconalab/percona-server-mysql-operator:PR-339-3a8483b"' - logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | ++ printf '.spec.mysql.image="%s"' perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | + yq eval '.spec.mysql.image="perconalab/percona-server-mysql-operator:main-psmysql"' - logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | + yq eval '.spec.sslSecretName="test-ssl"' - logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | + yq eval '.spec.upgradeOptions.apply="disabled"' - logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | + '[' -n '' ']' logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | + yq eval - logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | ++ printf '.spec.backup.image="%s"' perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | ++ printf '.spec.orchestrator.image="%s"' perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | + yq eval '.spec.backup.image="perconalab/percona-server-mysql-operator:main-backup"' - logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | + yq eval '.spec.orchestrator.image="perconalab/percona-server-mysql-operator:main-orchestrator"' - logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | ++ printf '.spec.toolkit.image="%s"' perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | + yq eval '.spec.toolkit.image="perconalab/percona-server-mysql-operator:main-toolkit"' - logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | ++ printf '.spec.proxy.router.image="%s"' perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | + yq eval '.spec.proxy.router.image="perconalab/percona-server-mysql-operator:main-router"' - logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | ++ printf '.spec.proxy.haproxy.image="%s"' perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | ++ printf '.spec.pmm.image="%s"' perconalab/pmm-client:dev-latest logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | + yq eval '.spec.proxy.haproxy.image="perconalab/percona-server-mysql-operator:main-haproxy"' - logger.go:42: 13:11:04 | self-healing-chaos/1-create-cluster | + yq eval '.spec.pmm.image="perconalab/pmm-client:dev-latest"' - logger.go:42: 13:11:09 | self-healing-chaos/1-create-cluster | perconaservermysql.ps.percona.com/self-healing-chaos created logger.go:42: 13:11:09 | self-healing-chaos/1-create-cluster | + sleep 60 logger.go:42: 13:14:23 | self-healing-chaos/1-create-cluster | test step completed 1-create-cluster logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | starting test step 2-write-data logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_mysql \ "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password" run_mysql \ "INSERT myDB.myTable (id) VALUES (100500)" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password"] logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | + source ../../functions logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ realpath ../../.. logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | ++++ mktemp -d logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ export TEMP_DIR=/tmp/tmp.pSbYSBxXjV logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ TEMP_DIR=/tmp/tmp.pSbYSBxXjV logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ export GIT_BRANCH=PR-339 logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ GIT_BRANCH=PR-339 logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ export VERSION=PR-339-3a8483b logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ VERSION=PR-339-3a8483b logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | ++++ which gdate logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | ++++ which date logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ date=/usr/bin/date logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ command -v oc logger.go:42: 13:14:23 | self-healing-chaos/2-write-data | +++ oc get projects logger.go:42: 13:14:29 | self-healing-chaos/2-write-data | error: the server doesn't have a resource type "projects" logger.go:42: 13:14:29 | self-healing-chaos/2-write-data | +++ kubectl get nodes logger.go:42: 13:14:29 | self-healing-chaos/2-write-data | +++ grep '^minikube' logger.go:42: 13:14:30 | self-healing-chaos/2-write-data | ++++ pwd logger.go:42: 13:14:30 | self-healing-chaos/2-write-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/self-healing-chaos logger.go:42: 13:14:30 | self-healing-chaos/2-write-data | ++ test_name=self-healing-chaos logger.go:42: 13:14:30 | self-healing-chaos/2-write-data | +++ get_cluster_name logger.go:42: 13:14:30 | self-healing-chaos/2-write-data | +++ kubectl -n kuttl-test-set-bluejay get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 13:14:31 | self-healing-chaos/2-write-data | ++ get_haproxy_svc self-healing-chaos logger.go:42: 13:14:31 | self-healing-chaos/2-write-data | ++ local cluster=self-healing-chaos logger.go:42: 13:14:31 | self-healing-chaos/2-write-data | ++ echo self-healing-chaos-haproxy logger.go:42: 13:14:31 | self-healing-chaos/2-write-data | + run_mysql 'CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' '-h self-healing-chaos-haproxy -uroot -proot_password' logger.go:42: 13:14:31 | self-healing-chaos/2-write-data | + local 'command=CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' logger.go:42: 13:14:31 | self-healing-chaos/2-write-data | + local 'uri=-h self-healing-chaos-haproxy -uroot -proot_password' logger.go:42: 13:14:31 | self-healing-chaos/2-write-data | + local pod= logger.go:42: 13:14:31 | self-healing-chaos/2-write-data | ++ get_client_pod logger.go:42: 13:14:31 | self-healing-chaos/2-write-data | ++ kubectl -n kuttl-test-set-bluejay get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:14:32 | self-healing-chaos/2-write-data | + client_pod=mysql-client logger.go:42: 13:14:32 | self-healing-chaos/2-write-data | + wait_pod mysql-client logger.go:42: 13:14:32 | self-healing-chaos/2-write-data | + local pod=mysql-client logger.go:42: 13:14:32 | self-healing-chaos/2-write-data | + set +o xtrace logger.go:42: 13:14:32 | self-healing-chaos/2-write-data | mysql-clienttrue logger.go:42: 13:14:32 | self-healing-chaos/2-write-data | + kubectl -n kuttl-test-set-bluejay exec mysql-client -- bash -c 'printf '\''%s\n'\'' "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" | mysql -sN -h self-healing-chaos-haproxy -uroot -proot_password' logger.go:42: 13:14:32 | self-healing-chaos/2-write-data | + sed -e 's/mysql: //' logger.go:42: 13:14:32 | self-healing-chaos/2-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:14:34 | self-healing-chaos/2-write-data | + : logger.go:42: 13:14:34 | self-healing-chaos/2-write-data | +++ get_cluster_name logger.go:42: 13:14:34 | self-healing-chaos/2-write-data | +++ kubectl -n kuttl-test-set-bluejay get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 13:14:35 | self-healing-chaos/2-write-data | ++ get_haproxy_svc self-healing-chaos logger.go:42: 13:14:35 | self-healing-chaos/2-write-data | ++ local cluster=self-healing-chaos logger.go:42: 13:14:35 | self-healing-chaos/2-write-data | ++ echo self-healing-chaos-haproxy logger.go:42: 13:14:35 | self-healing-chaos/2-write-data | + run_mysql 'INSERT myDB.myTable (id) VALUES (100500)' '-h self-healing-chaos-haproxy -uroot -proot_password' logger.go:42: 13:14:35 | self-healing-chaos/2-write-data | + local 'command=INSERT myDB.myTable (id) VALUES (100500)' logger.go:42: 13:14:35 | self-healing-chaos/2-write-data | + local 'uri=-h self-healing-chaos-haproxy -uroot -proot_password' logger.go:42: 13:14:35 | self-healing-chaos/2-write-data | + local pod= logger.go:42: 13:14:35 | self-healing-chaos/2-write-data | ++ get_client_pod logger.go:42: 13:14:35 | self-healing-chaos/2-write-data | ++ kubectl -n kuttl-test-set-bluejay get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:14:35 | self-healing-chaos/2-write-data | + client_pod=mysql-client logger.go:42: 13:14:35 | self-healing-chaos/2-write-data | + wait_pod mysql-client logger.go:42: 13:14:35 | self-healing-chaos/2-write-data | + local pod=mysql-client logger.go:42: 13:14:35 | self-healing-chaos/2-write-data | + set +o xtrace logger.go:42: 13:14:36 | self-healing-chaos/2-write-data | mysql-clienttrue logger.go:42: 13:14:36 | self-healing-chaos/2-write-data | + sed -e 's/mysql: //' logger.go:42: 13:14:36 | self-healing-chaos/2-write-data | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:14:36 | self-healing-chaos/2-write-data | + kubectl -n kuttl-test-set-bluejay exec mysql-client -- bash -c 'printf '\''%s\n'\'' "INSERT myDB.myTable (id) VALUES (100500)" | mysql -sN -h self-healing-chaos-haproxy -uroot -proot_password' logger.go:42: 13:14:38 | self-healing-chaos/2-write-data | + : logger.go:42: 13:14:39 | self-healing-chaos/2-write-data | test step completed 2-write-data logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | starting test step 3-kill-pod logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | running command: [sh -c set -o errexit set -o xtrace source ../../functions pod=self-healing-chaos-mysql-0 yq eval '.metadata.name = "chaos-pod-kill" | del(.spec.selector.pods.test-namespace) | .metadata.namespace = "'${NAMESPACE}'" | .spec.selector.pods.'${NAMESPACE}'[0] = "'$pod'"' "${TESTS_CONFIG_DIR}/chaos-pod-kill.yml" \ | kubectl apply -f - sleep 5] logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | + source ../../functions logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ realpath ../../.. logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | ++++ mktemp -d logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ export TEMP_DIR=/tmp/tmp.FkuRBwrkL5 logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ TEMP_DIR=/tmp/tmp.FkuRBwrkL5 logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ export GIT_BRANCH=PR-339 logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ GIT_BRANCH=PR-339 logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ export VERSION=PR-339-3a8483b logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ VERSION=PR-339-3a8483b logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | ++++ which gdate logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | ++++ which date logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ date=/usr/bin/date logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ command -v oc logger.go:42: 13:14:39 | self-healing-chaos/3-kill-pod | +++ oc get projects logger.go:42: 13:14:45 | self-healing-chaos/3-kill-pod | error: the server doesn't have a resource type "projects" logger.go:42: 13:14:45 | self-healing-chaos/3-kill-pod | +++ kubectl get nodes logger.go:42: 13:14:45 | self-healing-chaos/3-kill-pod | +++ grep '^minikube' logger.go:42: 13:14:45 | self-healing-chaos/3-kill-pod | ++++ pwd logger.go:42: 13:14:45 | self-healing-chaos/3-kill-pod | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/self-healing-chaos logger.go:42: 13:14:45 | self-healing-chaos/3-kill-pod | ++ test_name=self-healing-chaos logger.go:42: 13:14:45 | self-healing-chaos/3-kill-pod | + pod=self-healing-chaos-mysql-0 logger.go:42: 13:14:45 | self-healing-chaos/3-kill-pod | + kubectl apply -f - logger.go:42: 13:14:45 | self-healing-chaos/3-kill-pod | + yq eval '.metadata.name = "chaos-pod-kill" | logger.go:42: 13:14:45 | self-healing-chaos/3-kill-pod | del(.spec.selector.pods.test-namespace) | logger.go:42: 13:14:45 | self-healing-chaos/3-kill-pod | .metadata.namespace = "kuttl-test-set-bluejay" | logger.go:42: 13:14:45 | self-healing-chaos/3-kill-pod | .spec.selector.pods.kuttl-test-set-bluejay[0] = "self-healing-chaos-mysql-0"' /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/chaos-pod-kill.yml logger.go:42: 13:14:47 | self-healing-chaos/3-kill-pod | podchaos.chaos-mesh.org/chaos-pod-kill created logger.go:42: 13:14:47 | self-healing-chaos/3-kill-pod | + sleep 5 logger.go:42: 13:15:48 | self-healing-chaos/3-kill-pod | test step completed 3-kill-pod logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | starting test step 4-read-data logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions cluster_name=$(get_cluster_name) for i in 0 1 2; do data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 04-read-data-${i} --from-literal=data="${data}" done] logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | + source ../../functions logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ realpath ../../.. logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | ++++ mktemp -d logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ export TEMP_DIR=/tmp/tmp.QKdYxRRChV logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ TEMP_DIR=/tmp/tmp.QKdYxRRChV logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ export GIT_BRANCH=PR-339 logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ GIT_BRANCH=PR-339 logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ export VERSION=PR-339-3a8483b logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ VERSION=PR-339-3a8483b logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | ++++ which gdate logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | ++++ which date logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ date=/usr/bin/date logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ command -v oc logger.go:42: 13:15:48 | self-healing-chaos/4-read-data | +++ oc get projects logger.go:42: 13:15:54 | self-healing-chaos/4-read-data | error: the server doesn't have a resource type "projects" logger.go:42: 13:15:54 | self-healing-chaos/4-read-data | +++ kubectl get nodes logger.go:42: 13:15:54 | self-healing-chaos/4-read-data | +++ grep '^minikube' logger.go:42: 13:15:55 | self-healing-chaos/4-read-data | ++++ pwd logger.go:42: 13:15:55 | self-healing-chaos/4-read-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/self-healing-chaos logger.go:42: 13:15:55 | self-healing-chaos/4-read-data | ++ test_name=self-healing-chaos logger.go:42: 13:15:55 | self-healing-chaos/4-read-data | ++ get_cluster_name logger.go:42: 13:15:55 | self-healing-chaos/4-read-data | ++ kubectl -n kuttl-test-set-bluejay get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 13:15:55 | self-healing-chaos/4-read-data | + cluster_name=self-healing-chaos logger.go:42: 13:15:55 | self-healing-chaos/4-read-data | + for i in 0 1 2 logger.go:42: 13:15:55 | self-healing-chaos/4-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h self-healing-chaos-mysql-0.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:15:55 | self-healing-chaos/4-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 13:15:55 | self-healing-chaos/4-read-data | ++ local 'uri=-h self-healing-chaos-mysql-0.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:15:55 | self-healing-chaos/4-read-data | ++ local pod= logger.go:42: 13:15:55 | self-healing-chaos/4-read-data | +++ get_client_pod logger.go:42: 13:15:55 | self-healing-chaos/4-read-data | +++ kubectl -n kuttl-test-set-bluejay get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:15:56 | self-healing-chaos/4-read-data | ++ client_pod=mysql-client logger.go:42: 13:15:56 | self-healing-chaos/4-read-data | ++ wait_pod mysql-client logger.go:42: 13:15:56 | self-healing-chaos/4-read-data | ++ local pod=mysql-client logger.go:42: 13:15:56 | self-healing-chaos/4-read-data | ++ set +o xtrace logger.go:42: 13:15:57 | self-healing-chaos/4-read-data | mysql-clienttrue logger.go:42: 13:15:57 | self-healing-chaos/4-read-data | ++ kubectl -n kuttl-test-set-bluejay exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h self-healing-chaos-mysql-0.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:15:57 | self-healing-chaos/4-read-data | ++ sed -e 's/mysql: //' logger.go:42: 13:15:57 | self-healing-chaos/4-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:15:58 | self-healing-chaos/4-read-data | + data=100500 logger.go:42: 13:15:58 | self-healing-chaos/4-read-data | + kubectl create configmap -n kuttl-test-set-bluejay 04-read-data-0 --from-literal=data=100500 logger.go:42: 13:15:59 | self-healing-chaos/4-read-data | configmap/04-read-data-0 created logger.go:42: 13:15:59 | self-healing-chaos/4-read-data | + for i in 0 1 2 logger.go:42: 13:15:59 | self-healing-chaos/4-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h self-healing-chaos-mysql-1.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:15:59 | self-healing-chaos/4-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 13:15:59 | self-healing-chaos/4-read-data | ++ local 'uri=-h self-healing-chaos-mysql-1.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:15:59 | self-healing-chaos/4-read-data | ++ local pod= logger.go:42: 13:15:59 | self-healing-chaos/4-read-data | +++ get_client_pod logger.go:42: 13:15:59 | self-healing-chaos/4-read-data | +++ kubectl -n kuttl-test-set-bluejay get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:16:00 | self-healing-chaos/4-read-data | ++ client_pod=mysql-client logger.go:42: 13:16:00 | self-healing-chaos/4-read-data | ++ wait_pod mysql-client logger.go:42: 13:16:00 | self-healing-chaos/4-read-data | ++ local pod=mysql-client logger.go:42: 13:16:00 | self-healing-chaos/4-read-data | ++ set +o xtrace logger.go:42: 13:16:01 | self-healing-chaos/4-read-data | mysql-clienttrue logger.go:42: 13:16:01 | self-healing-chaos/4-read-data | ++ kubectl -n kuttl-test-set-bluejay exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h self-healing-chaos-mysql-1.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:16:01 | self-healing-chaos/4-read-data | ++ sed -e 's/mysql: //' logger.go:42: 13:16:01 | self-healing-chaos/4-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:16:02 | self-healing-chaos/4-read-data | + data=100500 logger.go:42: 13:16:02 | self-healing-chaos/4-read-data | + kubectl create configmap -n kuttl-test-set-bluejay 04-read-data-1 --from-literal=data=100500 logger.go:42: 13:16:03 | self-healing-chaos/4-read-data | configmap/04-read-data-1 created logger.go:42: 13:16:03 | self-healing-chaos/4-read-data | + for i in 0 1 2 logger.go:42: 13:16:03 | self-healing-chaos/4-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h self-healing-chaos-mysql-2.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:16:03 | self-healing-chaos/4-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 13:16:03 | self-healing-chaos/4-read-data | ++ local 'uri=-h self-healing-chaos-mysql-2.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:16:03 | self-healing-chaos/4-read-data | ++ local pod= logger.go:42: 13:16:03 | self-healing-chaos/4-read-data | +++ get_client_pod logger.go:42: 13:16:03 | self-healing-chaos/4-read-data | +++ kubectl -n kuttl-test-set-bluejay get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:16:03 | self-healing-chaos/4-read-data | ++ client_pod=mysql-client logger.go:42: 13:16:03 | self-healing-chaos/4-read-data | ++ wait_pod mysql-client logger.go:42: 13:16:03 | self-healing-chaos/4-read-data | ++ local pod=mysql-client logger.go:42: 13:16:03 | self-healing-chaos/4-read-data | ++ set +o xtrace logger.go:42: 13:16:04 | self-healing-chaos/4-read-data | mysql-clienttrue logger.go:42: 13:16:04 | self-healing-chaos/4-read-data | ++ kubectl -n kuttl-test-set-bluejay exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h self-healing-chaos-mysql-2.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:16:04 | self-healing-chaos/4-read-data | ++ sed -e 's/mysql: //' logger.go:42: 13:16:04 | self-healing-chaos/4-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:16:06 | self-healing-chaos/4-read-data | + data=100500 logger.go:42: 13:16:06 | self-healing-chaos/4-read-data | + kubectl create configmap -n kuttl-test-set-bluejay 04-read-data-2 --from-literal=data=100500 logger.go:42: 13:16:06 | self-healing-chaos/4-read-data | configmap/04-read-data-2 created logger.go:42: 13:16:08 | self-healing-chaos/4-read-data | test step completed 4-read-data logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | starting test step 5-fail-pod logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | running command: [sh -c set -o errexit set -o xtrace source ../../functions pod=self-healing-chaos-mysql-0 yq eval '.metadata.name = "chaos-pod-failure" | del(.spec.selector.pods.test-namespace) | .metadata.namespace = "'${NAMESPACE}'" | .spec.selector.pods.'${NAMESPACE}'[0] = "'$pod'"' "${TESTS_CONFIG_DIR}/chaos-pod-failure.yml" \ | kubectl apply -f - run_mysql \ "INSERT myDB.myTable (id) VALUES (100501)" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password" sleep 60] logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | + source ../../functions logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ realpath ../../.. logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | ++++ mktemp -d logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ export TEMP_DIR=/tmp/tmp.BDpD69WEi5 logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ TEMP_DIR=/tmp/tmp.BDpD69WEi5 logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ export GIT_BRANCH=PR-339 logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ GIT_BRANCH=PR-339 logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ export VERSION=PR-339-3a8483b logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ VERSION=PR-339-3a8483b logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | ++++ which gdate logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | ++++ which date logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ date=/usr/bin/date logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ command -v oc logger.go:42: 13:16:08 | self-healing-chaos/5-fail-pod | +++ oc get projects logger.go:42: 13:16:14 | self-healing-chaos/5-fail-pod | error: the server doesn't have a resource type "projects" logger.go:42: 13:16:14 | self-healing-chaos/5-fail-pod | +++ kubectl get nodes logger.go:42: 13:16:14 | self-healing-chaos/5-fail-pod | +++ grep '^minikube' logger.go:42: 13:16:15 | self-healing-chaos/5-fail-pod | ++++ pwd logger.go:42: 13:16:15 | self-healing-chaos/5-fail-pod | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/self-healing-chaos logger.go:42: 13:16:15 | self-healing-chaos/5-fail-pod | ++ test_name=self-healing-chaos logger.go:42: 13:16:15 | self-healing-chaos/5-fail-pod | + pod=self-healing-chaos-mysql-0 logger.go:42: 13:16:15 | self-healing-chaos/5-fail-pod | + yq eval '.metadata.name = "chaos-pod-failure" | logger.go:42: 13:16:15 | self-healing-chaos/5-fail-pod | del(.spec.selector.pods.test-namespace) | logger.go:42: 13:16:15 | self-healing-chaos/5-fail-pod | .metadata.namespace = "kuttl-test-set-bluejay" | logger.go:42: 13:16:15 | self-healing-chaos/5-fail-pod | .spec.selector.pods.kuttl-test-set-bluejay[0] = "self-healing-chaos-mysql-0"' /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/chaos-pod-failure.yml logger.go:42: 13:16:15 | self-healing-chaos/5-fail-pod | + kubectl apply -f - logger.go:42: 13:16:16 | self-healing-chaos/5-fail-pod | podchaos.chaos-mesh.org/chaos-pod-failure created logger.go:42: 13:16:16 | self-healing-chaos/5-fail-pod | +++ get_cluster_name logger.go:42: 13:16:16 | self-healing-chaos/5-fail-pod | +++ kubectl -n kuttl-test-set-bluejay get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 13:16:17 | self-healing-chaos/5-fail-pod | ++ get_haproxy_svc self-healing-chaos logger.go:42: 13:16:17 | self-healing-chaos/5-fail-pod | ++ local cluster=self-healing-chaos logger.go:42: 13:16:17 | self-healing-chaos/5-fail-pod | ++ echo self-healing-chaos-haproxy logger.go:42: 13:16:17 | self-healing-chaos/5-fail-pod | + run_mysql 'INSERT myDB.myTable (id) VALUES (100501)' '-h self-healing-chaos-haproxy -uroot -proot_password' logger.go:42: 13:16:17 | self-healing-chaos/5-fail-pod | + local 'command=INSERT myDB.myTable (id) VALUES (100501)' logger.go:42: 13:16:17 | self-healing-chaos/5-fail-pod | + local 'uri=-h self-healing-chaos-haproxy -uroot -proot_password' logger.go:42: 13:16:17 | self-healing-chaos/5-fail-pod | + local pod= logger.go:42: 13:16:17 | self-healing-chaos/5-fail-pod | ++ get_client_pod logger.go:42: 13:16:17 | self-healing-chaos/5-fail-pod | ++ kubectl -n kuttl-test-set-bluejay get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:16:18 | self-healing-chaos/5-fail-pod | + client_pod=mysql-client logger.go:42: 13:16:18 | self-healing-chaos/5-fail-pod | + wait_pod mysql-client logger.go:42: 13:16:18 | self-healing-chaos/5-fail-pod | + local pod=mysql-client logger.go:42: 13:16:18 | self-healing-chaos/5-fail-pod | + set +o xtrace logger.go:42: 13:16:19 | self-healing-chaos/5-fail-pod | mysql-clienttrue logger.go:42: 13:16:19 | self-healing-chaos/5-fail-pod | + kubectl -n kuttl-test-set-bluejay exec mysql-client -- bash -c 'printf '\''%s\n'\'' "INSERT myDB.myTable (id) VALUES (100501)" | mysql -sN -h self-healing-chaos-haproxy -uroot -proot_password' logger.go:42: 13:16:19 | self-healing-chaos/5-fail-pod | + sed -e 's/mysql: //' logger.go:42: 13:16:19 | self-healing-chaos/5-fail-pod | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:16:20 | self-healing-chaos/5-fail-pod | + : logger.go:42: 13:16:20 | self-healing-chaos/5-fail-pod | + sleep 60 logger.go:42: 13:18:21 | self-healing-chaos/5-fail-pod | test step completed 5-fail-pod logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | starting test step 6-read-data logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions cluster_name=$(get_cluster_name) for i in 0 1 2; do data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 06-read-data-${i} --from-literal=data="${data}" done] logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | + source ../../functions logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ realpath ../../.. logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | ++++ mktemp -d logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ export TEMP_DIR=/tmp/tmp.FQ9GSeEhM2 logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ TEMP_DIR=/tmp/tmp.FQ9GSeEhM2 logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ export GIT_BRANCH=PR-339 logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ GIT_BRANCH=PR-339 logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ export VERSION=PR-339-3a8483b logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ VERSION=PR-339-3a8483b logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | ++++ which gdate logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | ++++ which date logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ date=/usr/bin/date logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ command -v oc logger.go:42: 13:18:21 | self-healing-chaos/6-read-data | +++ oc get projects logger.go:42: 13:18:26 | self-healing-chaos/6-read-data | error: the server doesn't have a resource type "projects" logger.go:42: 13:18:26 | self-healing-chaos/6-read-data | +++ kubectl get nodes logger.go:42: 13:18:26 | self-healing-chaos/6-read-data | +++ grep '^minikube' logger.go:42: 13:18:27 | self-healing-chaos/6-read-data | ++++ pwd logger.go:42: 13:18:27 | self-healing-chaos/6-read-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/self-healing-chaos logger.go:42: 13:18:27 | self-healing-chaos/6-read-data | ++ test_name=self-healing-chaos logger.go:42: 13:18:27 | self-healing-chaos/6-read-data | ++ get_cluster_name logger.go:42: 13:18:27 | self-healing-chaos/6-read-data | ++ kubectl -n kuttl-test-set-bluejay get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 13:18:28 | self-healing-chaos/6-read-data | + cluster_name=self-healing-chaos logger.go:42: 13:18:28 | self-healing-chaos/6-read-data | + for i in 0 1 2 logger.go:42: 13:18:28 | self-healing-chaos/6-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h self-healing-chaos-mysql-0.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:18:28 | self-healing-chaos/6-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 13:18:28 | self-healing-chaos/6-read-data | ++ local 'uri=-h self-healing-chaos-mysql-0.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:18:28 | self-healing-chaos/6-read-data | ++ local pod= logger.go:42: 13:18:28 | self-healing-chaos/6-read-data | +++ get_client_pod logger.go:42: 13:18:28 | self-healing-chaos/6-read-data | +++ kubectl -n kuttl-test-set-bluejay get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:18:29 | self-healing-chaos/6-read-data | ++ client_pod=mysql-client logger.go:42: 13:18:29 | self-healing-chaos/6-read-data | ++ wait_pod mysql-client logger.go:42: 13:18:29 | self-healing-chaos/6-read-data | ++ local pod=mysql-client logger.go:42: 13:18:29 | self-healing-chaos/6-read-data | ++ set +o xtrace logger.go:42: 13:18:30 | self-healing-chaos/6-read-data | mysql-clienttrue logger.go:42: 13:18:30 | self-healing-chaos/6-read-data | ++ kubectl -n kuttl-test-set-bluejay exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h self-healing-chaos-mysql-0.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:18:30 | self-healing-chaos/6-read-data | ++ sed -e 's/mysql: //' logger.go:42: 13:18:30 | self-healing-chaos/6-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:18:31 | self-healing-chaos/6-read-data | + data='100500 logger.go:42: 13:18:31 | self-healing-chaos/6-read-data | 100501' logger.go:42: 13:18:31 | self-healing-chaos/6-read-data | + kubectl create configmap -n kuttl-test-set-bluejay 06-read-data-0 '--from-literal=data=100500 logger.go:42: 13:18:31 | self-healing-chaos/6-read-data | 100501' logger.go:42: 13:18:32 | self-healing-chaos/6-read-data | configmap/06-read-data-0 created logger.go:42: 13:18:32 | self-healing-chaos/6-read-data | + for i in 0 1 2 logger.go:42: 13:18:32 | self-healing-chaos/6-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h self-healing-chaos-mysql-1.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:18:32 | self-healing-chaos/6-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 13:18:32 | self-healing-chaos/6-read-data | ++ local 'uri=-h self-healing-chaos-mysql-1.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:18:32 | self-healing-chaos/6-read-data | ++ local pod= logger.go:42: 13:18:32 | self-healing-chaos/6-read-data | +++ get_client_pod logger.go:42: 13:18:32 | self-healing-chaos/6-read-data | +++ kubectl -n kuttl-test-set-bluejay get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:18:33 | self-healing-chaos/6-read-data | ++ client_pod=mysql-client logger.go:42: 13:18:33 | self-healing-chaos/6-read-data | ++ wait_pod mysql-client logger.go:42: 13:18:33 | self-healing-chaos/6-read-data | ++ local pod=mysql-client logger.go:42: 13:18:33 | self-healing-chaos/6-read-data | ++ set +o xtrace logger.go:42: 13:18:33 | self-healing-chaos/6-read-data | mysql-clienttrue logger.go:42: 13:18:33 | self-healing-chaos/6-read-data | ++ kubectl -n kuttl-test-set-bluejay exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h self-healing-chaos-mysql-1.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:18:33 | self-healing-chaos/6-read-data | ++ sed -e 's/mysql: //' logger.go:42: 13:18:33 | self-healing-chaos/6-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:18:35 | self-healing-chaos/6-read-data | + data='100500 logger.go:42: 13:18:35 | self-healing-chaos/6-read-data | 100501' logger.go:42: 13:18:35 | self-healing-chaos/6-read-data | + kubectl create configmap -n kuttl-test-set-bluejay 06-read-data-1 '--from-literal=data=100500 logger.go:42: 13:18:35 | self-healing-chaos/6-read-data | 100501' logger.go:42: 13:18:36 | self-healing-chaos/6-read-data | configmap/06-read-data-1 created logger.go:42: 13:18:36 | self-healing-chaos/6-read-data | + for i in 0 1 2 logger.go:42: 13:18:36 | self-healing-chaos/6-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h self-healing-chaos-mysql-2.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:18:36 | self-healing-chaos/6-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 13:18:36 | self-healing-chaos/6-read-data | ++ local 'uri=-h self-healing-chaos-mysql-2.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:18:36 | self-healing-chaos/6-read-data | ++ local pod= logger.go:42: 13:18:36 | self-healing-chaos/6-read-data | +++ get_client_pod logger.go:42: 13:18:36 | self-healing-chaos/6-read-data | +++ kubectl -n kuttl-test-set-bluejay get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:18:36 | self-healing-chaos/6-read-data | ++ client_pod=mysql-client logger.go:42: 13:18:36 | self-healing-chaos/6-read-data | ++ wait_pod mysql-client logger.go:42: 13:18:36 | self-healing-chaos/6-read-data | ++ local pod=mysql-client logger.go:42: 13:18:36 | self-healing-chaos/6-read-data | ++ set +o xtrace logger.go:42: 13:18:37 | self-healing-chaos/6-read-data | mysql-clienttrue logger.go:42: 13:18:37 | self-healing-chaos/6-read-data | ++ kubectl -n kuttl-test-set-bluejay exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h self-healing-chaos-mysql-2.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:18:37 | self-healing-chaos/6-read-data | ++ sed -e 's/mysql: //' logger.go:42: 13:18:37 | self-healing-chaos/6-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:18:39 | self-healing-chaos/6-read-data | + data='100500 logger.go:42: 13:18:39 | self-healing-chaos/6-read-data | 100501' logger.go:42: 13:18:39 | self-healing-chaos/6-read-data | + kubectl create configmap -n kuttl-test-set-bluejay 06-read-data-2 '--from-literal=data=100500 logger.go:42: 13:18:39 | self-healing-chaos/6-read-data | 100501' logger.go:42: 13:18:39 | self-healing-chaos/6-read-data | configmap/06-read-data-2 created logger.go:42: 13:18:41 | self-healing-chaos/6-read-data | test step completed 6-read-data logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | starting test step 7-network-loss logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | running command: [sh -c set -o errexit set -o xtrace source ../../functions pod=self-healing-chaos-mysql-0 yq eval '.metadata.name = "chaos-pod-network-loss" | del(.spec.selector.pods.test-namespace) | .metadata.namespace = "'${NAMESPACE}'" | .spec.selector.pods.'${NAMESPACE}'[0] = "'$pod'"' "${TESTS_CONFIG_DIR}/chaos-network-loss.yml" \ | kubectl apply -f - sleep 30 run_mysql \ "INSERT myDB.myTable (id) VALUES (100502)" \ "-h $(get_haproxy_svc $(get_cluster_name)) -uroot -proot_password" sleep 30] logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | + source ../../functions logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ realpath ../../.. logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | ++++ mktemp -d logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ export TEMP_DIR=/tmp/tmp.TYj6rJJVRY logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ TEMP_DIR=/tmp/tmp.TYj6rJJVRY logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ export GIT_BRANCH=PR-339 logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ GIT_BRANCH=PR-339 logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ export VERSION=PR-339-3a8483b logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ VERSION=PR-339-3a8483b logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | ++++ which gdate logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | ++++ which date logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ date=/usr/bin/date logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ command -v oc logger.go:42: 13:18:41 | self-healing-chaos/7-network-loss | +++ oc get projects logger.go:42: 13:18:47 | self-healing-chaos/7-network-loss | error: the server doesn't have a resource type "projects" logger.go:42: 13:18:47 | self-healing-chaos/7-network-loss | +++ kubectl get nodes logger.go:42: 13:18:47 | self-healing-chaos/7-network-loss | +++ grep '^minikube' logger.go:42: 13:18:48 | self-healing-chaos/7-network-loss | ++++ pwd logger.go:42: 13:18:48 | self-healing-chaos/7-network-loss | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/self-healing-chaos logger.go:42: 13:18:48 | self-healing-chaos/7-network-loss | ++ test_name=self-healing-chaos logger.go:42: 13:18:48 | self-healing-chaos/7-network-loss | + pod=self-healing-chaos-mysql-0 logger.go:42: 13:18:48 | self-healing-chaos/7-network-loss | + yq eval '.metadata.name = "chaos-pod-network-loss" | logger.go:42: 13:18:48 | self-healing-chaos/7-network-loss | del(.spec.selector.pods.test-namespace) | logger.go:42: 13:18:48 | self-healing-chaos/7-network-loss | .metadata.namespace = "kuttl-test-set-bluejay" | logger.go:42: 13:18:48 | self-healing-chaos/7-network-loss | .spec.selector.pods.kuttl-test-set-bluejay[0] = "self-healing-chaos-mysql-0"' /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/chaos-network-loss.yml logger.go:42: 13:18:48 | self-healing-chaos/7-network-loss | + kubectl apply -f - logger.go:42: 13:18:49 | self-healing-chaos/7-network-loss | networkchaos.chaos-mesh.org/chaos-pod-network-loss created logger.go:42: 13:18:49 | self-healing-chaos/7-network-loss | + sleep 30 logger.go:42: 13:19:19 | self-healing-chaos/7-network-loss | +++ get_cluster_name logger.go:42: 13:19:19 | self-healing-chaos/7-network-loss | +++ kubectl -n kuttl-test-set-bluejay get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 13:19:20 | self-healing-chaos/7-network-loss | ++ get_haproxy_svc self-healing-chaos logger.go:42: 13:19:20 | self-healing-chaos/7-network-loss | ++ local cluster=self-healing-chaos logger.go:42: 13:19:20 | self-healing-chaos/7-network-loss | ++ echo self-healing-chaos-haproxy logger.go:42: 13:19:20 | self-healing-chaos/7-network-loss | + run_mysql 'INSERT myDB.myTable (id) VALUES (100502)' '-h self-healing-chaos-haproxy -uroot -proot_password' logger.go:42: 13:19:20 | self-healing-chaos/7-network-loss | + local 'command=INSERT myDB.myTable (id) VALUES (100502)' logger.go:42: 13:19:20 | self-healing-chaos/7-network-loss | + local 'uri=-h self-healing-chaos-haproxy -uroot -proot_password' logger.go:42: 13:19:20 | self-healing-chaos/7-network-loss | + local pod= logger.go:42: 13:19:20 | self-healing-chaos/7-network-loss | ++ get_client_pod logger.go:42: 13:19:20 | self-healing-chaos/7-network-loss | ++ kubectl -n kuttl-test-set-bluejay get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:19:21 | self-healing-chaos/7-network-loss | + client_pod=mysql-client logger.go:42: 13:19:21 | self-healing-chaos/7-network-loss | + wait_pod mysql-client logger.go:42: 13:19:21 | self-healing-chaos/7-network-loss | + local pod=mysql-client logger.go:42: 13:19:21 | self-healing-chaos/7-network-loss | + set +o xtrace logger.go:42: 13:19:22 | self-healing-chaos/7-network-loss | mysql-clienttrue logger.go:42: 13:19:22 | self-healing-chaos/7-network-loss | + sed -e 's/mysql: //' logger.go:42: 13:19:22 | self-healing-chaos/7-network-loss | + grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:19:22 | self-healing-chaos/7-network-loss | + kubectl -n kuttl-test-set-bluejay exec mysql-client -- bash -c 'printf '\''%s\n'\'' "INSERT myDB.myTable (id) VALUES (100502)" | mysql -sN -h self-healing-chaos-haproxy -uroot -proot_password' logger.go:42: 13:19:23 | self-healing-chaos/7-network-loss | + : logger.go:42: 13:19:23 | self-healing-chaos/7-network-loss | + sleep 30 logger.go:42: 13:19:55 | self-healing-chaos/7-network-loss | test step completed 7-network-loss logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | starting test step 8-read-data logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions cluster_name=$(get_cluster_name) for i in 0 1 2; do data=$(run_mysql "SELECT * FROM myDB.myTable" "-h ${cluster_name}-mysql-${i}.${cluster_name}-mysql -uroot -proot_password") kubectl create configmap -n "${NAMESPACE}" 08-read-data-${i} --from-literal=data="${data}" done] logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | + source ../../functions logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ realpath ../../.. logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | ++++ mktemp -d logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ export TEMP_DIR=/tmp/tmp.Wj7rzopR85 logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ TEMP_DIR=/tmp/tmp.Wj7rzopR85 logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ export GIT_BRANCH=PR-339 logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ GIT_BRANCH=PR-339 logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ export VERSION=PR-339-3a8483b logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ VERSION=PR-339-3a8483b logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | ++++ which gdate logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | ++++ which date logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ date=/usr/bin/date logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ command -v oc logger.go:42: 13:19:55 | self-healing-chaos/8-read-data | +++ oc get projects logger.go:42: 13:20:01 | self-healing-chaos/8-read-data | error: the server doesn't have a resource type "projects" logger.go:42: 13:20:01 | self-healing-chaos/8-read-data | +++ kubectl get nodes logger.go:42: 13:20:01 | self-healing-chaos/8-read-data | +++ grep '^minikube' logger.go:42: 13:20:02 | self-healing-chaos/8-read-data | ++++ pwd logger.go:42: 13:20:02 | self-healing-chaos/8-read-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/self-healing-chaos logger.go:42: 13:20:02 | self-healing-chaos/8-read-data | ++ test_name=self-healing-chaos logger.go:42: 13:20:02 | self-healing-chaos/8-read-data | ++ get_cluster_name logger.go:42: 13:20:02 | self-healing-chaos/8-read-data | ++ kubectl -n kuttl-test-set-bluejay get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 13:20:03 | self-healing-chaos/8-read-data | + cluster_name=self-healing-chaos logger.go:42: 13:20:03 | self-healing-chaos/8-read-data | + for i in 0 1 2 logger.go:42: 13:20:03 | self-healing-chaos/8-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h self-healing-chaos-mysql-0.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:20:03 | self-healing-chaos/8-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 13:20:03 | self-healing-chaos/8-read-data | ++ local 'uri=-h self-healing-chaos-mysql-0.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:20:03 | self-healing-chaos/8-read-data | ++ local pod= logger.go:42: 13:20:03 | self-healing-chaos/8-read-data | +++ get_client_pod logger.go:42: 13:20:03 | self-healing-chaos/8-read-data | +++ kubectl -n kuttl-test-set-bluejay get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:20:03 | self-healing-chaos/8-read-data | ++ client_pod=mysql-client logger.go:42: 13:20:03 | self-healing-chaos/8-read-data | ++ wait_pod mysql-client logger.go:42: 13:20:03 | self-healing-chaos/8-read-data | ++ local pod=mysql-client logger.go:42: 13:20:03 | self-healing-chaos/8-read-data | ++ set +o xtrace logger.go:42: 13:20:04 | self-healing-chaos/8-read-data | mysql-clienttrue logger.go:42: 13:20:04 | self-healing-chaos/8-read-data | ++ kubectl -n kuttl-test-set-bluejay exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h self-healing-chaos-mysql-0.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:20:04 | self-healing-chaos/8-read-data | ++ sed -e 's/mysql: //' logger.go:42: 13:20:04 | self-healing-chaos/8-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:20:06 | self-healing-chaos/8-read-data | + data='100500 logger.go:42: 13:20:06 | self-healing-chaos/8-read-data | 100501 logger.go:42: 13:20:06 | self-healing-chaos/8-read-data | 100502' logger.go:42: 13:20:06 | self-healing-chaos/8-read-data | + kubectl create configmap -n kuttl-test-set-bluejay 08-read-data-0 '--from-literal=data=100500 logger.go:42: 13:20:06 | self-healing-chaos/8-read-data | 100501 logger.go:42: 13:20:06 | self-healing-chaos/8-read-data | 100502' logger.go:42: 13:20:06 | self-healing-chaos/8-read-data | configmap/08-read-data-0 created logger.go:42: 13:20:06 | self-healing-chaos/8-read-data | + for i in 0 1 2 logger.go:42: 13:20:06 | self-healing-chaos/8-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h self-healing-chaos-mysql-1.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:20:06 | self-healing-chaos/8-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 13:20:06 | self-healing-chaos/8-read-data | ++ local 'uri=-h self-healing-chaos-mysql-1.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:20:06 | self-healing-chaos/8-read-data | ++ local pod= logger.go:42: 13:20:06 | self-healing-chaos/8-read-data | +++ get_client_pod logger.go:42: 13:20:06 | self-healing-chaos/8-read-data | +++ kubectl -n kuttl-test-set-bluejay get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:20:07 | self-healing-chaos/8-read-data | ++ client_pod=mysql-client logger.go:42: 13:20:07 | self-healing-chaos/8-read-data | ++ wait_pod mysql-client logger.go:42: 13:20:07 | self-healing-chaos/8-read-data | ++ local pod=mysql-client logger.go:42: 13:20:07 | self-healing-chaos/8-read-data | ++ set +o xtrace logger.go:42: 13:20:08 | self-healing-chaos/8-read-data | mysql-clienttrue logger.go:42: 13:20:08 | self-healing-chaos/8-read-data | ++ kubectl -n kuttl-test-set-bluejay exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h self-healing-chaos-mysql-1.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:20:08 | self-healing-chaos/8-read-data | ++ sed -e 's/mysql: //' logger.go:42: 13:20:08 | self-healing-chaos/8-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:20:09 | self-healing-chaos/8-read-data | + data='100500 logger.go:42: 13:20:09 | self-healing-chaos/8-read-data | 100501 logger.go:42: 13:20:09 | self-healing-chaos/8-read-data | 100502' logger.go:42: 13:20:09 | self-healing-chaos/8-read-data | + kubectl create configmap -n kuttl-test-set-bluejay 08-read-data-1 '--from-literal=data=100500 logger.go:42: 13:20:09 | self-healing-chaos/8-read-data | 100501 logger.go:42: 13:20:09 | self-healing-chaos/8-read-data | 100502' logger.go:42: 13:20:10 | self-healing-chaos/8-read-data | configmap/08-read-data-1 created logger.go:42: 13:20:10 | self-healing-chaos/8-read-data | + for i in 0 1 2 logger.go:42: 13:20:10 | self-healing-chaos/8-read-data | ++ run_mysql 'SELECT * FROM myDB.myTable' '-h self-healing-chaos-mysql-2.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:20:10 | self-healing-chaos/8-read-data | ++ local 'command=SELECT * FROM myDB.myTable' logger.go:42: 13:20:10 | self-healing-chaos/8-read-data | ++ local 'uri=-h self-healing-chaos-mysql-2.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:20:10 | self-healing-chaos/8-read-data | ++ local pod= logger.go:42: 13:20:10 | self-healing-chaos/8-read-data | +++ get_client_pod logger.go:42: 13:20:10 | self-healing-chaos/8-read-data | +++ kubectl -n kuttl-test-set-bluejay get pods --selector=name=mysql-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 13:20:11 | self-healing-chaos/8-read-data | ++ client_pod=mysql-client logger.go:42: 13:20:11 | self-healing-chaos/8-read-data | ++ wait_pod mysql-client logger.go:42: 13:20:11 | self-healing-chaos/8-read-data | ++ local pod=mysql-client logger.go:42: 13:20:11 | self-healing-chaos/8-read-data | ++ set +o xtrace logger.go:42: 13:20:11 | self-healing-chaos/8-read-data | mysql-clienttrue logger.go:42: 13:20:11 | self-healing-chaos/8-read-data | ++ kubectl -n kuttl-test-set-bluejay exec mysql-client -- bash -c 'printf '\''%s\n'\'' "SELECT * FROM myDB.myTable" | mysql -sN -h self-healing-chaos-mysql-2.self-healing-chaos-mysql -uroot -proot_password' logger.go:42: 13:20:11 | self-healing-chaos/8-read-data | ++ sed -e 's/mysql: //' logger.go:42: 13:20:11 | self-healing-chaos/8-read-data | ++ grep -v 'Using a password on the command line interface can be insecure.' logger.go:42: 13:20:13 | self-healing-chaos/8-read-data | + data='100500 logger.go:42: 13:20:13 | self-healing-chaos/8-read-data | 100501 logger.go:42: 13:20:13 | self-healing-chaos/8-read-data | 100502' logger.go:42: 13:20:13 | self-healing-chaos/8-read-data | + kubectl create configmap -n kuttl-test-set-bluejay 08-read-data-2 '--from-literal=data=100500 logger.go:42: 13:20:13 | self-healing-chaos/8-read-data | 100501 logger.go:42: 13:20:13 | self-healing-chaos/8-read-data | 100502' logger.go:42: 13:20:14 | self-healing-chaos/8-read-data | configmap/08-read-data-2 created logger.go:42: 13:20:15 | self-healing-chaos/8-read-data | test step completed 8-read-data logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | starting test step 9-destroy-chaos-mesh logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | running command: [sh -c set -o errexit set -o xtrace source ../../functions destroy_chaos_mesh] logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | + source ../../functions logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ realpath ../../.. logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | ++++ mktemp -d logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ export TEMP_DIR=/tmp/tmp.JP1wcEW3Tk logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ TEMP_DIR=/tmp/tmp.JP1wcEW3Tk logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ export GIT_BRANCH=PR-339 logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ GIT_BRANCH=PR-339 logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ export VERSION=PR-339-3a8483b logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ VERSION=PR-339-3a8483b logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | ++++ which gdate logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | ++++ which date logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ date=/usr/bin/date logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ command -v oc logger.go:42: 13:20:15 | self-healing-chaos/9-destroy-chaos-mesh | +++ oc get projects logger.go:42: 13:20:21 | self-healing-chaos/9-destroy-chaos-mesh | error: the server doesn't have a resource type "projects" logger.go:42: 13:20:21 | self-healing-chaos/9-destroy-chaos-mesh | +++ grep '^minikube' logger.go:42: 13:20:21 | self-healing-chaos/9-destroy-chaos-mesh | +++ kubectl get nodes logger.go:42: 13:20:22 | self-healing-chaos/9-destroy-chaos-mesh | ++++ pwd logger.go:42: 13:20:22 | self-healing-chaos/9-destroy-chaos-mesh | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/self-healing-chaos logger.go:42: 13:20:22 | self-healing-chaos/9-destroy-chaos-mesh | ++ test_name=self-healing-chaos logger.go:42: 13:20:22 | self-healing-chaos/9-destroy-chaos-mesh | + destroy_chaos_mesh logger.go:42: 13:20:22 | self-healing-chaos/9-destroy-chaos-mesh | + local chaos_mesh_ns logger.go:42: 13:20:22 | self-healing-chaos/9-destroy-chaos-mesh | ++ helm list --all-namespaces --filter chaos-mesh logger.go:42: 13:20:22 | self-healing-chaos/9-destroy-chaos-mesh | ++ tail -n1 logger.go:42: 13:20:22 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '-F ' '{print $2}' logger.go:42: 13:20:22 | self-healing-chaos/9-destroy-chaos-mesh | ++ sed s/NAMESPACE// logger.go:42: 13:20:22 | self-healing-chaos/9-destroy-chaos-mesh | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 13:20:22 | self-healing-chaos/9-destroy-chaos-mesh | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 13:20:23 | self-healing-chaos/9-destroy-chaos-mesh | + chaos_mesh_ns=kuttl-test-set-bluejay logger.go:42: 13:20:23 | self-healing-chaos/9-destroy-chaos-mesh | + '[' -n kuttl-test-set-bluejay ']' logger.go:42: 13:20:23 | self-healing-chaos/9-destroy-chaos-mesh | + helm uninstall chaos-mesh --namespace kuttl-test-set-bluejay logger.go:42: 13:20:23 | self-healing-chaos/9-destroy-chaos-mesh | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 13:20:23 | self-healing-chaos/9-destroy-chaos-mesh | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 13:20:28 | self-healing-chaos/9-destroy-chaos-mesh | release "chaos-mesh" uninstalled logger.go:42: 13:20:28 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl api-resources logger.go:42: 13:20:28 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep chaos-mesh logger.go:42: 13:20:28 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:20:30 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 13:20:30 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer awschaos --all logger.go:42: 13:20:30 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=awschaos logger.go:42: 13:20:30 | self-healing-chaos/9-destroy-chaos-mesh | + local name=--all logger.go:42: 13:20:30 | self-healing-chaos/9-destroy-chaos-mesh | + '[' --all == --all ']' logger.go:42: 13:20:30 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get awschaos logger.go:42: 13:20:30 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep -v NAME logger.go:42: 13:20:30 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:20:30 | self-healing-chaos/9-destroy-chaos-mesh | No resources found in default namespace. logger.go:42: 13:20:30 | self-healing-chaos/9-destroy-chaos-mesh | + return logger.go:42: 13:20:30 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 13:20:30 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer azurechaos --all logger.go:42: 13:20:30 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=azurechaos logger.go:42: 13:20:30 | self-healing-chaos/9-destroy-chaos-mesh | + local name=--all logger.go:42: 13:20:30 | self-healing-chaos/9-destroy-chaos-mesh | + '[' --all == --all ']' logger.go:42: 13:20:30 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get azurechaos logger.go:42: 13:20:30 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep -v NAME logger.go:42: 13:20:30 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:20:31 | self-healing-chaos/9-destroy-chaos-mesh | No resources found in default namespace. logger.go:42: 13:20:31 | self-healing-chaos/9-destroy-chaos-mesh | + return logger.go:42: 13:20:31 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 13:20:31 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer blockchaos --all logger.go:42: 13:20:31 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=blockchaos logger.go:42: 13:20:31 | self-healing-chaos/9-destroy-chaos-mesh | + local name=--all logger.go:42: 13:20:31 | self-healing-chaos/9-destroy-chaos-mesh | + '[' --all == --all ']' logger.go:42: 13:20:31 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get blockchaos logger.go:42: 13:20:31 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep -v NAME logger.go:42: 13:20:31 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:20:32 | self-healing-chaos/9-destroy-chaos-mesh | No resources found in default namespace. logger.go:42: 13:20:32 | self-healing-chaos/9-destroy-chaos-mesh | + return logger.go:42: 13:20:32 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 13:20:32 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer dnschaos --all logger.go:42: 13:20:32 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=dnschaos logger.go:42: 13:20:32 | self-healing-chaos/9-destroy-chaos-mesh | + local name=--all logger.go:42: 13:20:32 | self-healing-chaos/9-destroy-chaos-mesh | + '[' --all == --all ']' logger.go:42: 13:20:32 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get dnschaos logger.go:42: 13:20:32 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep -v NAME logger.go:42: 13:20:32 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:20:33 | self-healing-chaos/9-destroy-chaos-mesh | No resources found in default namespace. logger.go:42: 13:20:33 | self-healing-chaos/9-destroy-chaos-mesh | + return logger.go:42: 13:20:33 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 13:20:33 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer gcpchaos --all logger.go:42: 13:20:33 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=gcpchaos logger.go:42: 13:20:33 | self-healing-chaos/9-destroy-chaos-mesh | + local name=--all logger.go:42: 13:20:33 | self-healing-chaos/9-destroy-chaos-mesh | + '[' --all == --all ']' logger.go:42: 13:20:33 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get gcpchaos logger.go:42: 13:20:33 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep -v NAME logger.go:42: 13:20:33 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:20:33 | self-healing-chaos/9-destroy-chaos-mesh | No resources found in default namespace. logger.go:42: 13:20:33 | self-healing-chaos/9-destroy-chaos-mesh | + return logger.go:42: 13:20:33 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 13:20:33 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer httpchaos --all logger.go:42: 13:20:33 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=httpchaos logger.go:42: 13:20:33 | self-healing-chaos/9-destroy-chaos-mesh | + local name=--all logger.go:42: 13:20:33 | self-healing-chaos/9-destroy-chaos-mesh | + '[' --all == --all ']' logger.go:42: 13:20:33 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get httpchaos logger.go:42: 13:20:33 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep -v NAME logger.go:42: 13:20:33 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:20:34 | self-healing-chaos/9-destroy-chaos-mesh | No resources found in default namespace. logger.go:42: 13:20:34 | self-healing-chaos/9-destroy-chaos-mesh | + return logger.go:42: 13:20:34 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 13:20:34 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer iochaos --all logger.go:42: 13:20:34 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=iochaos logger.go:42: 13:20:34 | self-healing-chaos/9-destroy-chaos-mesh | + local name=--all logger.go:42: 13:20:34 | self-healing-chaos/9-destroy-chaos-mesh | + '[' --all == --all ']' logger.go:42: 13:20:34 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get iochaos logger.go:42: 13:20:34 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep -v NAME logger.go:42: 13:20:34 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:20:35 | self-healing-chaos/9-destroy-chaos-mesh | No resources found in default namespace. logger.go:42: 13:20:35 | self-healing-chaos/9-destroy-chaos-mesh | + return logger.go:42: 13:20:35 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 13:20:35 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer jvmchaos --all logger.go:42: 13:20:35 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=jvmchaos logger.go:42: 13:20:35 | self-healing-chaos/9-destroy-chaos-mesh | + local name=--all logger.go:42: 13:20:35 | self-healing-chaos/9-destroy-chaos-mesh | + '[' --all == --all ']' logger.go:42: 13:20:35 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get jvmchaos logger.go:42: 13:20:35 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep -v NAME logger.go:42: 13:20:35 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:20:35 | self-healing-chaos/9-destroy-chaos-mesh | No resources found in default namespace. logger.go:42: 13:20:35 | self-healing-chaos/9-destroy-chaos-mesh | + return logger.go:42: 13:20:35 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 13:20:35 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer kernelchaos --all logger.go:42: 13:20:35 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=kernelchaos logger.go:42: 13:20:35 | self-healing-chaos/9-destroy-chaos-mesh | + local name=--all logger.go:42: 13:20:35 | self-healing-chaos/9-destroy-chaos-mesh | + '[' --all == --all ']' logger.go:42: 13:20:35 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get kernelchaos logger.go:42: 13:20:35 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep -v NAME logger.go:42: 13:20:35 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:20:36 | self-healing-chaos/9-destroy-chaos-mesh | No resources found in default namespace. logger.go:42: 13:20:36 | self-healing-chaos/9-destroy-chaos-mesh | + return logger.go:42: 13:20:36 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 13:20:36 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer networkchaos --all logger.go:42: 13:20:36 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=networkchaos logger.go:42: 13:20:36 | self-healing-chaos/9-destroy-chaos-mesh | + local name=--all logger.go:42: 13:20:36 | self-healing-chaos/9-destroy-chaos-mesh | + '[' --all == --all ']' logger.go:42: 13:20:36 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get networkchaos logger.go:42: 13:20:36 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep -v NAME logger.go:42: 13:20:36 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:20:37 | self-healing-chaos/9-destroy-chaos-mesh | No resources found in default namespace. logger.go:42: 13:20:37 | self-healing-chaos/9-destroy-chaos-mesh | + return logger.go:42: 13:20:37 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 13:20:37 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer physicalmachinechaos --all logger.go:42: 13:20:37 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=physicalmachinechaos logger.go:42: 13:20:37 | self-healing-chaos/9-destroy-chaos-mesh | + local name=--all logger.go:42: 13:20:37 | self-healing-chaos/9-destroy-chaos-mesh | + '[' --all == --all ']' logger.go:42: 13:20:37 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get physicalmachinechaos logger.go:42: 13:20:37 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep -v NAME logger.go:42: 13:20:37 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:20:38 | self-healing-chaos/9-destroy-chaos-mesh | No resources found in default namespace. logger.go:42: 13:20:38 | self-healing-chaos/9-destroy-chaos-mesh | + return logger.go:42: 13:20:38 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 13:20:38 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer physicalmachines --all logger.go:42: 13:20:38 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=physicalmachines logger.go:42: 13:20:38 | self-healing-chaos/9-destroy-chaos-mesh | + local name=--all logger.go:42: 13:20:38 | self-healing-chaos/9-destroy-chaos-mesh | + '[' --all == --all ']' logger.go:42: 13:20:38 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get physicalmachines logger.go:42: 13:20:38 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep -v NAME logger.go:42: 13:20:38 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:20:38 | self-healing-chaos/9-destroy-chaos-mesh | No resources found in default namespace. logger.go:42: 13:20:38 | self-healing-chaos/9-destroy-chaos-mesh | + return logger.go:42: 13:20:38 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 13:20:38 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer podchaos --all logger.go:42: 13:20:38 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=podchaos logger.go:42: 13:20:38 | self-healing-chaos/9-destroy-chaos-mesh | + local name=--all logger.go:42: 13:20:38 | self-healing-chaos/9-destroy-chaos-mesh | + '[' --all == --all ']' logger.go:42: 13:20:38 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get podchaos logger.go:42: 13:20:38 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep -v NAME logger.go:42: 13:20:38 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:20:39 | self-healing-chaos/9-destroy-chaos-mesh | No resources found in default namespace. logger.go:42: 13:20:39 | self-healing-chaos/9-destroy-chaos-mesh | + return logger.go:42: 13:20:39 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 13:20:39 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer podhttpchaos --all logger.go:42: 13:20:39 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=podhttpchaos logger.go:42: 13:20:39 | self-healing-chaos/9-destroy-chaos-mesh | + local name=--all logger.go:42: 13:20:39 | self-healing-chaos/9-destroy-chaos-mesh | + '[' --all == --all ']' logger.go:42: 13:20:39 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get podhttpchaos logger.go:42: 13:20:39 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep -v NAME logger.go:42: 13:20:39 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:20:40 | self-healing-chaos/9-destroy-chaos-mesh | No resources found in default namespace. logger.go:42: 13:20:40 | self-healing-chaos/9-destroy-chaos-mesh | + return logger.go:42: 13:20:40 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 13:20:40 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer podiochaos --all logger.go:42: 13:20:40 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=podiochaos logger.go:42: 13:20:40 | self-healing-chaos/9-destroy-chaos-mesh | + local name=--all logger.go:42: 13:20:40 | self-healing-chaos/9-destroy-chaos-mesh | + '[' --all == --all ']' logger.go:42: 13:20:40 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get podiochaos logger.go:42: 13:20:40 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep -v NAME logger.go:42: 13:20:40 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:20:41 | self-healing-chaos/9-destroy-chaos-mesh | No resources found in default namespace. logger.go:42: 13:20:41 | self-healing-chaos/9-destroy-chaos-mesh | + return logger.go:42: 13:20:41 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 13:20:41 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer podnetworkchaos --all logger.go:42: 13:20:41 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=podnetworkchaos logger.go:42: 13:20:41 | self-healing-chaos/9-destroy-chaos-mesh | + local name=--all logger.go:42: 13:20:41 | self-healing-chaos/9-destroy-chaos-mesh | + '[' --all == --all ']' logger.go:42: 13:20:41 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get podnetworkchaos logger.go:42: 13:20:41 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep -v NAME logger.go:42: 13:20:41 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:20:41 | self-healing-chaos/9-destroy-chaos-mesh | No resources found in default namespace. logger.go:42: 13:20:41 | self-healing-chaos/9-destroy-chaos-mesh | + return logger.go:42: 13:20:41 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 13:20:41 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer remoteclusters --all logger.go:42: 13:20:41 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=remoteclusters logger.go:42: 13:20:41 | self-healing-chaos/9-destroy-chaos-mesh | + local name=--all logger.go:42: 13:20:41 | self-healing-chaos/9-destroy-chaos-mesh | + '[' --all == --all ']' logger.go:42: 13:20:41 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get remoteclusters logger.go:42: 13:20:41 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep -v NAME logger.go:42: 13:20:41 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:20:42 | self-healing-chaos/9-destroy-chaos-mesh | No resources found logger.go:42: 13:20:42 | self-healing-chaos/9-destroy-chaos-mesh | + return logger.go:42: 13:20:42 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 13:20:42 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer schedules --all logger.go:42: 13:20:42 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=schedules logger.go:42: 13:20:42 | self-healing-chaos/9-destroy-chaos-mesh | + local name=--all logger.go:42: 13:20:42 | self-healing-chaos/9-destroy-chaos-mesh | + '[' --all == --all ']' logger.go:42: 13:20:42 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get schedules logger.go:42: 13:20:42 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep -v NAME logger.go:42: 13:20:42 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:20:43 | self-healing-chaos/9-destroy-chaos-mesh | No resources found in default namespace. logger.go:42: 13:20:43 | self-healing-chaos/9-destroy-chaos-mesh | + return logger.go:42: 13:20:43 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 13:20:43 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer statuschecks --all logger.go:42: 13:20:43 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=statuschecks logger.go:42: 13:20:43 | self-healing-chaos/9-destroy-chaos-mesh | + local name=--all logger.go:42: 13:20:43 | self-healing-chaos/9-destroy-chaos-mesh | + '[' --all == --all ']' logger.go:42: 13:20:43 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get statuschecks logger.go:42: 13:20:43 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep -v NAME logger.go:42: 13:20:43 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:20:44 | self-healing-chaos/9-destroy-chaos-mesh | No resources found in default namespace. logger.go:42: 13:20:44 | self-healing-chaos/9-destroy-chaos-mesh | + return logger.go:42: 13:20:44 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 13:20:44 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer stresschaos --all logger.go:42: 13:20:44 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=stresschaos logger.go:42: 13:20:44 | self-healing-chaos/9-destroy-chaos-mesh | + local name=--all logger.go:42: 13:20:44 | self-healing-chaos/9-destroy-chaos-mesh | + '[' --all == --all ']' logger.go:42: 13:20:44 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get stresschaos logger.go:42: 13:20:44 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep -v NAME logger.go:42: 13:20:44 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:20:44 | self-healing-chaos/9-destroy-chaos-mesh | No resources found in default namespace. logger.go:42: 13:20:44 | self-healing-chaos/9-destroy-chaos-mesh | + return logger.go:42: 13:20:44 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 13:20:44 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer timechaos --all logger.go:42: 13:20:44 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=timechaos logger.go:42: 13:20:44 | self-healing-chaos/9-destroy-chaos-mesh | + local name=--all logger.go:42: 13:20:44 | self-healing-chaos/9-destroy-chaos-mesh | + '[' --all == --all ']' logger.go:42: 13:20:44 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get timechaos logger.go:42: 13:20:44 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep -v NAME logger.go:42: 13:20:44 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:20:45 | self-healing-chaos/9-destroy-chaos-mesh | No resources found in default namespace. logger.go:42: 13:20:45 | self-healing-chaos/9-destroy-chaos-mesh | + return logger.go:42: 13:20:45 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 13:20:45 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer workflownodes --all logger.go:42: 13:20:45 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=workflownodes logger.go:42: 13:20:45 | self-healing-chaos/9-destroy-chaos-mesh | + local name=--all logger.go:42: 13:20:45 | self-healing-chaos/9-destroy-chaos-mesh | + '[' --all == --all ']' logger.go:42: 13:20:45 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get workflownodes logger.go:42: 13:20:45 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep -v NAME logger.go:42: 13:20:45 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:20:46 | self-healing-chaos/9-destroy-chaos-mesh | No resources found in default namespace. logger.go:42: 13:20:46 | self-healing-chaos/9-destroy-chaos-mesh | + return logger.go:42: 13:20:46 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 13:20:46 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer workflows --all logger.go:42: 13:20:46 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=workflows logger.go:42: 13:20:46 | self-healing-chaos/9-destroy-chaos-mesh | + local name=--all logger.go:42: 13:20:46 | self-healing-chaos/9-destroy-chaos-mesh | + '[' --all == --all ']' logger.go:42: 13:20:46 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get workflows logger.go:42: 13:20:46 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep -v NAME logger.go:42: 13:20:46 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:20:47 | self-healing-chaos/9-destroy-chaos-mesh | No resources found in default namespace. logger.go:42: 13:20:47 | self-healing-chaos/9-destroy-chaos-mesh | + return logger.go:42: 13:20:47 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get crd -A logger.go:42: 13:20:47 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep chaos-mesh.org logger.go:42: 13:20:47 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:20:48 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 13:20:48 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer crd awschaos.chaos-mesh.org logger.go:42: 13:20:48 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=crd logger.go:42: 13:20:48 | self-healing-chaos/9-destroy-chaos-mesh | + local name=awschaos.chaos-mesh.org logger.go:42: 13:20:48 | self-healing-chaos/9-destroy-chaos-mesh | + '[' awschaos.chaos-mesh.org == --all ']' logger.go:42: 13:20:48 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl delete crd awschaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 13:20:49 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "awschaos.chaos-mesh.org" deleted logger.go:42: 13:20:49 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl patch crd awschaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 13:20:50 | self-healing-chaos/9-destroy-chaos-mesh | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "awschaos.chaos-mesh.org" not found logger.go:42: 13:20:50 | self-healing-chaos/9-destroy-chaos-mesh | + : logger.go:42: 13:20:50 | self-healing-chaos/9-destroy-chaos-mesh | + timeout 60 kubectl delete crd awschaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 13:20:51 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 13:20:51 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer crd azurechaos.chaos-mesh.org logger.go:42: 13:20:51 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=crd logger.go:42: 13:20:51 | self-healing-chaos/9-destroy-chaos-mesh | + local name=azurechaos.chaos-mesh.org logger.go:42: 13:20:51 | self-healing-chaos/9-destroy-chaos-mesh | + '[' azurechaos.chaos-mesh.org == --all ']' logger.go:42: 13:20:51 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl delete crd azurechaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 13:20:51 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "azurechaos.chaos-mesh.org" deleted logger.go:42: 13:20:51 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl patch crd azurechaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 13:20:52 | self-healing-chaos/9-destroy-chaos-mesh | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "azurechaos.chaos-mesh.org" not found logger.go:42: 13:20:52 | self-healing-chaos/9-destroy-chaos-mesh | + : logger.go:42: 13:20:52 | self-healing-chaos/9-destroy-chaos-mesh | + timeout 60 kubectl delete crd azurechaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 13:20:53 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 13:20:53 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer crd blockchaos.chaos-mesh.org logger.go:42: 13:20:53 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=crd logger.go:42: 13:20:53 | self-healing-chaos/9-destroy-chaos-mesh | + local name=blockchaos.chaos-mesh.org logger.go:42: 13:20:53 | self-healing-chaos/9-destroy-chaos-mesh | + '[' blockchaos.chaos-mesh.org == --all ']' logger.go:42: 13:20:53 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl delete crd blockchaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 13:20:54 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "blockchaos.chaos-mesh.org" deleted logger.go:42: 13:20:54 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl patch crd blockchaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 13:20:55 | self-healing-chaos/9-destroy-chaos-mesh | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "blockchaos.chaos-mesh.org" not found logger.go:42: 13:20:55 | self-healing-chaos/9-destroy-chaos-mesh | + : logger.go:42: 13:20:55 | self-healing-chaos/9-destroy-chaos-mesh | + timeout 60 kubectl delete crd blockchaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 13:20:55 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 13:20:55 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer crd dnschaos.chaos-mesh.org logger.go:42: 13:20:55 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=crd logger.go:42: 13:20:55 | self-healing-chaos/9-destroy-chaos-mesh | + local name=dnschaos.chaos-mesh.org logger.go:42: 13:20:55 | self-healing-chaos/9-destroy-chaos-mesh | + '[' dnschaos.chaos-mesh.org == --all ']' logger.go:42: 13:20:55 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl delete crd dnschaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 13:20:56 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "dnschaos.chaos-mesh.org" deleted logger.go:42: 13:20:56 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl patch crd dnschaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 13:20:57 | self-healing-chaos/9-destroy-chaos-mesh | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "dnschaos.chaos-mesh.org" not found logger.go:42: 13:20:57 | self-healing-chaos/9-destroy-chaos-mesh | + : logger.go:42: 13:20:57 | self-healing-chaos/9-destroy-chaos-mesh | + timeout 60 kubectl delete crd dnschaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 13:20:58 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 13:20:58 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer crd gcpchaos.chaos-mesh.org logger.go:42: 13:20:58 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=crd logger.go:42: 13:20:58 | self-healing-chaos/9-destroy-chaos-mesh | + local name=gcpchaos.chaos-mesh.org logger.go:42: 13:20:58 | self-healing-chaos/9-destroy-chaos-mesh | + '[' gcpchaos.chaos-mesh.org == --all ']' logger.go:42: 13:20:58 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl delete crd gcpchaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 13:20:59 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "gcpchaos.chaos-mesh.org" deleted logger.go:42: 13:20:59 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl patch crd gcpchaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 13:21:00 | self-healing-chaos/9-destroy-chaos-mesh | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "gcpchaos.chaos-mesh.org" not found logger.go:42: 13:21:00 | self-healing-chaos/9-destroy-chaos-mesh | + : logger.go:42: 13:21:00 | self-healing-chaos/9-destroy-chaos-mesh | + timeout 60 kubectl delete crd gcpchaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 13:21:01 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 13:21:01 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer crd httpchaos.chaos-mesh.org logger.go:42: 13:21:01 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=crd logger.go:42: 13:21:01 | self-healing-chaos/9-destroy-chaos-mesh | + local name=httpchaos.chaos-mesh.org logger.go:42: 13:21:01 | self-healing-chaos/9-destroy-chaos-mesh | + '[' httpchaos.chaos-mesh.org == --all ']' logger.go:42: 13:21:01 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl delete crd httpchaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 13:21:02 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "httpchaos.chaos-mesh.org" deleted logger.go:42: 13:21:02 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl patch crd httpchaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 13:21:02 | self-healing-chaos/9-destroy-chaos-mesh | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "httpchaos.chaos-mesh.org" not found logger.go:42: 13:21:02 | self-healing-chaos/9-destroy-chaos-mesh | + : logger.go:42: 13:21:02 | self-healing-chaos/9-destroy-chaos-mesh | + timeout 60 kubectl delete crd httpchaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 13:21:03 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 13:21:03 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer crd iochaos.chaos-mesh.org logger.go:42: 13:21:03 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=crd logger.go:42: 13:21:03 | self-healing-chaos/9-destroy-chaos-mesh | + local name=iochaos.chaos-mesh.org logger.go:42: 13:21:03 | self-healing-chaos/9-destroy-chaos-mesh | + '[' iochaos.chaos-mesh.org == --all ']' logger.go:42: 13:21:03 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl delete crd iochaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 13:21:04 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "iochaos.chaos-mesh.org" deleted logger.go:42: 13:21:04 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl patch crd iochaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 13:21:05 | self-healing-chaos/9-destroy-chaos-mesh | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "iochaos.chaos-mesh.org" not found logger.go:42: 13:21:05 | self-healing-chaos/9-destroy-chaos-mesh | + : logger.go:42: 13:21:05 | self-healing-chaos/9-destroy-chaos-mesh | + timeout 60 kubectl delete crd iochaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 13:21:06 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 13:21:06 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer crd jvmchaos.chaos-mesh.org logger.go:42: 13:21:06 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=crd logger.go:42: 13:21:06 | self-healing-chaos/9-destroy-chaos-mesh | + local name=jvmchaos.chaos-mesh.org logger.go:42: 13:21:06 | self-healing-chaos/9-destroy-chaos-mesh | + '[' jvmchaos.chaos-mesh.org == --all ']' logger.go:42: 13:21:06 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl delete crd jvmchaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 13:21:07 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "jvmchaos.chaos-mesh.org" deleted logger.go:42: 13:21:07 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl patch crd jvmchaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 13:21:07 | self-healing-chaos/9-destroy-chaos-mesh | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "jvmchaos.chaos-mesh.org" not found logger.go:42: 13:21:07 | self-healing-chaos/9-destroy-chaos-mesh | + : logger.go:42: 13:21:07 | self-healing-chaos/9-destroy-chaos-mesh | + timeout 60 kubectl delete crd jvmchaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 13:21:08 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 13:21:08 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer crd kernelchaos.chaos-mesh.org logger.go:42: 13:21:08 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=crd logger.go:42: 13:21:08 | self-healing-chaos/9-destroy-chaos-mesh | + local name=kernelchaos.chaos-mesh.org logger.go:42: 13:21:08 | self-healing-chaos/9-destroy-chaos-mesh | + '[' kernelchaos.chaos-mesh.org == --all ']' logger.go:42: 13:21:08 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl delete crd kernelchaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 13:21:09 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "kernelchaos.chaos-mesh.org" deleted logger.go:42: 13:21:09 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl patch crd kernelchaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 13:21:10 | self-healing-chaos/9-destroy-chaos-mesh | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "kernelchaos.chaos-mesh.org" not found logger.go:42: 13:21:10 | self-healing-chaos/9-destroy-chaos-mesh | + : logger.go:42: 13:21:10 | self-healing-chaos/9-destroy-chaos-mesh | + timeout 60 kubectl delete crd kernelchaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 13:21:11 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 13:21:11 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer crd networkchaos.chaos-mesh.org logger.go:42: 13:21:11 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=crd logger.go:42: 13:21:11 | self-healing-chaos/9-destroy-chaos-mesh | + local name=networkchaos.chaos-mesh.org logger.go:42: 13:21:11 | self-healing-chaos/9-destroy-chaos-mesh | + '[' networkchaos.chaos-mesh.org == --all ']' logger.go:42: 13:21:11 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl delete crd networkchaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 13:21:12 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "networkchaos.chaos-mesh.org" deleted logger.go:42: 13:21:12 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl patch crd networkchaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 13:21:13 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io/networkchaos.chaos-mesh.org patched logger.go:42: 13:21:13 | self-healing-chaos/9-destroy-chaos-mesh | + timeout 60 kubectl delete crd networkchaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 13:21:14 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 13:21:14 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer crd physicalmachinechaos.chaos-mesh.org logger.go:42: 13:21:14 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=crd logger.go:42: 13:21:14 | self-healing-chaos/9-destroy-chaos-mesh | + local name=physicalmachinechaos.chaos-mesh.org logger.go:42: 13:21:14 | self-healing-chaos/9-destroy-chaos-mesh | + '[' physicalmachinechaos.chaos-mesh.org == --all ']' logger.go:42: 13:21:14 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl delete crd physicalmachinechaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 13:21:15 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "physicalmachinechaos.chaos-mesh.org" deleted logger.go:42: 13:21:15 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl patch crd physicalmachinechaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 13:21:15 | self-healing-chaos/9-destroy-chaos-mesh | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "physicalmachinechaos.chaos-mesh.org" not found logger.go:42: 13:21:15 | self-healing-chaos/9-destroy-chaos-mesh | + : logger.go:42: 13:21:15 | self-healing-chaos/9-destroy-chaos-mesh | + timeout 60 kubectl delete crd physicalmachinechaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 13:21:16 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 13:21:16 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer crd physicalmachines.chaos-mesh.org logger.go:42: 13:21:16 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=crd logger.go:42: 13:21:16 | self-healing-chaos/9-destroy-chaos-mesh | + local name=physicalmachines.chaos-mesh.org logger.go:42: 13:21:16 | self-healing-chaos/9-destroy-chaos-mesh | + '[' physicalmachines.chaos-mesh.org == --all ']' logger.go:42: 13:21:16 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl delete crd physicalmachines.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 13:21:17 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "physicalmachines.chaos-mesh.org" deleted logger.go:42: 13:21:17 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl patch crd physicalmachines.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 13:21:18 | self-healing-chaos/9-destroy-chaos-mesh | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "physicalmachines.chaos-mesh.org" not found logger.go:42: 13:21:18 | self-healing-chaos/9-destroy-chaos-mesh | + : logger.go:42: 13:21:18 | self-healing-chaos/9-destroy-chaos-mesh | + timeout 60 kubectl delete crd physicalmachines.chaos-mesh.org --ignore-not-found=true logger.go:42: 13:21:19 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 13:21:19 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer crd podchaos.chaos-mesh.org logger.go:42: 13:21:19 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=crd logger.go:42: 13:21:19 | self-healing-chaos/9-destroy-chaos-mesh | + local name=podchaos.chaos-mesh.org logger.go:42: 13:21:19 | self-healing-chaos/9-destroy-chaos-mesh | + '[' podchaos.chaos-mesh.org == --all ']' logger.go:42: 13:21:19 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl delete crd podchaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 13:21:19 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "podchaos.chaos-mesh.org" deleted logger.go:42: 13:21:19 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl patch crd podchaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 13:21:20 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io/podchaos.chaos-mesh.org patched logger.go:42: 13:21:20 | self-healing-chaos/9-destroy-chaos-mesh | + timeout 60 kubectl delete crd podchaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 13:21:21 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 13:21:21 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer crd podhttpchaos.chaos-mesh.org logger.go:42: 13:21:21 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=crd logger.go:42: 13:21:21 | self-healing-chaos/9-destroy-chaos-mesh | + local name=podhttpchaos.chaos-mesh.org logger.go:42: 13:21:21 | self-healing-chaos/9-destroy-chaos-mesh | + '[' podhttpchaos.chaos-mesh.org == --all ']' logger.go:42: 13:21:21 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl delete crd podhttpchaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 13:21:22 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "podhttpchaos.chaos-mesh.org" deleted logger.go:42: 13:21:22 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl patch crd podhttpchaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 13:21:23 | self-healing-chaos/9-destroy-chaos-mesh | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "podhttpchaos.chaos-mesh.org" not found logger.go:42: 13:21:23 | self-healing-chaos/9-destroy-chaos-mesh | + : logger.go:42: 13:21:23 | self-healing-chaos/9-destroy-chaos-mesh | + timeout 60 kubectl delete crd podhttpchaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 13:21:24 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 13:21:24 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer crd podiochaos.chaos-mesh.org logger.go:42: 13:21:24 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=crd logger.go:42: 13:21:24 | self-healing-chaos/9-destroy-chaos-mesh | + local name=podiochaos.chaos-mesh.org logger.go:42: 13:21:24 | self-healing-chaos/9-destroy-chaos-mesh | + '[' podiochaos.chaos-mesh.org == --all ']' logger.go:42: 13:21:24 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl delete crd podiochaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 13:21:24 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "podiochaos.chaos-mesh.org" deleted logger.go:42: 13:21:24 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl patch crd podiochaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 13:21:25 | self-healing-chaos/9-destroy-chaos-mesh | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "podiochaos.chaos-mesh.org" not found logger.go:42: 13:21:25 | self-healing-chaos/9-destroy-chaos-mesh | + : logger.go:42: 13:21:25 | self-healing-chaos/9-destroy-chaos-mesh | + timeout 60 kubectl delete crd podiochaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 13:21:26 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 13:21:26 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer crd podnetworkchaos.chaos-mesh.org logger.go:42: 13:21:26 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=crd logger.go:42: 13:21:26 | self-healing-chaos/9-destroy-chaos-mesh | + local name=podnetworkchaos.chaos-mesh.org logger.go:42: 13:21:26 | self-healing-chaos/9-destroy-chaos-mesh | + '[' podnetworkchaos.chaos-mesh.org == --all ']' logger.go:42: 13:21:26 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl delete crd podnetworkchaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 13:21:27 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "podnetworkchaos.chaos-mesh.org" deleted logger.go:42: 13:21:27 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl patch crd podnetworkchaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 13:21:28 | self-healing-chaos/9-destroy-chaos-mesh | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "podnetworkchaos.chaos-mesh.org" not found logger.go:42: 13:21:28 | self-healing-chaos/9-destroy-chaos-mesh | + : logger.go:42: 13:21:28 | self-healing-chaos/9-destroy-chaos-mesh | + timeout 60 kubectl delete crd podnetworkchaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 13:21:29 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 13:21:29 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer crd remoteclusters.chaos-mesh.org logger.go:42: 13:21:29 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=crd logger.go:42: 13:21:29 | self-healing-chaos/9-destroy-chaos-mesh | + local name=remoteclusters.chaos-mesh.org logger.go:42: 13:21:29 | self-healing-chaos/9-destroy-chaos-mesh | + '[' remoteclusters.chaos-mesh.org == --all ']' logger.go:42: 13:21:29 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl delete crd remoteclusters.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 13:21:29 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "remoteclusters.chaos-mesh.org" deleted logger.go:42: 13:21:29 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl patch crd remoteclusters.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 13:21:30 | self-healing-chaos/9-destroy-chaos-mesh | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "remoteclusters.chaos-mesh.org" not found logger.go:42: 13:21:30 | self-healing-chaos/9-destroy-chaos-mesh | + : logger.go:42: 13:21:30 | self-healing-chaos/9-destroy-chaos-mesh | + timeout 60 kubectl delete crd remoteclusters.chaos-mesh.org --ignore-not-found=true logger.go:42: 13:21:31 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 13:21:31 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer crd schedules.chaos-mesh.org logger.go:42: 13:21:31 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=crd logger.go:42: 13:21:31 | self-healing-chaos/9-destroy-chaos-mesh | + local name=schedules.chaos-mesh.org logger.go:42: 13:21:31 | self-healing-chaos/9-destroy-chaos-mesh | + '[' schedules.chaos-mesh.org == --all ']' logger.go:42: 13:21:31 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl delete crd schedules.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 13:21:32 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "schedules.chaos-mesh.org" deleted logger.go:42: 13:21:32 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl patch crd schedules.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 13:21:36 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io/schedules.chaos-mesh.org patched logger.go:42: 13:21:36 | self-healing-chaos/9-destroy-chaos-mesh | + timeout 60 kubectl delete crd schedules.chaos-mesh.org --ignore-not-found=true logger.go:42: 13:21:37 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 13:21:37 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer crd statuschecks.chaos-mesh.org logger.go:42: 13:21:37 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=crd logger.go:42: 13:21:37 | self-healing-chaos/9-destroy-chaos-mesh | + local name=statuschecks.chaos-mesh.org logger.go:42: 13:21:37 | self-healing-chaos/9-destroy-chaos-mesh | + '[' statuschecks.chaos-mesh.org == --all ']' logger.go:42: 13:21:37 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl delete crd statuschecks.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 13:21:37 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "statuschecks.chaos-mesh.org" deleted logger.go:42: 13:21:37 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl patch crd statuschecks.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 13:21:38 | self-healing-chaos/9-destroy-chaos-mesh | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "statuschecks.chaos-mesh.org" not found logger.go:42: 13:21:38 | self-healing-chaos/9-destroy-chaos-mesh | + : logger.go:42: 13:21:38 | self-healing-chaos/9-destroy-chaos-mesh | + timeout 60 kubectl delete crd statuschecks.chaos-mesh.org --ignore-not-found=true logger.go:42: 13:21:39 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 13:21:39 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer crd stresschaos.chaos-mesh.org logger.go:42: 13:21:39 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=crd logger.go:42: 13:21:39 | self-healing-chaos/9-destroy-chaos-mesh | + local name=stresschaos.chaos-mesh.org logger.go:42: 13:21:39 | self-healing-chaos/9-destroy-chaos-mesh | + '[' stresschaos.chaos-mesh.org == --all ']' logger.go:42: 13:21:39 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl delete crd stresschaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 13:21:40 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "stresschaos.chaos-mesh.org" deleted logger.go:42: 13:21:40 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl patch crd stresschaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 13:21:41 | self-healing-chaos/9-destroy-chaos-mesh | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "stresschaos.chaos-mesh.org" not found logger.go:42: 13:21:41 | self-healing-chaos/9-destroy-chaos-mesh | + : logger.go:42: 13:21:41 | self-healing-chaos/9-destroy-chaos-mesh | + timeout 60 kubectl delete crd stresschaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 13:21:42 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 13:21:42 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer crd timechaos.chaos-mesh.org logger.go:42: 13:21:42 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=crd logger.go:42: 13:21:42 | self-healing-chaos/9-destroy-chaos-mesh | + local name=timechaos.chaos-mesh.org logger.go:42: 13:21:42 | self-healing-chaos/9-destroy-chaos-mesh | + '[' timechaos.chaos-mesh.org == --all ']' logger.go:42: 13:21:42 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl delete crd timechaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 13:21:43 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "timechaos.chaos-mesh.org" deleted logger.go:42: 13:21:43 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl patch crd timechaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 13:21:44 | self-healing-chaos/9-destroy-chaos-mesh | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "timechaos.chaos-mesh.org" not found logger.go:42: 13:21:44 | self-healing-chaos/9-destroy-chaos-mesh | + : logger.go:42: 13:21:44 | self-healing-chaos/9-destroy-chaos-mesh | + timeout 60 kubectl delete crd timechaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 13:21:44 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 13:21:44 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer crd workflownodes.chaos-mesh.org logger.go:42: 13:21:44 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=crd logger.go:42: 13:21:44 | self-healing-chaos/9-destroy-chaos-mesh | + local name=workflownodes.chaos-mesh.org logger.go:42: 13:21:44 | self-healing-chaos/9-destroy-chaos-mesh | + '[' workflownodes.chaos-mesh.org == --all ']' logger.go:42: 13:21:44 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl delete crd workflownodes.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 13:21:46 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "workflownodes.chaos-mesh.org" deleted logger.go:42: 13:21:46 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl patch crd workflownodes.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 13:21:50 | self-healing-chaos/9-destroy-chaos-mesh | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "workflownodes.chaos-mesh.org" not found logger.go:42: 13:21:50 | self-healing-chaos/9-destroy-chaos-mesh | + : logger.go:42: 13:21:50 | self-healing-chaos/9-destroy-chaos-mesh | + timeout 60 kubectl delete crd workflownodes.chaos-mesh.org --ignore-not-found=true logger.go:42: 13:21:51 | self-healing-chaos/9-destroy-chaos-mesh | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 13:21:51 | self-healing-chaos/9-destroy-chaos-mesh | + delete_resource_with_finalizer crd workflows.chaos-mesh.org logger.go:42: 13:21:51 | self-healing-chaos/9-destroy-chaos-mesh | + local resource=crd logger.go:42: 13:21:51 | self-healing-chaos/9-destroy-chaos-mesh | + local name=workflows.chaos-mesh.org logger.go:42: 13:21:51 | self-healing-chaos/9-destroy-chaos-mesh | + '[' workflows.chaos-mesh.org == --all ']' logger.go:42: 13:21:51 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl delete crd workflows.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 13:21:53 | self-healing-chaos/9-destroy-chaos-mesh | customresourcedefinition.apiextensions.k8s.io "workflows.chaos-mesh.org" deleted logger.go:42: 13:21:53 | self-healing-chaos/9-destroy-chaos-mesh | + kubectl patch crd workflows.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 13:21:55 | self-healing-chaos/9-destroy-chaos-mesh | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "workflows.chaos-mesh.org" not found logger.go:42: 13:21:55 | self-healing-chaos/9-destroy-chaos-mesh | + : logger.go:42: 13:21:55 | self-healing-chaos/9-destroy-chaos-mesh | + timeout 60 kubectl delete crd workflows.chaos-mesh.org --ignore-not-found=true logger.go:42: 13:21:56 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get clusterrolebinding -A logger.go:42: 13:21:56 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep chaos-mesh logger.go:42: 13:21:56 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:21:57 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get clusterrole -A logger.go:42: 13:21:57 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep chaos-mesh logger.go:42: 13:21:57 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:21:58 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get mutatingwebhookconfiguration -A logger.go:42: 13:21:58 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep chaos-mesh logger.go:42: 13:21:58 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:21:59 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get validatingwebhookconfiguration -A logger.go:42: 13:21:59 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep chaos-mesh logger.go:42: 13:21:59 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:22:00 | self-healing-chaos/9-destroy-chaos-mesh | ++ kubectl get validatingwebhookconfiguration -A logger.go:42: 13:22:00 | self-healing-chaos/9-destroy-chaos-mesh | ++ grep validate-auth logger.go:42: 13:22:00 | self-healing-chaos/9-destroy-chaos-mesh | ++ awk '{print $1}' logger.go:42: 13:22:02 | self-healing-chaos/9-destroy-chaos-mesh | test step completed 9-destroy-chaos-mesh logger.go:42: 13:22:02 | self-healing-chaos/10-drop-finalizer | starting test step 10-drop-finalizer logger.go:42: 13:22:03 | self-healing-chaos/10-drop-finalizer | PerconaServerMySQL:kuttl-test-set-bluejay/self-healing-chaos updated logger.go:42: 13:22:03 | self-healing-chaos/10-drop-finalizer | test step completed 10-drop-finalizer logger.go:42: 13:22:03 | self-healing-chaos | self-healing-chaos events from ns kuttl-test-set-bluejay: logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:09:58 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-xs9bm Scheduled Successfully assigned kuttl-test-set-bluejay/percona-server-mysql-operator-6dcd78d857-xs9bm to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:09:58 +0000 UTC Normal ReplicaSet.apps percona-server-mysql-operator-6dcd78d857 SuccessfulCreate Created pod: percona-server-mysql-operator-6dcd78d857-xs9bm logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:09:58 +0000 UTC Normal Deployment.apps percona-server-mysql-operator ScalingReplicaSet Scaled up replica set percona-server-mysql-operator-6dcd78d857 to 1 logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:00 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-xs9bm.spec.containers{manager} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:01 +0000 UTC Normal Lease.coordination.k8s.io 08db2feb.percona.com LeaderElection percona-server-mysql-operator-6dcd78d857-xs9bm_c9f7ba9e-0e66-464a-89e7-8b146f148633 became leader logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:01 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-xs9bm.spec.containers{manager} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 458.071526ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:01 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-xs9bm.spec.containers{manager} Created Created container manager logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:01 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-xs9bm.spec.containers{manager} Started Started container manager logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:02 +0000 UTC Normal Pod mysql-client Scheduled Successfully assigned kuttl-test-set-bluejay/mysql-client to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:03 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Pulled Container image "percona/percona-server:8.0.25" already present on machine logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:03 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Created Created container mysql-client logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:03 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Started Started container mysql-client logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:53 +0000 UTC Normal Pod chaos-controller-manager-66674457c8-gl4lm Scheduled Successfully assigned kuttl-test-set-bluejay/chaos-controller-manager-66674457c8-gl4lm to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:53 +0000 UTC Normal ReplicaSet.apps chaos-controller-manager-66674457c8 SuccessfulCreate Created pod: chaos-controller-manager-66674457c8-gl4lm logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:53 +0000 UTC Normal ReplicaSet.apps chaos-controller-manager-66674457c8 SuccessfulCreate Created pod: chaos-controller-manager-66674457c8-r5pl8 logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:53 +0000 UTC Normal ReplicaSet.apps chaos-controller-manager-66674457c8 SuccessfulCreate Created pod: chaos-controller-manager-66674457c8-bh98d logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:53 +0000 UTC Normal Deployment.apps chaos-controller-manager ScalingReplicaSet Scaled up replica set chaos-controller-manager-66674457c8 to 3 logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:53 +0000 UTC Normal Pod chaos-daemon-mdxm8 Scheduled Successfully assigned kuttl-test-set-bluejay/chaos-daemon-mdxm8 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w381 logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:53 +0000 UTC Normal DaemonSet.apps chaos-daemon SuccessfulCreate Created pod: chaos-daemon-mdxm8 logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:53 +0000 UTC Normal DaemonSet.apps chaos-daemon SuccessfulCreate Created pod: chaos-daemon-dtcv6 logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:53 +0000 UTC Normal DaemonSet.apps chaos-daemon SuccessfulCreate Created pod: chaos-daemon-hd6dz logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:54 +0000 UTC Normal Pod chaos-controller-manager-66674457c8-bh98d Scheduled Successfully assigned kuttl-test-set-bluejay/chaos-controller-manager-66674457c8-bh98d to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w381 logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:54 +0000 UTC Normal Pod chaos-controller-manager-66674457c8-bh98d.spec.containers{chaos-mesh} Pulling Pulling image "ghcr.io/chaos-mesh/chaos-mesh:v2.5.1" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:54 +0000 UTC Normal Pod chaos-controller-manager-66674457c8-gl4lm.spec.containers{chaos-mesh} Pulling Pulling image "ghcr.io/chaos-mesh/chaos-mesh:v2.5.1" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:54 +0000 UTC Normal Pod chaos-controller-manager-66674457c8-r5pl8 Scheduled Successfully assigned kuttl-test-set-bluejay/chaos-controller-manager-66674457c8-r5pl8 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:54 +0000 UTC Normal Pod chaos-controller-manager-66674457c8-r5pl8.spec.containers{chaos-mesh} Pulling Pulling image "ghcr.io/chaos-mesh/chaos-mesh:v2.5.1" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:54 +0000 UTC Normal Pod chaos-daemon-dtcv6 Scheduled Successfully assigned kuttl-test-set-bluejay/chaos-daemon-dtcv6 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:54 +0000 UTC Normal Pod chaos-daemon-dtcv6.spec.containers{chaos-daemon} Pulling Pulling image "ghcr.io/chaos-mesh/chaos-daemon:v2.5.1" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:54 +0000 UTC Normal Pod chaos-daemon-hd6dz Scheduled Successfully assigned kuttl-test-set-bluejay/chaos-daemon-hd6dz to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:54 +0000 UTC Normal Pod chaos-daemon-hd6dz.spec.containers{chaos-daemon} Pulling Pulling image "ghcr.io/chaos-mesh/chaos-daemon:v2.5.1" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:10:54 +0000 UTC Normal Pod chaos-daemon-mdxm8.spec.containers{chaos-daemon} Pulling Pulling image "ghcr.io/chaos-mesh/chaos-daemon:v2.5.1" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:05 +0000 UTC Normal Pod chaos-daemon-dtcv6.spec.containers{chaos-daemon} Pulled Successfully pulled image "ghcr.io/chaos-mesh/chaos-daemon:v2.5.1" in 11.058073256s logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:05 +0000 UTC Normal Pod chaos-daemon-dtcv6.spec.containers{chaos-daemon} Created Created container chaos-daemon logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:05 +0000 UTC Normal Pod chaos-daemon-dtcv6.spec.containers{chaos-daemon} Started Started container chaos-daemon logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:05 +0000 UTC Normal Pod chaos-daemon-hd6dz.spec.containers{chaos-daemon} Pulled Successfully pulled image "ghcr.io/chaos-mesh/chaos-daemon:v2.5.1" in 10.624488112s logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:05 +0000 UTC Normal Pod chaos-daemon-hd6dz.spec.containers{chaos-daemon} Created Created container chaos-daemon logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:05 +0000 UTC Normal Pod chaos-daemon-hd6dz.spec.containers{chaos-daemon} Started Started container chaos-daemon logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:05 +0000 UTC Normal Pod chaos-daemon-mdxm8.spec.containers{chaos-daemon} Pulled Successfully pulled image "ghcr.io/chaos-mesh/chaos-daemon:v2.5.1" in 10.570159025s logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:05 +0000 UTC Normal Pod chaos-daemon-mdxm8.spec.containers{chaos-daemon} Created Created container chaos-daemon logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:05 +0000 UTC Normal Pod chaos-daemon-mdxm8.spec.containers{chaos-daemon} Started Started container chaos-daemon logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:07 +0000 UTC Normal Pod chaos-controller-manager-66674457c8-bh98d.spec.containers{chaos-mesh} Pulled Successfully pulled image "ghcr.io/chaos-mesh/chaos-mesh:v2.5.1" in 12.385011406s logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:07 +0000 UTC Normal Pod chaos-controller-manager-66674457c8-bh98d.spec.containers{chaos-mesh} Created Created container chaos-mesh logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:07 +0000 UTC Normal Pod chaos-controller-manager-66674457c8-bh98d.spec.containers{chaos-mesh} Started Started container chaos-mesh logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:07 +0000 UTC Normal Pod chaos-controller-manager-66674457c8-gl4lm.spec.containers{chaos-mesh} Pulled Successfully pulled image "ghcr.io/chaos-mesh/chaos-mesh:v2.5.1" in 12.56918363s logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:07 +0000 UTC Normal Pod chaos-controller-manager-66674457c8-gl4lm.spec.containers{chaos-mesh} Created Created container chaos-mesh logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:07 +0000 UTC Normal Pod chaos-controller-manager-66674457c8-gl4lm.spec.containers{chaos-mesh} Started Started container chaos-mesh logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:07 +0000 UTC Normal Pod chaos-controller-manager-66674457c8-r5pl8.spec.containers{chaos-mesh} Pulled Successfully pulled image "ghcr.io/chaos-mesh/chaos-mesh:v2.5.1" in 12.414697255s logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:07 +0000 UTC Normal Pod chaos-controller-manager-66674457c8-r5pl8.spec.containers{chaos-mesh} Created Created container chaos-mesh logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:07 +0000 UTC Normal Pod chaos-controller-manager-66674457c8-r5pl8.spec.containers{chaos-mesh} Started Started container chaos-mesh logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:08 +0000 UTC Normal ConfigMap chaos-mesh LeaderElection chaos-controller-manager-66674457c8-bh98d_0fe6dade-4aa3-450d-a7df-4a3ea7e8c7c3 became leader logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:08 +0000 UTC Normal Lease.coordination.k8s.io chaos-mesh LeaderElection chaos-controller-manager-66674457c8-bh98d_0fe6dade-4aa3-450d-a7df-4a3ea7e8c7c3 became leader logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:10 +0000 UTC Normal CertificateRequest.cert-manager.io self-healing-chaos-ca-cert-p7h7z WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:10 +0000 UTC Normal CertificateRequest.cert-manager.io self-healing-chaos-ca-cert-p7h7z WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:10 +0000 UTC Normal CertificateRequest.cert-manager.io self-healing-chaos-ca-cert-p7h7z WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:10 +0000 UTC Normal CertificateRequest.cert-manager.io self-healing-chaos-ca-cert-p7h7z WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:10 +0000 UTC Normal CertificateRequest.cert-manager.io self-healing-chaos-ca-cert-p7h7z WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:10 +0000 UTC Normal CertificateRequest.cert-manager.io self-healing-chaos-ca-cert-p7h7z cert-manager.io Certificate request has been approved by cert-manager.io logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:10 +0000 UTC Normal CertificateRequest.cert-manager.io self-healing-chaos-ca-cert-p7h7z CertificateIssued Certificate fetched from issuer successfully logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:10 +0000 UTC Normal Certificate.cert-manager.io self-healing-chaos-ca-cert Issuing Issuing certificate as Secret does not exist logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:10 +0000 UTC Normal Certificate.cert-manager.io self-healing-chaos-ca-cert Generated Stored new private key in temporary Secret resource "self-healing-chaos-ca-cert-bk5mc" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:10 +0000 UTC Normal Certificate.cert-manager.io self-healing-chaos-ca-cert Requested Created new CertificateRequest resource "self-healing-chaos-ca-cert-p7h7z" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:10 +0000 UTC Normal Certificate.cert-manager.io self-healing-chaos-ca-cert Issuing The certificate has been successfully issued logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:12 +0000 UTC Normal Issuer.cert-manager.io self-healing-chaos-pso-issuer KeyPairVerified Signing CA verified logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:13 +0000 UTC Normal CertificateRequest.cert-manager.io self-healing-chaos-ssl-9rd8g WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:13 +0000 UTC Normal CertificateRequest.cert-manager.io self-healing-chaos-ssl-9rd8g WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:13 +0000 UTC Normal CertificateRequest.cert-manager.io self-healing-chaos-ssl-9rd8g WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:13 +0000 UTC Normal CertificateRequest.cert-manager.io self-healing-chaos-ssl-9rd8g WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:13 +0000 UTC Normal CertificateRequest.cert-manager.io self-healing-chaos-ssl-9rd8g WaitingForApproval Not signing CertificateRequest until it is Approved logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:13 +0000 UTC Normal CertificateRequest.cert-manager.io self-healing-chaos-ssl-9rd8g cert-manager.io Certificate request has been approved by cert-manager.io logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:13 +0000 UTC Normal CertificateRequest.cert-manager.io self-healing-chaos-ssl-9rd8g CertificateIssued Certificate fetched from issuer successfully logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:13 +0000 UTC Normal Certificate.cert-manager.io self-healing-chaos-ssl Issuing Existing issued Secret is not up to date for spec: [spec.dnsNames] logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:13 +0000 UTC Normal Certificate.cert-manager.io self-healing-chaos-ssl Reused Reusing private key stored in existing Secret resource "test-ssl" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:13 +0000 UTC Normal Certificate.cert-manager.io self-healing-chaos-ssl Requested Created new CertificateRequest resource "self-healing-chaos-ssl-9rd8g" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:13 +0000 UTC Normal Certificate.cert-manager.io self-healing-chaos-ssl Issuing The certificate has been successfully issued logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:16 +0000 UTC Normal PersistentVolumeClaim datadir-self-healing-chaos-mysql-0 WaitForFirstConsumer waiting for first consumer to be created before binding logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:16 +0000 UTC Normal PersistentVolumeClaim datadir-self-healing-chaos-mysql-0 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:16 +0000 UTC Normal StatefulSet.apps self-healing-chaos-mysql SuccessfulCreate create Claim datadir-self-healing-chaos-mysql-0 Pod self-healing-chaos-mysql-0 in StatefulSet self-healing-chaos-mysql success logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:16 +0000 UTC Normal StatefulSet.apps self-healing-chaos-mysql SuccessfulCreate create Pod self-healing-chaos-mysql-0 in StatefulSet self-healing-chaos-mysql successful logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:16 +0000 UTC Normal Pod self-healing-chaos-orc-0 Scheduled Successfully assigned kuttl-test-set-bluejay/self-healing-chaos-orc-0 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:16 +0000 UTC Normal StatefulSet.apps self-healing-chaos-orc SuccessfulCreate create Pod self-healing-chaos-orc-0 in StatefulSet self-healing-chaos-orc successful logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:17 +0000 UTC Normal PersistentVolumeClaim datadir-self-healing-chaos-mysql-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-set-bluejay/datadir-self-healing-chaos-mysql-0" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:17 +0000 UTC Normal Pod self-healing-chaos-orc-0.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:18 +0000 UTC Normal Pod self-healing-chaos-orc-0.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 428.821828ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:18 +0000 UTC Normal Pod self-healing-chaos-orc-0.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:18 +0000 UTC Normal Pod self-healing-chaos-orc-0.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:21 +0000 UTC Normal PersistentVolumeClaim datadir-self-healing-chaos-mysql-0 ProvisioningSucceeded Successfully provisioned volume pvc-d4736157-3199-48bf-89c0-6f7ffa82c1a8 logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:21 +0000 UTC Normal Pod self-healing-chaos-mysql-0 Scheduled Successfully assigned kuttl-test-set-bluejay/self-healing-chaos-mysql-0 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w381 logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:25 +0000 UTC Normal Pod self-healing-chaos-orc-0.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:25 +0000 UTC Normal Pod self-healing-chaos-orc-0.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 366.661355ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:25 +0000 UTC Normal Pod self-healing-chaos-orc-0.spec.containers{orc} Created Created container orc logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:25 +0000 UTC Normal Pod self-healing-chaos-orc-0.spec.containers{orc} Started Started container orc logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:25 +0000 UTC Normal Pod self-healing-chaos-orc-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:26 +0000 UTC Normal Pod self-healing-chaos-orc-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 376.775047ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:26 +0000 UTC Normal Pod self-healing-chaos-orc-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:26 +0000 UTC Normal Pod self-healing-chaos-orc-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:29 +0000 UTC Normal Pod self-healing-chaos-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-d4736157-3199-48bf-89c0-6f7ffa82c1a8" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:33 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:34 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 442.745889ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:34 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:34 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:38 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:38 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 388.423871ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:38 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{mysql} Created Created container mysql logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:38 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{mysql} Started Started container mysql logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:38 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:39 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 353.912123ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:39 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:39 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:39 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:39 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 374.607197ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:39 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:39 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:57 +0000 UTC Normal Pod self-healing-chaos-orc-1 Scheduled Successfully assigned kuttl-test-set-bluejay/self-healing-chaos-orc-1 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:57 +0000 UTC Normal StatefulSet.apps self-healing-chaos-orc SuccessfulCreate create Pod self-healing-chaos-orc-1 in StatefulSet self-healing-chaos-orc successful logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:58 +0000 UTC Normal Pod self-healing-chaos-orc-1.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:58 +0000 UTC Normal Pod self-healing-chaos-orc-1.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 434.130699ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:58 +0000 UTC Normal Pod self-healing-chaos-orc-1.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:11:58 +0000 UTC Normal Pod self-healing-chaos-orc-1.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:00 +0000 UTC Normal Pod self-healing-chaos-orc-1.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:01 +0000 UTC Normal Pod self-healing-chaos-orc-1.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 408.707788ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:01 +0000 UTC Normal Pod self-healing-chaos-orc-1.spec.containers{orc} Created Created container orc logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:01 +0000 UTC Normal Pod self-healing-chaos-orc-1.spec.containers{orc} Started Started container orc logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:01 +0000 UTC Normal Pod self-healing-chaos-orc-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:01 +0000 UTC Normal Pod self-healing-chaos-orc-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 373.622577ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:01 +0000 UTC Normal Pod self-healing-chaos-orc-1.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:01 +0000 UTC Normal Pod self-healing-chaos-orc-1.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:08 +0000 UTC Normal PersistentVolumeClaim datadir-self-healing-chaos-mysql-1 WaitForFirstConsumer waiting for first consumer to be created before binding logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:08 +0000 UTC Normal PersistentVolumeClaim datadir-self-healing-chaos-mysql-1 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:08 +0000 UTC Normal PersistentVolumeClaim datadir-self-healing-chaos-mysql-1 Provisioning External provisioner is provisioning volume for claim "kuttl-test-set-bluejay/datadir-self-healing-chaos-mysql-1" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:08 +0000 UTC Normal StatefulSet.apps self-healing-chaos-mysql SuccessfulCreate create Claim datadir-self-healing-chaos-mysql-1 Pod self-healing-chaos-mysql-1 in StatefulSet self-healing-chaos-mysql success logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:08 +0000 UTC Normal StatefulSet.apps self-healing-chaos-mysql SuccessfulCreate create Pod self-healing-chaos-mysql-1 in StatefulSet self-healing-chaos-mysql successful logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:11 +0000 UTC Normal Pod self-healing-chaos-haproxy-0 Scheduled Successfully assigned kuttl-test-set-bluejay/self-healing-chaos-haproxy-0 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:11 +0000 UTC Normal StatefulSet.apps self-healing-chaos-haproxy SuccessfulCreate create Pod self-healing-chaos-haproxy-0 in StatefulSet self-healing-chaos-haproxy successful logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:12 +0000 UTC Normal PersistentVolumeClaim datadir-self-healing-chaos-mysql-1 ProvisioningSucceeded Successfully provisioned volume pvc-033d6b90-dd49-4680-9bd6-48ed01823e81 logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:12 +0000 UTC Normal Pod self-healing-chaos-haproxy-0.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:12 +0000 UTC Normal Pod self-healing-chaos-haproxy-0.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 415.105202ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:12 +0000 UTC Normal Pod self-healing-chaos-haproxy-0.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:12 +0000 UTC Normal Pod self-healing-chaos-haproxy-0.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:12 +0000 UTC Normal Pod self-healing-chaos-mysql-1 Scheduled Successfully assigned kuttl-test-set-bluejay/self-healing-chaos-mysql-1 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:14 +0000 UTC Normal Pod self-healing-chaos-haproxy-0.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:15 +0000 UTC Normal Pod self-healing-chaos-haproxy-0.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 372.206926ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:15 +0000 UTC Normal Pod self-healing-chaos-haproxy-0.spec.containers{haproxy} Created Created container haproxy logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:15 +0000 UTC Normal Pod self-healing-chaos-haproxy-0.spec.containers{haproxy} Started Started container haproxy logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:15 +0000 UTC Normal Pod self-healing-chaos-haproxy-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:15 +0000 UTC Normal Pod self-healing-chaos-haproxy-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 384.476988ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:15 +0000 UTC Normal Pod self-healing-chaos-haproxy-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:15 +0000 UTC Normal Pod self-healing-chaos-haproxy-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:15 +0000 UTC Normal Pod self-healing-chaos-haproxy-1 Scheduled Successfully assigned kuttl-test-set-bluejay/self-healing-chaos-haproxy-1 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w381 logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:15 +0000 UTC Normal StatefulSet.apps self-healing-chaos-haproxy SuccessfulCreate create Pod self-healing-chaos-haproxy-1 in StatefulSet self-healing-chaos-haproxy successful logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:16 +0000 UTC Normal Pod self-healing-chaos-haproxy-1.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:17 +0000 UTC Normal Pod self-healing-chaos-haproxy-1.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 404.382118ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:17 +0000 UTC Normal Pod self-healing-chaos-haproxy-1.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:17 +0000 UTC Normal Pod self-healing-chaos-haproxy-1.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:17 +0000 UTC Normal Pod self-healing-chaos-mysql-1 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-033d6b90-dd49-4680-9bd6-48ed01823e81" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:19 +0000 UTC Normal Pod self-healing-chaos-haproxy-1.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:19 +0000 UTC Normal Pod self-healing-chaos-haproxy-1.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 373.782964ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:19 +0000 UTC Normal Pod self-healing-chaos-haproxy-1.spec.containers{haproxy} Created Created container haproxy logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:19 +0000 UTC Normal Pod self-healing-chaos-haproxy-1.spec.containers{haproxy} Started Started container haproxy logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:19 +0000 UTC Normal Pod self-healing-chaos-haproxy-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:19 +0000 UTC Normal Pod self-healing-chaos-mysql-1.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:19 +0000 UTC Normal Pod self-healing-chaos-mysql-1.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 422.211481ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:19 +0000 UTC Normal Pod self-healing-chaos-mysql-1.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:19 +0000 UTC Normal Pod self-healing-chaos-mysql-1.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:20 +0000 UTC Normal Pod self-healing-chaos-haproxy-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 381.004029ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:20 +0000 UTC Normal Pod self-healing-chaos-haproxy-1.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:20 +0000 UTC Normal Pod self-healing-chaos-haproxy-1.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:20 +0000 UTC Normal Pod self-healing-chaos-haproxy-2 Scheduled Successfully assigned kuttl-test-set-bluejay/self-healing-chaos-haproxy-2 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w76w logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:20 +0000 UTC Normal StatefulSet.apps self-healing-chaos-haproxy SuccessfulCreate create Pod self-healing-chaos-haproxy-2 in StatefulSet self-healing-chaos-haproxy successful logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:21 +0000 UTC Normal Pod self-healing-chaos-haproxy-2.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:21 +0000 UTC Normal Pod self-healing-chaos-mysql-1.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:21 +0000 UTC Normal Pod self-healing-chaos-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 408.137834ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:21 +0000 UTC Normal Pod self-healing-chaos-mysql-1.spec.containers{mysql} Created Created container mysql logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:21 +0000 UTC Normal Pod self-healing-chaos-mysql-1.spec.containers{mysql} Started Started container mysql logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:21 +0000 UTC Normal Pod self-healing-chaos-mysql-1.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:22 +0000 UTC Normal Pod self-healing-chaos-haproxy-2.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 382.389976ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:22 +0000 UTC Normal Pod self-healing-chaos-haproxy-2.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:22 +0000 UTC Normal Pod self-healing-chaos-haproxy-2.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:22 +0000 UTC Normal Pod self-healing-chaos-mysql-1.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 766.207785ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:22 +0000 UTC Normal Pod self-healing-chaos-mysql-1.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:22 +0000 UTC Normal Pod self-healing-chaos-mysql-1.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:22 +0000 UTC Normal Pod self-healing-chaos-mysql-1.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:23 +0000 UTC Normal Pod self-healing-chaos-mysql-1.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 1.151284301s logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:24 +0000 UTC Normal Pod self-healing-chaos-haproxy-2.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:24 +0000 UTC Normal Pod self-healing-chaos-haproxy-2.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 391.412634ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:24 +0000 UTC Normal Pod self-healing-chaos-haproxy-2.spec.containers{haproxy} Created Created container haproxy logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:24 +0000 UTC Normal Pod self-healing-chaos-haproxy-2.spec.containers{haproxy} Started Started container haproxy logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:24 +0000 UTC Normal Pod self-healing-chaos-haproxy-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:24 +0000 UTC Normal Pod self-healing-chaos-mysql-1.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:24 +0000 UTC Normal Pod self-healing-chaos-mysql-1.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:25 +0000 UTC Normal Pod self-healing-chaos-haproxy-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 449.628712ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:25 +0000 UTC Normal Pod self-healing-chaos-haproxy-2.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:25 +0000 UTC Normal Pod self-healing-chaos-haproxy-2.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:32 +0000 UTC Normal Pod self-healing-chaos-orc-2 Scheduled Successfully assigned kuttl-test-set-bluejay/self-healing-chaos-orc-2 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w381 logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:32 +0000 UTC Normal StatefulSet.apps self-healing-chaos-orc SuccessfulCreate create Pod self-healing-chaos-orc-2 in StatefulSet self-healing-chaos-orc successful logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:33 +0000 UTC Normal Pod self-healing-chaos-orc-2.spec.initContainers{orc-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:34 +0000 UTC Normal Pod self-healing-chaos-orc-2.spec.initContainers{orc-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 412.342698ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:34 +0000 UTC Normal Pod self-healing-chaos-orc-2.spec.initContainers{orc-init} Created Created container orc-init logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:34 +0000 UTC Normal Pod self-healing-chaos-orc-2.spec.initContainers{orc-init} Started Started container orc-init logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:36 +0000 UTC Normal Pod self-healing-chaos-orc-2.spec.containers{orc} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:36 +0000 UTC Normal Pod self-healing-chaos-orc-2.spec.containers{orc} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 414.325484ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:36 +0000 UTC Normal Pod self-healing-chaos-orc-2.spec.containers{orc} Created Created container orc logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:37 +0000 UTC Normal Pod self-healing-chaos-orc-2.spec.containers{orc} Started Started container orc logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:37 +0000 UTC Normal Pod self-healing-chaos-orc-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:37 +0000 UTC Normal Pod self-healing-chaos-orc-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 408.402806ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:37 +0000 UTC Normal Pod self-healing-chaos-orc-2.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:37 +0000 UTC Normal Pod self-healing-chaos-orc-2.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:39 +0000 UTC Warning Pod self-healing-chaos-mysql-1.spec.containers{mysql} Unhealthy Startup probe failed: logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:39 +0000 UTC Normal Pod self-healing-chaos-mysql-1.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:12:43 +0000 UTC Normal Pod self-healing-chaos-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 422.991125ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:13 +0000 UTC Normal PersistentVolumeClaim datadir-self-healing-chaos-mysql-2 WaitForFirstConsumer waiting for first consumer to be created before binding logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:13 +0000 UTC Normal PersistentVolumeClaim datadir-self-healing-chaos-mysql-2 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:13 +0000 UTC Normal PersistentVolumeClaim datadir-self-healing-chaos-mysql-2 Provisioning External provisioner is provisioning volume for claim "kuttl-test-set-bluejay/datadir-self-healing-chaos-mysql-2" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:13 +0000 UTC Normal StatefulSet.apps self-healing-chaos-mysql SuccessfulCreate create Claim datadir-self-healing-chaos-mysql-2 Pod self-healing-chaos-mysql-2 in StatefulSet self-healing-chaos-mysql success logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:13 +0000 UTC Normal StatefulSet.apps self-healing-chaos-mysql SuccessfulCreate create Pod self-healing-chaos-mysql-2 in StatefulSet self-healing-chaos-mysql successful logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:17 +0000 UTC Normal PersistentVolumeClaim datadir-self-healing-chaos-mysql-2 ProvisioningSucceeded Successfully provisioned volume pvc-9c1de7ff-bd3a-46e0-b2c0-6849157eb30c logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:18 +0000 UTC Normal Pod self-healing-chaos-mysql-2 Scheduled Successfully assigned kuttl-test-set-bluejay/self-healing-chaos-mysql-2 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-6qj2 logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:26 +0000 UTC Normal Pod self-healing-chaos-mysql-2 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-9c1de7ff-bd3a-46e0-b2c0-6849157eb30c" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:27 +0000 UTC Normal Pod self-healing-chaos-mysql-2.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:28 +0000 UTC Normal Pod self-healing-chaos-mysql-2.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 417.350289ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:28 +0000 UTC Normal Pod self-healing-chaos-mysql-2.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:28 +0000 UTC Normal Pod self-healing-chaos-mysql-2.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:30 +0000 UTC Normal Pod self-healing-chaos-mysql-2.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:30 +0000 UTC Normal Pod self-healing-chaos-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 395.168487ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:30 +0000 UTC Normal Pod self-healing-chaos-mysql-2.spec.containers{mysql} Created Created container mysql logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:30 +0000 UTC Normal Pod self-healing-chaos-mysql-2.spec.containers{mysql} Started Started container mysql logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:30 +0000 UTC Normal Pod self-healing-chaos-mysql-2.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:30 +0000 UTC Normal Pod self-healing-chaos-mysql-2.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 370.925998ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:30 +0000 UTC Normal Pod self-healing-chaos-mysql-2.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:31 +0000 UTC Normal Pod self-healing-chaos-mysql-2.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:31 +0000 UTC Normal Pod self-healing-chaos-mysql-2.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:31 +0000 UTC Normal Pod self-healing-chaos-mysql-2.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 375.491456ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:31 +0000 UTC Normal Pod self-healing-chaos-mysql-2.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:31 +0000 UTC Normal Pod self-healing-chaos-mysql-2.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:48 +0000 UTC Warning Pod self-healing-chaos-mysql-2.spec.containers{mysql} Unhealthy Startup probe failed: logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:48 +0000 UTC Normal Pod self-healing-chaos-mysql-2.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:13:51 +0000 UTC Normal Pod self-healing-chaos-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 429.086123ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:47 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-kill FinalizerInited Finalizer has been inited logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:47 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-kill Updated Successfully update finalizer of resource logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:47 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-kill Updated Successfully update desiredPhase of resource logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:47 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-kill Applied Successfully apply chaos for kuttl-test-set-bluejay/self-healing-chaos-mysql-0 logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:47 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-kill Updated Successfully update records of resource logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:47 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{mysql} Killing Stopping container mysql logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:47 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{pt-heartbeat} Killing Stopping container pt-heartbeat logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:47 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{xtrabackup} Killing Stopping container xtrabackup logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:48 +0000 UTC Warning Pod self-healing-chaos-mysql-0.spec.containers{mysql} Unhealthy Readiness probe failed: 2023/08/09 13:14:48 readiness check failed: connect to db: ping database: dial tcp 10.167.209.39:33062: connect: connection refused logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:48 +0000 UTC Normal Pod self-healing-chaos-mysql-0 Scheduled Successfully assigned kuttl-test-set-bluejay/self-healing-chaos-mysql-0 to gke-jen-ps-339-3a8483b-1-default-pool-82bc8527-w381 logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:48 +0000 UTC Warning Endpoints self-healing-chaos-mysql FailedToUpdateEndpoint Failed to update endpoint kuttl-test-set-bluejay/self-healing-chaos-mysql: Operation cannot be fulfilled on endpoints "self-healing-chaos-mysql": the object has been modified; please apply your changes to the latest version and try again logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:49 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:50 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 403.632359ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:50 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:50 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:51 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:52 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 402.066743ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:52 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{mysql} Created Created container mysql logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:52 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{mysql} Started Started container mysql logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:52 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:52 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 398.68057ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:52 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:53 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:53 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:53 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 416.558518ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:53 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{pt-heartbeat} Created Created container pt-heartbeat logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:14:53 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{pt-heartbeat} Started Started container pt-heartbeat logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:15:10 +0000 UTC Warning Pod self-healing-chaos-mysql-0.spec.containers{mysql} Unhealthy Startup probe failed: logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:15:10 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:15:13 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 397.544314ms logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:16:16 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-failure FinalizerInited Finalizer has been inited logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:16:16 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-failure Updated Successfully update finalizer of resource logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:16:16 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-failure Started Experiment has started logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:16:16 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-failure Updated Successfully update desiredPhase of resource logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:16:17 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-failure Applied Successfully apply chaos for kuttl-test-set-bluejay/self-healing-chaos-mysql-0 logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:16:17 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-failure Updated Successfully update records of resource logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:16:17 +0000 UTC Normal Pod self-healing-chaos-mysql-0.spec.containers{mysql} Killing Container mysql definition changed, will be restarted logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:16:19 +0000 UTC Warning Pod self-healing-chaos-mysql-0.spec.containers{mysql} Unhealthy Readiness probe failed: 2023/08/09 13:16:19 readiness check failed: connect to db: ping database: dial tcp 10.167.209.42:33062: connect: connection refused logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:16:19 +0000 UTC Warning Pod self-healing-chaos-mysql-0.spec.containers{mysql} Unhealthy Liveness probe failed: 2023/08/09 13:16:19 liveness check failed: connect to db: ping database: dial tcp 10.167.209.42:33062: connect: connection refused logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:17:16 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-failure TimeUp Time up according to the duration logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:17:16 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-failure Updated Successfully update desiredPhase of resource logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:17:16 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-failure Recovered Successfully recover chaos for kuttl-test-set-bluejay/self-healing-chaos-mysql-0 logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:17:16 +0000 UTC Normal PodChaos.chaos-mesh.org chaos-pod-failure Updated Successfully update records of resource logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:18:49 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss FinalizerInited Finalizer has been inited logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:18:49 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss Updated Successfully update finalizer of resource logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:18:49 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss Started Experiment has started logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:18:49 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss Updated Successfully update desiredPhase of resource logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:18:49 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss Updated Successfully update records of resource logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:18:50 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss Applied Successfully apply chaos for kuttl-test-set-bluejay/self-healing-chaos-mysql-0 logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:18:50 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss Updated Successfully update records of resource logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:18:50 +0000 UTC Normal PodNetworkChaos.chaos-mesh.org self-healing-chaos-mysql-0 Updated Successfully update ObservedGeneration and FailedMessage of resource logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:19:49 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss TimeUp Time up according to the duration logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:19:49 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss Updated Successfully update desiredPhase of resource logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:19:49 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss Updated Successfully update records of resource logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:19:49 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss Recovered Successfully recover chaos for kuttl-test-set-bluejay/self-healing-chaos-mysql-0 logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:19:49 +0000 UTC Normal NetworkChaos.chaos-mesh.org chaos-pod-network-loss Updated Successfully update records of resource logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:19:49 +0000 UTC Normal PodNetworkChaos.chaos-mesh.org self-healing-chaos-mysql-0 Updated Successfully update ObservedGeneration and FailedMessage of resource logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:20:27 +0000 UTC Normal Pod chaos-controller-manager-66674457c8-bh98d.spec.containers{chaos-mesh} Killing Stopping container chaos-mesh logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:20:27 +0000 UTC Normal Pod chaos-controller-manager-66674457c8-gl4lm.spec.containers{chaos-mesh} Killing Stopping container chaos-mesh logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:20:27 +0000 UTC Normal Pod chaos-controller-manager-66674457c8-r5pl8.spec.containers{chaos-mesh} Killing Stopping container chaos-mesh logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:20:27 +0000 UTC Normal Pod chaos-daemon-dtcv6.spec.containers{chaos-daemon} Killing Stopping container chaos-daemon logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:20:27 +0000 UTC Normal Pod chaos-daemon-hd6dz.spec.containers{chaos-daemon} Killing Stopping container chaos-daemon logger.go:42: 13:22:03 | self-healing-chaos | 2023-08-09 13:20:27 +0000 UTC Normal Pod chaos-daemon-mdxm8.spec.containers{chaos-daemon} Killing Stopping container chaos-daemon logger.go:42: 13:22:03 | self-healing-chaos | Deleting namespace: kuttl-test-set-bluejay === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (781.94s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/self-healing-chaos (778.22s) PASS