=== RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://34.30.151.116 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 180 seconds for each step harness.go:372: testsuite: e2e-tests/tests has 26 tests === RUN kuttl/harness === RUN kuttl/harness/gr-self-healing-chaos === PAUSE kuttl/harness/gr-self-healing-chaos === CONT kuttl/harness/gr-self-healing-chaos logger.go:42: 12:56:47 | gr-self-healing-chaos | Creating namespace: kuttl-test-alive-corgi logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | starting test step 0-deploy-operator logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | running command: [sh -c set -o errexit set -o xtrace source ../../functions deploy_operator deploy_non_tls_cluster_secrets deploy_tls_cluster_secrets deploy_client] logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | + source ../../functions logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ realpath ../../.. logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | ++++ mktemp -d logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ export TEMP_DIR=/tmp/tmp.DogDwvOPKD logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ TEMP_DIR=/tmp/tmp.DogDwvOPKD logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ GIT_BRANCH=PR-339 logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ VERSION=PR-339-3a8483b logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | ++++ which gdate logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | ++++ which date logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ date=/usr/bin/date logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ command -v oc logger.go:42: 12:56:47 | gr-self-healing-chaos/0-deploy-operator | +++ oc get projects logger.go:42: 12:56:53 | gr-self-healing-chaos/0-deploy-operator | error: the server doesn't have a resource type "projects" logger.go:42: 12:56:53 | gr-self-healing-chaos/0-deploy-operator | +++ grep '^minikube' logger.go:42: 12:56:53 | gr-self-healing-chaos/0-deploy-operator | +++ kubectl get nodes logger.go:42: 12:56:54 | gr-self-healing-chaos/0-deploy-operator | ++++ pwd logger.go:42: 12:56:54 | gr-self-healing-chaos/0-deploy-operator | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/gr-self-healing-chaos logger.go:42: 12:56:54 | gr-self-healing-chaos/0-deploy-operator | ++ test_name=gr-self-healing-chaos logger.go:42: 12:56:54 | gr-self-healing-chaos/0-deploy-operator | + deploy_operator logger.go:42: 12:56:54 | gr-self-healing-chaos/0-deploy-operator | + kubectl -n kuttl-test-alive-corgi apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/crd.yaml logger.go:42: 12:56:55 | gr-self-healing-chaos/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlbackups.ps.percona.com serverside-applied logger.go:42: 12:56:55 | gr-self-healing-chaos/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlrestores.ps.percona.com serverside-applied logger.go:42: 12:56:56 | gr-self-healing-chaos/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqls.ps.percona.com serverside-applied logger.go:42: 12:56:56 | gr-self-healing-chaos/0-deploy-operator | + kubectl -n kuttl-test-alive-corgi apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/rbac.yaml logger.go:42: 12:56:57 | gr-self-healing-chaos/0-deploy-operator | serviceaccount/percona-server-mysql-operator created logger.go:42: 12:56:58 | gr-self-healing-chaos/0-deploy-operator | serviceaccount/percona-server-mysql-operator-orchestrator created logger.go:42: 12:56:58 | gr-self-healing-chaos/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 12:56:58 | gr-self-healing-chaos/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 12:56:59 | gr-self-healing-chaos/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 12:56:59 | gr-self-healing-chaos/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 12:56:59 | gr-self-healing-chaos/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator created logger.go:42: 12:57:00 | gr-self-healing-chaos/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-orchestrator created logger.go:42: 12:57:00 | gr-self-healing-chaos/0-deploy-operator | + yq eval '(select(documentIndex==1).spec.template.spec.containers[] | select(.name=="manager").env[] | select(.name=="DISABLE_TELEMETRY").value) = "true"' logger.go:42: 12:57:00 | gr-self-healing-chaos/0-deploy-operator | + kubectl -n kuttl-test-alive-corgi apply -f - logger.go:42: 12:57:00 | gr-self-healing-chaos/0-deploy-operator | ++ printf 'select(documentIndex==1).spec.template.spec.containers[0].image="%s"' perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:57:00 | gr-self-healing-chaos/0-deploy-operator | + yq eval 'select(documentIndex==1).spec.template.spec.containers[0].image="perconalab/percona-server-mysql-operator:PR-339-3a8483b"' /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/operator.yaml logger.go:42: 12:57:01 | gr-self-healing-chaos/0-deploy-operator | configmap/percona-server-mysql-operator-config created logger.go:42: 12:57:01 | gr-self-healing-chaos/0-deploy-operator | deployment.apps/percona-server-mysql-operator created logger.go:42: 12:57:01 | gr-self-healing-chaos/0-deploy-operator | + deploy_non_tls_cluster_secrets logger.go:42: 12:57:01 | gr-self-healing-chaos/0-deploy-operator | + kubectl -n kuttl-test-alive-corgi apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/secrets.yaml logger.go:42: 12:57:03 | gr-self-healing-chaos/0-deploy-operator | secret/test-secrets created logger.go:42: 12:57:03 | gr-self-healing-chaos/0-deploy-operator | + deploy_tls_cluster_secrets logger.go:42: 12:57:03 | gr-self-healing-chaos/0-deploy-operator | + kubectl -n kuttl-test-alive-corgi apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/ssl-secret.yaml logger.go:42: 12:57:04 | gr-self-healing-chaos/0-deploy-operator | secret/test-ssl created logger.go:42: 12:57:04 | gr-self-healing-chaos/0-deploy-operator | + deploy_client logger.go:42: 12:57:04 | gr-self-healing-chaos/0-deploy-operator | + kubectl -n kuttl-test-alive-corgi apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf/client.yaml logger.go:42: 12:57:05 | gr-self-healing-chaos/0-deploy-operator | pod/mysql-client created logger.go:42: 12:57:13 | gr-self-healing-chaos/0-deploy-operator | test step completed 0-deploy-operator logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | starting test step 1-create-cluster logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | running command: [sh -c set -o errexit set -o xtrace source ../../functions deploy_chaos_mesh "${NAMESPACE}" get_cr \ | yq eval '.spec.mysql.clusterType="group-replication"' - \ | kubectl -n "${NAMESPACE}" apply -f - sleep 60] logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | + source ../../functions logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ realpath ../../.. logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/vars.sh logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-339 logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/conf logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | ++++ mktemp -d logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ export TEMP_DIR=/tmp/tmp.NqVMRYcL1P logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ TEMP_DIR=/tmp/tmp.NqVMRYcL1P logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ export GIT_BRANCH=PR-339 logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ GIT_BRANCH=PR-339 logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ export VERSION=PR-339-3a8483b logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ VERSION=PR-339-3a8483b logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | ++++ which gdate logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-339/bin/:/home/ec2-user/google-cloud-sdk/bin:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | ++++ which date logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ date=/usr/bin/date logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ command -v oc logger.go:42: 12:57:13 | gr-self-healing-chaos/1-create-cluster | +++ oc get projects logger.go:42: 12:57:18 | gr-self-healing-chaos/1-create-cluster | error: the server doesn't have a resource type "projects" logger.go:42: 12:57:18 | gr-self-healing-chaos/1-create-cluster | +++ kubectl get nodes logger.go:42: 12:57:18 | gr-self-healing-chaos/1-create-cluster | +++ grep '^minikube' logger.go:42: 12:57:19 | gr-self-healing-chaos/1-create-cluster | ++++ pwd logger.go:42: 12:57:19 | gr-self-healing-chaos/1-create-cluster | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-339/e2e-tests/tests/gr-self-healing-chaos logger.go:42: 12:57:19 | gr-self-healing-chaos/1-create-cluster | ++ test_name=gr-self-healing-chaos logger.go:42: 12:57:19 | gr-self-healing-chaos/1-create-cluster | + deploy_chaos_mesh kuttl-test-alive-corgi logger.go:42: 12:57:19 | gr-self-healing-chaos/1-create-cluster | + local chaos_mesh_ns=kuttl-test-alive-corgi logger.go:42: 12:57:19 | gr-self-healing-chaos/1-create-cluster | + destroy_chaos_mesh logger.go:42: 12:57:19 | gr-self-healing-chaos/1-create-cluster | + local chaos_mesh_ns logger.go:42: 12:57:19 | gr-self-healing-chaos/1-create-cluster | ++ helm list --all-namespaces --filter chaos-mesh logger.go:42: 12:57:19 | gr-self-healing-chaos/1-create-cluster | ++ tail -n1 logger.go:42: 12:57:19 | gr-self-healing-chaos/1-create-cluster | ++ awk '-F ' '{print $2}' logger.go:42: 12:57:19 | gr-self-healing-chaos/1-create-cluster | ++ sed s/NAMESPACE// logger.go:42: 12:57:19 | gr-self-healing-chaos/1-create-cluster | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 12:57:19 | gr-self-healing-chaos/1-create-cluster | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 12:57:20 | gr-self-healing-chaos/1-create-cluster | + chaos_mesh_ns= logger.go:42: 12:57:20 | gr-self-healing-chaos/1-create-cluster | + '[' -n '' ']' logger.go:42: 12:57:20 | gr-self-healing-chaos/1-create-cluster | ++ kubectl api-resources logger.go:42: 12:57:20 | gr-self-healing-chaos/1-create-cluster | ++ grep chaos-mesh logger.go:42: 12:57:20 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:57:21 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 12:57:21 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer awschaos --all logger.go:42: 12:57:21 | gr-self-healing-chaos/1-create-cluster | + local resource=awschaos logger.go:42: 12:57:21 | gr-self-healing-chaos/1-create-cluster | + local name=--all logger.go:42: 12:57:21 | gr-self-healing-chaos/1-create-cluster | + '[' --all == --all ']' logger.go:42: 12:57:21 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get awschaos logger.go:42: 12:57:21 | gr-self-healing-chaos/1-create-cluster | ++ grep -v NAME logger.go:42: 12:57:21 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:57:21 | gr-self-healing-chaos/1-create-cluster | No resources found in default namespace. logger.go:42: 12:57:21 | gr-self-healing-chaos/1-create-cluster | + return logger.go:42: 12:57:21 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 12:57:21 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer azurechaos --all logger.go:42: 12:57:21 | gr-self-healing-chaos/1-create-cluster | + local resource=azurechaos logger.go:42: 12:57:21 | gr-self-healing-chaos/1-create-cluster | + local name=--all logger.go:42: 12:57:21 | gr-self-healing-chaos/1-create-cluster | + '[' --all == --all ']' logger.go:42: 12:57:21 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get azurechaos logger.go:42: 12:57:21 | gr-self-healing-chaos/1-create-cluster | ++ grep -v NAME logger.go:42: 12:57:21 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:57:22 | gr-self-healing-chaos/1-create-cluster | No resources found in default namespace. logger.go:42: 12:57:22 | gr-self-healing-chaos/1-create-cluster | + return logger.go:42: 12:57:22 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 12:57:22 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer blockchaos --all logger.go:42: 12:57:22 | gr-self-healing-chaos/1-create-cluster | + local resource=blockchaos logger.go:42: 12:57:22 | gr-self-healing-chaos/1-create-cluster | + local name=--all logger.go:42: 12:57:22 | gr-self-healing-chaos/1-create-cluster | + '[' --all == --all ']' logger.go:42: 12:57:22 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get blockchaos logger.go:42: 12:57:22 | gr-self-healing-chaos/1-create-cluster | ++ grep -v NAME logger.go:42: 12:57:22 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:57:23 | gr-self-healing-chaos/1-create-cluster | No resources found in default namespace. logger.go:42: 12:57:23 | gr-self-healing-chaos/1-create-cluster | + return logger.go:42: 12:57:23 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 12:57:23 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer dnschaos --all logger.go:42: 12:57:23 | gr-self-healing-chaos/1-create-cluster | + local resource=dnschaos logger.go:42: 12:57:23 | gr-self-healing-chaos/1-create-cluster | + local name=--all logger.go:42: 12:57:23 | gr-self-healing-chaos/1-create-cluster | + '[' --all == --all ']' logger.go:42: 12:57:23 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get dnschaos logger.go:42: 12:57:23 | gr-self-healing-chaos/1-create-cluster | ++ grep -v NAME logger.go:42: 12:57:23 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:57:24 | gr-self-healing-chaos/1-create-cluster | No resources found in default namespace. logger.go:42: 12:57:24 | gr-self-healing-chaos/1-create-cluster | + return logger.go:42: 12:57:24 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 12:57:24 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer gcpchaos --all logger.go:42: 12:57:24 | gr-self-healing-chaos/1-create-cluster | + local resource=gcpchaos logger.go:42: 12:57:24 | gr-self-healing-chaos/1-create-cluster | + local name=--all logger.go:42: 12:57:24 | gr-self-healing-chaos/1-create-cluster | + '[' --all == --all ']' logger.go:42: 12:57:24 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get gcpchaos logger.go:42: 12:57:24 | gr-self-healing-chaos/1-create-cluster | ++ grep -v NAME logger.go:42: 12:57:24 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:57:24 | gr-self-healing-chaos/1-create-cluster | No resources found in default namespace. logger.go:42: 12:57:24 | gr-self-healing-chaos/1-create-cluster | + return logger.go:42: 12:57:24 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 12:57:24 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer httpchaos --all logger.go:42: 12:57:24 | gr-self-healing-chaos/1-create-cluster | + local resource=httpchaos logger.go:42: 12:57:24 | gr-self-healing-chaos/1-create-cluster | + local name=--all logger.go:42: 12:57:24 | gr-self-healing-chaos/1-create-cluster | + '[' --all == --all ']' logger.go:42: 12:57:24 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get httpchaos logger.go:42: 12:57:24 | gr-self-healing-chaos/1-create-cluster | ++ grep -v NAME logger.go:42: 12:57:24 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:57:25 | gr-self-healing-chaos/1-create-cluster | No resources found in default namespace. logger.go:42: 12:57:25 | gr-self-healing-chaos/1-create-cluster | + return logger.go:42: 12:57:25 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 12:57:25 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer iochaos --all logger.go:42: 12:57:25 | gr-self-healing-chaos/1-create-cluster | + local resource=iochaos logger.go:42: 12:57:25 | gr-self-healing-chaos/1-create-cluster | + local name=--all logger.go:42: 12:57:25 | gr-self-healing-chaos/1-create-cluster | + '[' --all == --all ']' logger.go:42: 12:57:25 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get iochaos logger.go:42: 12:57:25 | gr-self-healing-chaos/1-create-cluster | ++ grep -v NAME logger.go:42: 12:57:25 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:57:26 | gr-self-healing-chaos/1-create-cluster | No resources found in default namespace. logger.go:42: 12:57:26 | gr-self-healing-chaos/1-create-cluster | + return logger.go:42: 12:57:26 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 12:57:26 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer jvmchaos --all logger.go:42: 12:57:26 | gr-self-healing-chaos/1-create-cluster | + local resource=jvmchaos logger.go:42: 12:57:26 | gr-self-healing-chaos/1-create-cluster | + local name=--all logger.go:42: 12:57:26 | gr-self-healing-chaos/1-create-cluster | + '[' --all == --all ']' logger.go:42: 12:57:26 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get jvmchaos logger.go:42: 12:57:26 | gr-self-healing-chaos/1-create-cluster | ++ grep -v NAME logger.go:42: 12:57:26 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:57:27 | gr-self-healing-chaos/1-create-cluster | No resources found in default namespace. logger.go:42: 12:57:27 | gr-self-healing-chaos/1-create-cluster | + return logger.go:42: 12:57:27 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 12:57:27 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer kernelchaos --all logger.go:42: 12:57:27 | gr-self-healing-chaos/1-create-cluster | + local resource=kernelchaos logger.go:42: 12:57:27 | gr-self-healing-chaos/1-create-cluster | + local name=--all logger.go:42: 12:57:27 | gr-self-healing-chaos/1-create-cluster | + '[' --all == --all ']' logger.go:42: 12:57:27 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get kernelchaos logger.go:42: 12:57:27 | gr-self-healing-chaos/1-create-cluster | ++ grep -v NAME logger.go:42: 12:57:27 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:57:27 | gr-self-healing-chaos/1-create-cluster | No resources found in default namespace. logger.go:42: 12:57:27 | gr-self-healing-chaos/1-create-cluster | + return logger.go:42: 12:57:27 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 12:57:27 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer networkchaos --all logger.go:42: 12:57:27 | gr-self-healing-chaos/1-create-cluster | + local resource=networkchaos logger.go:42: 12:57:27 | gr-self-healing-chaos/1-create-cluster | + local name=--all logger.go:42: 12:57:27 | gr-self-healing-chaos/1-create-cluster | + '[' --all == --all ']' logger.go:42: 12:57:27 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get networkchaos logger.go:42: 12:57:27 | gr-self-healing-chaos/1-create-cluster | ++ grep -v NAME logger.go:42: 12:57:27 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:57:28 | gr-self-healing-chaos/1-create-cluster | No resources found in default namespace. logger.go:42: 12:57:28 | gr-self-healing-chaos/1-create-cluster | + return logger.go:42: 12:57:28 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 12:57:28 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer physicalmachinechaos --all logger.go:42: 12:57:28 | gr-self-healing-chaos/1-create-cluster | + local resource=physicalmachinechaos logger.go:42: 12:57:28 | gr-self-healing-chaos/1-create-cluster | + local name=--all logger.go:42: 12:57:28 | gr-self-healing-chaos/1-create-cluster | + '[' --all == --all ']' logger.go:42: 12:57:28 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get physicalmachinechaos logger.go:42: 12:57:28 | gr-self-healing-chaos/1-create-cluster | ++ grep -v NAME logger.go:42: 12:57:28 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:57:29 | gr-self-healing-chaos/1-create-cluster | No resources found in default namespace. logger.go:42: 12:57:29 | gr-self-healing-chaos/1-create-cluster | + return logger.go:42: 12:57:29 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 12:57:29 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer physicalmachines --all logger.go:42: 12:57:29 | gr-self-healing-chaos/1-create-cluster | + local resource=physicalmachines logger.go:42: 12:57:29 | gr-self-healing-chaos/1-create-cluster | + local name=--all logger.go:42: 12:57:29 | gr-self-healing-chaos/1-create-cluster | + '[' --all == --all ']' logger.go:42: 12:57:29 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get physicalmachines logger.go:42: 12:57:29 | gr-self-healing-chaos/1-create-cluster | ++ grep -v NAME logger.go:42: 12:57:29 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:57:29 | gr-self-healing-chaos/1-create-cluster | No resources found in default namespace. logger.go:42: 12:57:29 | gr-self-healing-chaos/1-create-cluster | + return logger.go:42: 12:57:29 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 12:57:29 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer podchaos --all logger.go:42: 12:57:29 | gr-self-healing-chaos/1-create-cluster | + local resource=podchaos logger.go:42: 12:57:29 | gr-self-healing-chaos/1-create-cluster | + local name=--all logger.go:42: 12:57:29 | gr-self-healing-chaos/1-create-cluster | + '[' --all == --all ']' logger.go:42: 12:57:29 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get podchaos logger.go:42: 12:57:29 | gr-self-healing-chaos/1-create-cluster | ++ grep -v NAME logger.go:42: 12:57:29 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:57:30 | gr-self-healing-chaos/1-create-cluster | No resources found in default namespace. logger.go:42: 12:57:30 | gr-self-healing-chaos/1-create-cluster | + return logger.go:42: 12:57:30 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 12:57:30 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer podhttpchaos --all logger.go:42: 12:57:30 | gr-self-healing-chaos/1-create-cluster | + local resource=podhttpchaos logger.go:42: 12:57:30 | gr-self-healing-chaos/1-create-cluster | + local name=--all logger.go:42: 12:57:30 | gr-self-healing-chaos/1-create-cluster | + '[' --all == --all ']' logger.go:42: 12:57:30 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get podhttpchaos logger.go:42: 12:57:30 | gr-self-healing-chaos/1-create-cluster | ++ grep -v NAME logger.go:42: 12:57:30 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:57:31 | gr-self-healing-chaos/1-create-cluster | No resources found in default namespace. logger.go:42: 12:57:31 | gr-self-healing-chaos/1-create-cluster | + return logger.go:42: 12:57:31 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 12:57:31 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer podiochaos --all logger.go:42: 12:57:31 | gr-self-healing-chaos/1-create-cluster | + local resource=podiochaos logger.go:42: 12:57:31 | gr-self-healing-chaos/1-create-cluster | + local name=--all logger.go:42: 12:57:31 | gr-self-healing-chaos/1-create-cluster | + '[' --all == --all ']' logger.go:42: 12:57:31 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get podiochaos logger.go:42: 12:57:31 | gr-self-healing-chaos/1-create-cluster | ++ grep -v NAME logger.go:42: 12:57:31 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:57:32 | gr-self-healing-chaos/1-create-cluster | No resources found in default namespace. logger.go:42: 12:57:32 | gr-self-healing-chaos/1-create-cluster | + return logger.go:42: 12:57:32 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 12:57:32 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer podnetworkchaos --all logger.go:42: 12:57:32 | gr-self-healing-chaos/1-create-cluster | + local resource=podnetworkchaos logger.go:42: 12:57:32 | gr-self-healing-chaos/1-create-cluster | + local name=--all logger.go:42: 12:57:32 | gr-self-healing-chaos/1-create-cluster | + '[' --all == --all ']' logger.go:42: 12:57:32 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get podnetworkchaos logger.go:42: 12:57:32 | gr-self-healing-chaos/1-create-cluster | ++ grep -v NAME logger.go:42: 12:57:32 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:57:32 | gr-self-healing-chaos/1-create-cluster | No resources found in default namespace. logger.go:42: 12:57:32 | gr-self-healing-chaos/1-create-cluster | + return logger.go:42: 12:57:32 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 12:57:32 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer remoteclusters --all logger.go:42: 12:57:32 | gr-self-healing-chaos/1-create-cluster | + local resource=remoteclusters logger.go:42: 12:57:32 | gr-self-healing-chaos/1-create-cluster | + local name=--all logger.go:42: 12:57:32 | gr-self-healing-chaos/1-create-cluster | + '[' --all == --all ']' logger.go:42: 12:57:32 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get remoteclusters logger.go:42: 12:57:32 | gr-self-healing-chaos/1-create-cluster | ++ grep -v NAME logger.go:42: 12:57:32 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:57:33 | gr-self-healing-chaos/1-create-cluster | No resources found logger.go:42: 12:57:33 | gr-self-healing-chaos/1-create-cluster | + return logger.go:42: 12:57:33 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 12:57:33 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer schedules --all logger.go:42: 12:57:33 | gr-self-healing-chaos/1-create-cluster | + local resource=schedules logger.go:42: 12:57:33 | gr-self-healing-chaos/1-create-cluster | + local name=--all logger.go:42: 12:57:33 | gr-self-healing-chaos/1-create-cluster | + '[' --all == --all ']' logger.go:42: 12:57:33 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get schedules logger.go:42: 12:57:33 | gr-self-healing-chaos/1-create-cluster | ++ grep -v NAME logger.go:42: 12:57:33 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:57:34 | gr-self-healing-chaos/1-create-cluster | No resources found in default namespace. logger.go:42: 12:57:34 | gr-self-healing-chaos/1-create-cluster | + return logger.go:42: 12:57:34 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 12:57:34 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer statuschecks --all logger.go:42: 12:57:34 | gr-self-healing-chaos/1-create-cluster | + local resource=statuschecks logger.go:42: 12:57:34 | gr-self-healing-chaos/1-create-cluster | + local name=--all logger.go:42: 12:57:34 | gr-self-healing-chaos/1-create-cluster | + '[' --all == --all ']' logger.go:42: 12:57:34 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get statuschecks logger.go:42: 12:57:34 | gr-self-healing-chaos/1-create-cluster | ++ grep -v NAME logger.go:42: 12:57:34 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:57:35 | gr-self-healing-chaos/1-create-cluster | No resources found in default namespace. logger.go:42: 12:57:35 | gr-self-healing-chaos/1-create-cluster | + return logger.go:42: 12:57:35 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 12:57:35 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer stresschaos --all logger.go:42: 12:57:35 | gr-self-healing-chaos/1-create-cluster | + local resource=stresschaos logger.go:42: 12:57:35 | gr-self-healing-chaos/1-create-cluster | + local name=--all logger.go:42: 12:57:35 | gr-self-healing-chaos/1-create-cluster | + '[' --all == --all ']' logger.go:42: 12:57:35 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get stresschaos logger.go:42: 12:57:35 | gr-self-healing-chaos/1-create-cluster | ++ grep -v NAME logger.go:42: 12:57:35 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:57:35 | gr-self-healing-chaos/1-create-cluster | No resources found in default namespace. logger.go:42: 12:57:35 | gr-self-healing-chaos/1-create-cluster | + return logger.go:42: 12:57:35 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 12:57:35 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer timechaos --all logger.go:42: 12:57:35 | gr-self-healing-chaos/1-create-cluster | + local resource=timechaos logger.go:42: 12:57:35 | gr-self-healing-chaos/1-create-cluster | + local name=--all logger.go:42: 12:57:35 | gr-self-healing-chaos/1-create-cluster | + '[' --all == --all ']' logger.go:42: 12:57:35 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get timechaos logger.go:42: 12:57:35 | gr-self-healing-chaos/1-create-cluster | ++ grep -v NAME logger.go:42: 12:57:35 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:57:36 | gr-self-healing-chaos/1-create-cluster | No resources found in default namespace. logger.go:42: 12:57:36 | gr-self-healing-chaos/1-create-cluster | + return logger.go:42: 12:57:36 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 12:57:36 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer workflownodes --all logger.go:42: 12:57:36 | gr-self-healing-chaos/1-create-cluster | + local resource=workflownodes logger.go:42: 12:57:36 | gr-self-healing-chaos/1-create-cluster | + local name=--all logger.go:42: 12:57:36 | gr-self-healing-chaos/1-create-cluster | + '[' --all == --all ']' logger.go:42: 12:57:36 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get workflownodes logger.go:42: 12:57:36 | gr-self-healing-chaos/1-create-cluster | ++ grep -v NAME logger.go:42: 12:57:36 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:57:37 | gr-self-healing-chaos/1-create-cluster | No resources found in default namespace. logger.go:42: 12:57:37 | gr-self-healing-chaos/1-create-cluster | + return logger.go:42: 12:57:37 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl api-resources | grep chaos-mesh | awk '\''{print $1}'\'')' logger.go:42: 12:57:37 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer workflows --all logger.go:42: 12:57:37 | gr-self-healing-chaos/1-create-cluster | + local resource=workflows logger.go:42: 12:57:37 | gr-self-healing-chaos/1-create-cluster | + local name=--all logger.go:42: 12:57:37 | gr-self-healing-chaos/1-create-cluster | + '[' --all == --all ']' logger.go:42: 12:57:37 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get workflows logger.go:42: 12:57:37 | gr-self-healing-chaos/1-create-cluster | ++ grep -v NAME logger.go:42: 12:57:37 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:57:37 | gr-self-healing-chaos/1-create-cluster | No resources found in default namespace. logger.go:42: 12:57:37 | gr-self-healing-chaos/1-create-cluster | + return logger.go:42: 12:57:37 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get crd -A logger.go:42: 12:57:37 | gr-self-healing-chaos/1-create-cluster | ++ grep chaos-mesh.org logger.go:42: 12:57:37 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:57:39 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:57:39 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer crd awschaos.chaos-mesh.org logger.go:42: 12:57:39 | gr-self-healing-chaos/1-create-cluster | + local resource=crd logger.go:42: 12:57:39 | gr-self-healing-chaos/1-create-cluster | + local name=awschaos.chaos-mesh.org logger.go:42: 12:57:39 | gr-self-healing-chaos/1-create-cluster | + '[' awschaos.chaos-mesh.org == --all ']' logger.go:42: 12:57:39 | gr-self-healing-chaos/1-create-cluster | + kubectl delete crd awschaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 12:57:39 | gr-self-healing-chaos/1-create-cluster | customresourcedefinition.apiextensions.k8s.io "awschaos.chaos-mesh.org" deleted logger.go:42: 12:57:39 | gr-self-healing-chaos/1-create-cluster | + kubectl patch crd awschaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:57:40 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "awschaos.chaos-mesh.org" not found logger.go:42: 12:57:40 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:57:40 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete crd awschaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 12:57:41 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:57:41 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer crd azurechaos.chaos-mesh.org logger.go:42: 12:57:41 | gr-self-healing-chaos/1-create-cluster | + local resource=crd logger.go:42: 12:57:41 | gr-self-healing-chaos/1-create-cluster | + local name=azurechaos.chaos-mesh.org logger.go:42: 12:57:41 | gr-self-healing-chaos/1-create-cluster | + '[' azurechaos.chaos-mesh.org == --all ']' logger.go:42: 12:57:41 | gr-self-healing-chaos/1-create-cluster | + kubectl delete crd azurechaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 12:57:42 | gr-self-healing-chaos/1-create-cluster | customresourcedefinition.apiextensions.k8s.io "azurechaos.chaos-mesh.org" deleted logger.go:42: 12:57:42 | gr-self-healing-chaos/1-create-cluster | + kubectl patch crd azurechaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:57:42 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "azurechaos.chaos-mesh.org" not found logger.go:42: 12:57:42 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:57:42 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete crd azurechaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 12:57:43 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:57:43 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer crd blockchaos.chaos-mesh.org logger.go:42: 12:57:43 | gr-self-healing-chaos/1-create-cluster | + local resource=crd logger.go:42: 12:57:43 | gr-self-healing-chaos/1-create-cluster | + local name=blockchaos.chaos-mesh.org logger.go:42: 12:57:43 | gr-self-healing-chaos/1-create-cluster | + '[' blockchaos.chaos-mesh.org == --all ']' logger.go:42: 12:57:43 | gr-self-healing-chaos/1-create-cluster | + kubectl delete crd blockchaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 12:57:44 | gr-self-healing-chaos/1-create-cluster | customresourcedefinition.apiextensions.k8s.io "blockchaos.chaos-mesh.org" deleted logger.go:42: 12:57:44 | gr-self-healing-chaos/1-create-cluster | + kubectl patch crd blockchaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:57:45 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "blockchaos.chaos-mesh.org" not found logger.go:42: 12:57:45 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:57:45 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete crd blockchaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 12:57:46 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:57:46 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer crd dnschaos.chaos-mesh.org logger.go:42: 12:57:46 | gr-self-healing-chaos/1-create-cluster | + local resource=crd logger.go:42: 12:57:46 | gr-self-healing-chaos/1-create-cluster | + local name=dnschaos.chaos-mesh.org logger.go:42: 12:57:46 | gr-self-healing-chaos/1-create-cluster | + '[' dnschaos.chaos-mesh.org == --all ']' logger.go:42: 12:57:46 | gr-self-healing-chaos/1-create-cluster | + kubectl delete crd dnschaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 12:57:47 | gr-self-healing-chaos/1-create-cluster | customresourcedefinition.apiextensions.k8s.io "dnschaos.chaos-mesh.org" deleted logger.go:42: 12:57:47 | gr-self-healing-chaos/1-create-cluster | + kubectl patch crd dnschaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:57:47 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "dnschaos.chaos-mesh.org" not found logger.go:42: 12:57:47 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:57:47 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete crd dnschaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 12:57:49 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:57:49 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer crd gcpchaos.chaos-mesh.org logger.go:42: 12:57:49 | gr-self-healing-chaos/1-create-cluster | + local resource=crd logger.go:42: 12:57:49 | gr-self-healing-chaos/1-create-cluster | + local name=gcpchaos.chaos-mesh.org logger.go:42: 12:57:49 | gr-self-healing-chaos/1-create-cluster | + '[' gcpchaos.chaos-mesh.org == --all ']' logger.go:42: 12:57:49 | gr-self-healing-chaos/1-create-cluster | + kubectl delete crd gcpchaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 12:57:50 | gr-self-healing-chaos/1-create-cluster | customresourcedefinition.apiextensions.k8s.io "gcpchaos.chaos-mesh.org" deleted logger.go:42: 12:57:50 | gr-self-healing-chaos/1-create-cluster | + kubectl patch crd gcpchaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:57:50 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "gcpchaos.chaos-mesh.org" not found logger.go:42: 12:57:50 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:57:50 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete crd gcpchaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 12:57:51 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:57:51 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer crd httpchaos.chaos-mesh.org logger.go:42: 12:57:51 | gr-self-healing-chaos/1-create-cluster | + local resource=crd logger.go:42: 12:57:51 | gr-self-healing-chaos/1-create-cluster | + local name=httpchaos.chaos-mesh.org logger.go:42: 12:57:51 | gr-self-healing-chaos/1-create-cluster | + '[' httpchaos.chaos-mesh.org == --all ']' logger.go:42: 12:57:51 | gr-self-healing-chaos/1-create-cluster | + kubectl delete crd httpchaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 12:57:52 | gr-self-healing-chaos/1-create-cluster | customresourcedefinition.apiextensions.k8s.io "httpchaos.chaos-mesh.org" deleted logger.go:42: 12:57:52 | gr-self-healing-chaos/1-create-cluster | + kubectl patch crd httpchaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:57:53 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "httpchaos.chaos-mesh.org" not found logger.go:42: 12:57:53 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:57:53 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete crd httpchaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 12:57:54 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:57:54 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer crd iochaos.chaos-mesh.org logger.go:42: 12:57:54 | gr-self-healing-chaos/1-create-cluster | + local resource=crd logger.go:42: 12:57:54 | gr-self-healing-chaos/1-create-cluster | + local name=iochaos.chaos-mesh.org logger.go:42: 12:57:54 | gr-self-healing-chaos/1-create-cluster | + '[' iochaos.chaos-mesh.org == --all ']' logger.go:42: 12:57:54 | gr-self-healing-chaos/1-create-cluster | + kubectl delete crd iochaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 12:57:55 | gr-self-healing-chaos/1-create-cluster | customresourcedefinition.apiextensions.k8s.io "iochaos.chaos-mesh.org" deleted logger.go:42: 12:57:55 | gr-self-healing-chaos/1-create-cluster | + kubectl patch crd iochaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:57:55 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "iochaos.chaos-mesh.org" not found logger.go:42: 12:57:55 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:57:55 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete crd iochaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 12:57:56 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:57:56 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer crd jvmchaos.chaos-mesh.org logger.go:42: 12:57:56 | gr-self-healing-chaos/1-create-cluster | + local resource=crd logger.go:42: 12:57:56 | gr-self-healing-chaos/1-create-cluster | + local name=jvmchaos.chaos-mesh.org logger.go:42: 12:57:56 | gr-self-healing-chaos/1-create-cluster | + '[' jvmchaos.chaos-mesh.org == --all ']' logger.go:42: 12:57:56 | gr-self-healing-chaos/1-create-cluster | + kubectl delete crd jvmchaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 12:57:57 | gr-self-healing-chaos/1-create-cluster | customresourcedefinition.apiextensions.k8s.io "jvmchaos.chaos-mesh.org" deleted logger.go:42: 12:57:57 | gr-self-healing-chaos/1-create-cluster | + kubectl patch crd jvmchaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:57:58 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "jvmchaos.chaos-mesh.org" not found logger.go:42: 12:57:58 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:57:58 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete crd jvmchaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 12:57:59 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:57:59 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer crd kernelchaos.chaos-mesh.org logger.go:42: 12:57:59 | gr-self-healing-chaos/1-create-cluster | + local resource=crd logger.go:42: 12:57:59 | gr-self-healing-chaos/1-create-cluster | + local name=kernelchaos.chaos-mesh.org logger.go:42: 12:57:59 | gr-self-healing-chaos/1-create-cluster | + '[' kernelchaos.chaos-mesh.org == --all ']' logger.go:42: 12:57:59 | gr-self-healing-chaos/1-create-cluster | + kubectl delete crd kernelchaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 12:57:59 | gr-self-healing-chaos/1-create-cluster | customresourcedefinition.apiextensions.k8s.io "kernelchaos.chaos-mesh.org" deleted logger.go:42: 12:57:59 | gr-self-healing-chaos/1-create-cluster | + kubectl patch crd kernelchaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:58:00 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "kernelchaos.chaos-mesh.org" not found logger.go:42: 12:58:00 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:58:00 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete crd kernelchaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 12:58:01 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:58:01 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer crd networkchaos.chaos-mesh.org logger.go:42: 12:58:01 | gr-self-healing-chaos/1-create-cluster | + local resource=crd logger.go:42: 12:58:01 | gr-self-healing-chaos/1-create-cluster | + local name=networkchaos.chaos-mesh.org logger.go:42: 12:58:01 | gr-self-healing-chaos/1-create-cluster | + '[' networkchaos.chaos-mesh.org == --all ']' logger.go:42: 12:58:01 | gr-self-healing-chaos/1-create-cluster | + kubectl delete crd networkchaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 12:58:02 | gr-self-healing-chaos/1-create-cluster | customresourcedefinition.apiextensions.k8s.io "networkchaos.chaos-mesh.org" deleted logger.go:42: 12:58:02 | gr-self-healing-chaos/1-create-cluster | + kubectl patch crd networkchaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:58:03 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "networkchaos.chaos-mesh.org" not found logger.go:42: 12:58:03 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:58:03 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete crd networkchaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 12:58:04 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:58:04 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer crd physicalmachinechaos.chaos-mesh.org logger.go:42: 12:58:04 | gr-self-healing-chaos/1-create-cluster | + local resource=crd logger.go:42: 12:58:04 | gr-self-healing-chaos/1-create-cluster | + local name=physicalmachinechaos.chaos-mesh.org logger.go:42: 12:58:04 | gr-self-healing-chaos/1-create-cluster | + '[' physicalmachinechaos.chaos-mesh.org == --all ']' logger.go:42: 12:58:04 | gr-self-healing-chaos/1-create-cluster | + kubectl delete crd physicalmachinechaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 12:58:05 | gr-self-healing-chaos/1-create-cluster | customresourcedefinition.apiextensions.k8s.io "physicalmachinechaos.chaos-mesh.org" deleted logger.go:42: 12:58:05 | gr-self-healing-chaos/1-create-cluster | + kubectl patch crd physicalmachinechaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:58:06 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "physicalmachinechaos.chaos-mesh.org" not found logger.go:42: 12:58:06 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:58:06 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete crd physicalmachinechaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 12:58:06 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:58:06 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer crd physicalmachines.chaos-mesh.org logger.go:42: 12:58:06 | gr-self-healing-chaos/1-create-cluster | + local resource=crd logger.go:42: 12:58:06 | gr-self-healing-chaos/1-create-cluster | + local name=physicalmachines.chaos-mesh.org logger.go:42: 12:58:06 | gr-self-healing-chaos/1-create-cluster | + '[' physicalmachines.chaos-mesh.org == --all ']' logger.go:42: 12:58:06 | gr-self-healing-chaos/1-create-cluster | + kubectl delete crd physicalmachines.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 12:58:07 | gr-self-healing-chaos/1-create-cluster | customresourcedefinition.apiextensions.k8s.io "physicalmachines.chaos-mesh.org" deleted logger.go:42: 12:58:07 | gr-self-healing-chaos/1-create-cluster | + kubectl patch crd physicalmachines.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:58:08 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "physicalmachines.chaos-mesh.org" not found logger.go:42: 12:58:08 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:58:08 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete crd physicalmachines.chaos-mesh.org --ignore-not-found=true logger.go:42: 12:58:09 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:58:09 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer crd podchaos.chaos-mesh.org logger.go:42: 12:58:09 | gr-self-healing-chaos/1-create-cluster | + local resource=crd logger.go:42: 12:58:09 | gr-self-healing-chaos/1-create-cluster | + local name=podchaos.chaos-mesh.org logger.go:42: 12:58:09 | gr-self-healing-chaos/1-create-cluster | + '[' podchaos.chaos-mesh.org == --all ']' logger.go:42: 12:58:09 | gr-self-healing-chaos/1-create-cluster | + kubectl delete crd podchaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 12:58:10 | gr-self-healing-chaos/1-create-cluster | customresourcedefinition.apiextensions.k8s.io "podchaos.chaos-mesh.org" deleted logger.go:42: 12:58:10 | gr-self-healing-chaos/1-create-cluster | + kubectl patch crd podchaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:58:10 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "podchaos.chaos-mesh.org" not found logger.go:42: 12:58:10 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:58:10 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete crd podchaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 12:58:11 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:58:11 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer crd podhttpchaos.chaos-mesh.org logger.go:42: 12:58:11 | gr-self-healing-chaos/1-create-cluster | + local resource=crd logger.go:42: 12:58:11 | gr-self-healing-chaos/1-create-cluster | + local name=podhttpchaos.chaos-mesh.org logger.go:42: 12:58:11 | gr-self-healing-chaos/1-create-cluster | + '[' podhttpchaos.chaos-mesh.org == --all ']' logger.go:42: 12:58:11 | gr-self-healing-chaos/1-create-cluster | + kubectl delete crd podhttpchaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 12:58:12 | gr-self-healing-chaos/1-create-cluster | customresourcedefinition.apiextensions.k8s.io "podhttpchaos.chaos-mesh.org" deleted logger.go:42: 12:58:12 | gr-self-healing-chaos/1-create-cluster | + kubectl patch crd podhttpchaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:58:13 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "podhttpchaos.chaos-mesh.org" not found logger.go:42: 12:58:13 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:58:13 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete crd podhttpchaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 12:58:14 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:58:14 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer crd podiochaos.chaos-mesh.org logger.go:42: 12:58:14 | gr-self-healing-chaos/1-create-cluster | + local resource=crd logger.go:42: 12:58:14 | gr-self-healing-chaos/1-create-cluster | + local name=podiochaos.chaos-mesh.org logger.go:42: 12:58:14 | gr-self-healing-chaos/1-create-cluster | + '[' podiochaos.chaos-mesh.org == --all ']' logger.go:42: 12:58:14 | gr-self-healing-chaos/1-create-cluster | + kubectl delete crd podiochaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 12:58:15 | gr-self-healing-chaos/1-create-cluster | customresourcedefinition.apiextensions.k8s.io "podiochaos.chaos-mesh.org" deleted logger.go:42: 12:58:15 | gr-self-healing-chaos/1-create-cluster | + kubectl patch crd podiochaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:58:15 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "podiochaos.chaos-mesh.org" not found logger.go:42: 12:58:15 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:58:15 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete crd podiochaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 12:58:16 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:58:16 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer crd podnetworkchaos.chaos-mesh.org logger.go:42: 12:58:16 | gr-self-healing-chaos/1-create-cluster | + local resource=crd logger.go:42: 12:58:16 | gr-self-healing-chaos/1-create-cluster | + local name=podnetworkchaos.chaos-mesh.org logger.go:42: 12:58:16 | gr-self-healing-chaos/1-create-cluster | + '[' podnetworkchaos.chaos-mesh.org == --all ']' logger.go:42: 12:58:16 | gr-self-healing-chaos/1-create-cluster | + kubectl delete crd podnetworkchaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 12:58:17 | gr-self-healing-chaos/1-create-cluster | customresourcedefinition.apiextensions.k8s.io "podnetworkchaos.chaos-mesh.org" deleted logger.go:42: 12:58:17 | gr-self-healing-chaos/1-create-cluster | + kubectl patch crd podnetworkchaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:58:18 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "podnetworkchaos.chaos-mesh.org" not found logger.go:42: 12:58:18 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:58:18 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete crd podnetworkchaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 12:58:19 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:58:19 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer crd remoteclusters.chaos-mesh.org logger.go:42: 12:58:19 | gr-self-healing-chaos/1-create-cluster | + local resource=crd logger.go:42: 12:58:19 | gr-self-healing-chaos/1-create-cluster | + local name=remoteclusters.chaos-mesh.org logger.go:42: 12:58:19 | gr-self-healing-chaos/1-create-cluster | + '[' remoteclusters.chaos-mesh.org == --all ']' logger.go:42: 12:58:19 | gr-self-healing-chaos/1-create-cluster | + kubectl delete crd remoteclusters.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 12:58:20 | gr-self-healing-chaos/1-create-cluster | customresourcedefinition.apiextensions.k8s.io "remoteclusters.chaos-mesh.org" deleted logger.go:42: 12:58:20 | gr-self-healing-chaos/1-create-cluster | + kubectl patch crd remoteclusters.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:58:20 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "remoteclusters.chaos-mesh.org" not found logger.go:42: 12:58:20 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:58:20 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete crd remoteclusters.chaos-mesh.org --ignore-not-found=true logger.go:42: 12:58:21 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:58:21 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer crd schedules.chaos-mesh.org logger.go:42: 12:58:21 | gr-self-healing-chaos/1-create-cluster | + local resource=crd logger.go:42: 12:58:21 | gr-self-healing-chaos/1-create-cluster | + local name=schedules.chaos-mesh.org logger.go:42: 12:58:21 | gr-self-healing-chaos/1-create-cluster | + '[' schedules.chaos-mesh.org == --all ']' logger.go:42: 12:58:21 | gr-self-healing-chaos/1-create-cluster | + kubectl delete crd schedules.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 12:58:23 | gr-self-healing-chaos/1-create-cluster | customresourcedefinition.apiextensions.k8s.io "schedules.chaos-mesh.org" deleted logger.go:42: 12:58:23 | gr-self-healing-chaos/1-create-cluster | + kubectl patch crd schedules.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:58:25 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "schedules.chaos-mesh.org" not found logger.go:42: 12:58:25 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:58:25 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete crd schedules.chaos-mesh.org --ignore-not-found=true logger.go:42: 12:58:26 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:58:26 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer crd statuschecks.chaos-mesh.org logger.go:42: 12:58:26 | gr-self-healing-chaos/1-create-cluster | + local resource=crd logger.go:42: 12:58:26 | gr-self-healing-chaos/1-create-cluster | + local name=statuschecks.chaos-mesh.org logger.go:42: 12:58:26 | gr-self-healing-chaos/1-create-cluster | + '[' statuschecks.chaos-mesh.org == --all ']' logger.go:42: 12:58:26 | gr-self-healing-chaos/1-create-cluster | + kubectl delete crd statuschecks.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 12:58:27 | gr-self-healing-chaos/1-create-cluster | customresourcedefinition.apiextensions.k8s.io "statuschecks.chaos-mesh.org" deleted logger.go:42: 12:58:27 | gr-self-healing-chaos/1-create-cluster | + kubectl patch crd statuschecks.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:58:28 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "statuschecks.chaos-mesh.org" not found logger.go:42: 12:58:28 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:58:28 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete crd statuschecks.chaos-mesh.org --ignore-not-found=true logger.go:42: 12:58:28 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:58:28 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer crd stresschaos.chaos-mesh.org logger.go:42: 12:58:28 | gr-self-healing-chaos/1-create-cluster | + local resource=crd logger.go:42: 12:58:28 | gr-self-healing-chaos/1-create-cluster | + local name=stresschaos.chaos-mesh.org logger.go:42: 12:58:28 | gr-self-healing-chaos/1-create-cluster | + '[' stresschaos.chaos-mesh.org == --all ']' logger.go:42: 12:58:28 | gr-self-healing-chaos/1-create-cluster | + kubectl delete crd stresschaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 12:58:29 | gr-self-healing-chaos/1-create-cluster | customresourcedefinition.apiextensions.k8s.io "stresschaos.chaos-mesh.org" deleted logger.go:42: 12:58:29 | gr-self-healing-chaos/1-create-cluster | + kubectl patch crd stresschaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:58:30 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "stresschaos.chaos-mesh.org" not found logger.go:42: 12:58:30 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:58:30 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete crd stresschaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 12:58:31 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:58:31 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer crd timechaos.chaos-mesh.org logger.go:42: 12:58:31 | gr-self-healing-chaos/1-create-cluster | + local resource=crd logger.go:42: 12:58:31 | gr-self-healing-chaos/1-create-cluster | + local name=timechaos.chaos-mesh.org logger.go:42: 12:58:31 | gr-self-healing-chaos/1-create-cluster | + '[' timechaos.chaos-mesh.org == --all ']' logger.go:42: 12:58:31 | gr-self-healing-chaos/1-create-cluster | + kubectl delete crd timechaos.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 12:58:32 | gr-self-healing-chaos/1-create-cluster | customresourcedefinition.apiextensions.k8s.io "timechaos.chaos-mesh.org" deleted logger.go:42: 12:58:32 | gr-self-healing-chaos/1-create-cluster | + kubectl patch crd timechaos.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:58:33 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "timechaos.chaos-mesh.org" not found logger.go:42: 12:58:33 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:58:33 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete crd timechaos.chaos-mesh.org --ignore-not-found=true logger.go:42: 12:58:34 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:58:34 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer crd workflownodes.chaos-mesh.org logger.go:42: 12:58:34 | gr-self-healing-chaos/1-create-cluster | + local resource=crd logger.go:42: 12:58:34 | gr-self-healing-chaos/1-create-cluster | + local name=workflownodes.chaos-mesh.org logger.go:42: 12:58:34 | gr-self-healing-chaos/1-create-cluster | + '[' workflownodes.chaos-mesh.org == --all ']' logger.go:42: 12:58:34 | gr-self-healing-chaos/1-create-cluster | + kubectl delete crd workflownodes.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 12:58:35 | gr-self-healing-chaos/1-create-cluster | customresourcedefinition.apiextensions.k8s.io "workflownodes.chaos-mesh.org" deleted logger.go:42: 12:58:35 | gr-self-healing-chaos/1-create-cluster | + kubectl patch crd workflownodes.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:58:39 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "workflownodes.chaos-mesh.org" not found logger.go:42: 12:58:39 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:58:39 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete crd workflownodes.chaos-mesh.org --ignore-not-found=true logger.go:42: 12:58:40 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get crd -A | grep '\''chaos-mesh.org'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:58:40 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer crd workflows.chaos-mesh.org logger.go:42: 12:58:40 | gr-self-healing-chaos/1-create-cluster | + local resource=crd logger.go:42: 12:58:40 | gr-self-healing-chaos/1-create-cluster | + local name=workflows.chaos-mesh.org logger.go:42: 12:58:40 | gr-self-healing-chaos/1-create-cluster | + '[' workflows.chaos-mesh.org == --all ']' logger.go:42: 12:58:40 | gr-self-healing-chaos/1-create-cluster | + kubectl delete crd workflows.chaos-mesh.org --wait=false --ignore-not-found=true logger.go:42: 12:58:41 | gr-self-healing-chaos/1-create-cluster | customresourcedefinition.apiextensions.k8s.io "workflows.chaos-mesh.org" deleted logger.go:42: 12:58:41 | gr-self-healing-chaos/1-create-cluster | + kubectl patch crd workflows.chaos-mesh.org -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:58:42 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "workflows.chaos-mesh.org" not found logger.go:42: 12:58:42 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:58:42 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete crd workflows.chaos-mesh.org --ignore-not-found=true logger.go:42: 12:58:43 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get clusterrolebinding -A logger.go:42: 12:58:43 | gr-self-healing-chaos/1-create-cluster | ++ grep chaos-mesh logger.go:42: 12:58:43 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:58:44 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get clusterrolebinding -A | grep '\''chaos-mesh'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:58:44 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer clusterrolebinding chaos-mesh-chaos-controller-manager-cluster-level logger.go:42: 12:58:44 | gr-self-healing-chaos/1-create-cluster | + local resource=clusterrolebinding logger.go:42: 12:58:44 | gr-self-healing-chaos/1-create-cluster | + local name=chaos-mesh-chaos-controller-manager-cluster-level logger.go:42: 12:58:44 | gr-self-healing-chaos/1-create-cluster | + '[' chaos-mesh-chaos-controller-manager-cluster-level == --all ']' logger.go:42: 12:58:44 | gr-self-healing-chaos/1-create-cluster | + kubectl delete clusterrolebinding chaos-mesh-chaos-controller-manager-cluster-level --wait=false --ignore-not-found=true logger.go:42: 12:58:45 | gr-self-healing-chaos/1-create-cluster | clusterrolebinding.rbac.authorization.k8s.io "chaos-mesh-chaos-controller-manager-cluster-level" deleted logger.go:42: 12:58:45 | gr-self-healing-chaos/1-create-cluster | + kubectl patch clusterrolebinding chaos-mesh-chaos-controller-manager-cluster-level -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:58:45 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "chaos-mesh-chaos-controller-manager-cluster-level" not found logger.go:42: 12:58:45 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:58:45 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete clusterrolebinding chaos-mesh-chaos-controller-manager-cluster-level --ignore-not-found=true logger.go:42: 12:58:46 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get clusterrolebinding -A | grep '\''chaos-mesh'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:58:46 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer clusterrolebinding chaos-mesh-chaos-controller-manager-target-namespace logger.go:42: 12:58:46 | gr-self-healing-chaos/1-create-cluster | + local resource=clusterrolebinding logger.go:42: 12:58:46 | gr-self-healing-chaos/1-create-cluster | + local name=chaos-mesh-chaos-controller-manager-target-namespace logger.go:42: 12:58:46 | gr-self-healing-chaos/1-create-cluster | + '[' chaos-mesh-chaos-controller-manager-target-namespace == --all ']' logger.go:42: 12:58:46 | gr-self-healing-chaos/1-create-cluster | + kubectl delete clusterrolebinding chaos-mesh-chaos-controller-manager-target-namespace --wait=false --ignore-not-found=true logger.go:42: 12:58:47 | gr-self-healing-chaos/1-create-cluster | clusterrolebinding.rbac.authorization.k8s.io "chaos-mesh-chaos-controller-manager-target-namespace" deleted logger.go:42: 12:58:47 | gr-self-healing-chaos/1-create-cluster | + kubectl patch clusterrolebinding chaos-mesh-chaos-controller-manager-target-namespace -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:58:48 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "chaos-mesh-chaos-controller-manager-target-namespace" not found logger.go:42: 12:58:48 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:58:48 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete clusterrolebinding chaos-mesh-chaos-controller-manager-target-namespace --ignore-not-found=true logger.go:42: 12:58:49 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get clusterrolebinding -A | grep '\''chaos-mesh'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:58:49 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer clusterrolebinding chaos-mesh-chaos-dashboard-cluster-level logger.go:42: 12:58:49 | gr-self-healing-chaos/1-create-cluster | + local resource=clusterrolebinding logger.go:42: 12:58:49 | gr-self-healing-chaos/1-create-cluster | + local name=chaos-mesh-chaos-dashboard-cluster-level logger.go:42: 12:58:49 | gr-self-healing-chaos/1-create-cluster | + '[' chaos-mesh-chaos-dashboard-cluster-level == --all ']' logger.go:42: 12:58:49 | gr-self-healing-chaos/1-create-cluster | + kubectl delete clusterrolebinding chaos-mesh-chaos-dashboard-cluster-level --wait=false --ignore-not-found=true logger.go:42: 12:58:49 | gr-self-healing-chaos/1-create-cluster | clusterrolebinding.rbac.authorization.k8s.io "chaos-mesh-chaos-dashboard-cluster-level" deleted logger.go:42: 12:58:49 | gr-self-healing-chaos/1-create-cluster | + kubectl patch clusterrolebinding chaos-mesh-chaos-dashboard-cluster-level -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:58:50 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "chaos-mesh-chaos-dashboard-cluster-level" not found logger.go:42: 12:58:50 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:58:50 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete clusterrolebinding chaos-mesh-chaos-dashboard-cluster-level --ignore-not-found=true logger.go:42: 12:58:51 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get clusterrolebinding -A | grep '\''chaos-mesh'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:58:51 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer clusterrolebinding chaos-mesh-chaos-dashboard-target-namespace logger.go:42: 12:58:51 | gr-self-healing-chaos/1-create-cluster | + local resource=clusterrolebinding logger.go:42: 12:58:51 | gr-self-healing-chaos/1-create-cluster | + local name=chaos-mesh-chaos-dashboard-target-namespace logger.go:42: 12:58:51 | gr-self-healing-chaos/1-create-cluster | + '[' chaos-mesh-chaos-dashboard-target-namespace == --all ']' logger.go:42: 12:58:51 | gr-self-healing-chaos/1-create-cluster | + kubectl delete clusterrolebinding chaos-mesh-chaos-dashboard-target-namespace --wait=false --ignore-not-found=true logger.go:42: 12:58:52 | gr-self-healing-chaos/1-create-cluster | clusterrolebinding.rbac.authorization.k8s.io "chaos-mesh-chaos-dashboard-target-namespace" deleted logger.go:42: 12:58:52 | gr-self-healing-chaos/1-create-cluster | + kubectl patch clusterrolebinding chaos-mesh-chaos-dashboard-target-namespace -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:58:52 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "chaos-mesh-chaos-dashboard-target-namespace" not found logger.go:42: 12:58:52 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:58:52 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete clusterrolebinding chaos-mesh-chaos-dashboard-target-namespace --ignore-not-found=true logger.go:42: 12:58:53 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get clusterrole -A logger.go:42: 12:58:53 | gr-self-healing-chaos/1-create-cluster | ++ grep chaos-mesh logger.go:42: 12:58:53 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:58:54 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get clusterrole -A | grep '\''chaos-mesh'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:58:54 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer clusterrole chaos-mesh-chaos-controller-manager-cluster-level logger.go:42: 12:58:54 | gr-self-healing-chaos/1-create-cluster | + local resource=clusterrole logger.go:42: 12:58:54 | gr-self-healing-chaos/1-create-cluster | + local name=chaos-mesh-chaos-controller-manager-cluster-level logger.go:42: 12:58:54 | gr-self-healing-chaos/1-create-cluster | + '[' chaos-mesh-chaos-controller-manager-cluster-level == --all ']' logger.go:42: 12:58:54 | gr-self-healing-chaos/1-create-cluster | + kubectl delete clusterrole chaos-mesh-chaos-controller-manager-cluster-level --wait=false --ignore-not-found=true logger.go:42: 12:58:55 | gr-self-healing-chaos/1-create-cluster | clusterrole.rbac.authorization.k8s.io "chaos-mesh-chaos-controller-manager-cluster-level" deleted logger.go:42: 12:58:55 | gr-self-healing-chaos/1-create-cluster | + kubectl patch clusterrole chaos-mesh-chaos-controller-manager-cluster-level -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:58:56 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "chaos-mesh-chaos-controller-manager-cluster-level" not found logger.go:42: 12:58:56 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:58:56 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete clusterrole chaos-mesh-chaos-controller-manager-cluster-level --ignore-not-found=true logger.go:42: 12:58:56 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get clusterrole -A | grep '\''chaos-mesh'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:58:56 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer clusterrole chaos-mesh-chaos-controller-manager-target-namespace logger.go:42: 12:58:56 | gr-self-healing-chaos/1-create-cluster | + local resource=clusterrole logger.go:42: 12:58:56 | gr-self-healing-chaos/1-create-cluster | + local name=chaos-mesh-chaos-controller-manager-target-namespace logger.go:42: 12:58:56 | gr-self-healing-chaos/1-create-cluster | + '[' chaos-mesh-chaos-controller-manager-target-namespace == --all ']' logger.go:42: 12:58:56 | gr-self-healing-chaos/1-create-cluster | + kubectl delete clusterrole chaos-mesh-chaos-controller-manager-target-namespace --wait=false --ignore-not-found=true logger.go:42: 12:58:57 | gr-self-healing-chaos/1-create-cluster | clusterrole.rbac.authorization.k8s.io "chaos-mesh-chaos-controller-manager-target-namespace" deleted logger.go:42: 12:58:57 | gr-self-healing-chaos/1-create-cluster | + kubectl patch clusterrole chaos-mesh-chaos-controller-manager-target-namespace -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:58:58 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "chaos-mesh-chaos-controller-manager-target-namespace" not found logger.go:42: 12:58:58 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:58:58 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete clusterrole chaos-mesh-chaos-controller-manager-target-namespace --ignore-not-found=true logger.go:42: 12:58:59 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get clusterrole -A | grep '\''chaos-mesh'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:58:59 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer clusterrole chaos-mesh-chaos-dashboard-cluster-level logger.go:42: 12:58:59 | gr-self-healing-chaos/1-create-cluster | + local resource=clusterrole logger.go:42: 12:58:59 | gr-self-healing-chaos/1-create-cluster | + local name=chaos-mesh-chaos-dashboard-cluster-level logger.go:42: 12:58:59 | gr-self-healing-chaos/1-create-cluster | + '[' chaos-mesh-chaos-dashboard-cluster-level == --all ']' logger.go:42: 12:58:59 | gr-self-healing-chaos/1-create-cluster | + kubectl delete clusterrole chaos-mesh-chaos-dashboard-cluster-level --wait=false --ignore-not-found=true logger.go:42: 12:58:59 | gr-self-healing-chaos/1-create-cluster | clusterrole.rbac.authorization.k8s.io "chaos-mesh-chaos-dashboard-cluster-level" deleted logger.go:42: 12:58:59 | gr-self-healing-chaos/1-create-cluster | + kubectl patch clusterrole chaos-mesh-chaos-dashboard-cluster-level -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:59:00 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "chaos-mesh-chaos-dashboard-cluster-level" not found logger.go:42: 12:59:00 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:59:00 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete clusterrole chaos-mesh-chaos-dashboard-cluster-level --ignore-not-found=true logger.go:42: 12:59:01 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get clusterrole -A | grep '\''chaos-mesh'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:59:01 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer clusterrole chaos-mesh-chaos-dashboard-target-namespace logger.go:42: 12:59:01 | gr-self-healing-chaos/1-create-cluster | + local resource=clusterrole logger.go:42: 12:59:01 | gr-self-healing-chaos/1-create-cluster | + local name=chaos-mesh-chaos-dashboard-target-namespace logger.go:42: 12:59:01 | gr-self-healing-chaos/1-create-cluster | + '[' chaos-mesh-chaos-dashboard-target-namespace == --all ']' logger.go:42: 12:59:01 | gr-self-healing-chaos/1-create-cluster | + kubectl delete clusterrole chaos-mesh-chaos-dashboard-target-namespace --wait=false --ignore-not-found=true logger.go:42: 12:59:02 | gr-self-healing-chaos/1-create-cluster | clusterrole.rbac.authorization.k8s.io "chaos-mesh-chaos-dashboard-target-namespace" deleted logger.go:42: 12:59:02 | gr-self-healing-chaos/1-create-cluster | + kubectl patch clusterrole chaos-mesh-chaos-dashboard-target-namespace -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:59:02 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "chaos-mesh-chaos-dashboard-target-namespace" not found logger.go:42: 12:59:02 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:59:02 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete clusterrole chaos-mesh-chaos-dashboard-target-namespace --ignore-not-found=true logger.go:42: 12:59:03 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get mutatingwebhookconfiguration -A logger.go:42: 12:59:03 | gr-self-healing-chaos/1-create-cluster | ++ grep chaos-mesh logger.go:42: 12:59:03 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:59:04 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get mutatingwebhookconfiguration -A | grep '\''chaos-mesh'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:59:04 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer mutatingwebhookconfiguration chaos-mesh-mutation logger.go:42: 12:59:04 | gr-self-healing-chaos/1-create-cluster | + local resource=mutatingwebhookconfiguration logger.go:42: 12:59:04 | gr-self-healing-chaos/1-create-cluster | + local name=chaos-mesh-mutation logger.go:42: 12:59:04 | gr-self-healing-chaos/1-create-cluster | + '[' chaos-mesh-mutation == --all ']' logger.go:42: 12:59:04 | gr-self-healing-chaos/1-create-cluster | + kubectl delete mutatingwebhookconfiguration chaos-mesh-mutation --wait=false --ignore-not-found=true logger.go:42: 12:59:05 | gr-self-healing-chaos/1-create-cluster | mutatingwebhookconfiguration.admissionregistration.k8s.io "chaos-mesh-mutation" deleted logger.go:42: 12:59:05 | gr-self-healing-chaos/1-create-cluster | + kubectl patch mutatingwebhookconfiguration chaos-mesh-mutation -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:59:06 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): mutatingwebhookconfigurations.admissionregistration.k8s.io "chaos-mesh-mutation" not found logger.go:42: 12:59:06 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:59:06 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete mutatingwebhookconfiguration chaos-mesh-mutation --ignore-not-found=true logger.go:42: 12:59:06 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get validatingwebhookconfiguration -A logger.go:42: 12:59:06 | gr-self-healing-chaos/1-create-cluster | ++ grep chaos-mesh logger.go:42: 12:59:06 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:59:07 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get validatingwebhookconfiguration -A | grep '\''chaos-mesh'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:59:07 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer validatingwebhookconfiguration chaos-mesh-validation logger.go:42: 12:59:07 | gr-self-healing-chaos/1-create-cluster | + local resource=validatingwebhookconfiguration logger.go:42: 12:59:07 | gr-self-healing-chaos/1-create-cluster | + local name=chaos-mesh-validation logger.go:42: 12:59:07 | gr-self-healing-chaos/1-create-cluster | + '[' chaos-mesh-validation == --all ']' logger.go:42: 12:59:07 | gr-self-healing-chaos/1-create-cluster | + kubectl delete validatingwebhookconfiguration chaos-mesh-validation --wait=false --ignore-not-found=true logger.go:42: 12:59:08 | gr-self-healing-chaos/1-create-cluster | validatingwebhookconfiguration.admissionregistration.k8s.io "chaos-mesh-validation" deleted logger.go:42: 12:59:08 | gr-self-healing-chaos/1-create-cluster | + kubectl patch validatingwebhookconfiguration chaos-mesh-validation -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:59:09 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): validatingwebhookconfigurations.admissionregistration.k8s.io "chaos-mesh-validation" not found logger.go:42: 12:59:09 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:59:09 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete validatingwebhookconfiguration chaos-mesh-validation --ignore-not-found=true logger.go:42: 12:59:10 | gr-self-healing-chaos/1-create-cluster | + for i in '$(kubectl get validatingwebhookconfiguration -A | grep '\''chaos-mesh'\'' | awk '\''{print $1}'\'')' logger.go:42: 12:59:10 | gr-self-healing-chaos/1-create-cluster | + delete_resource_with_finalizer validatingwebhookconfiguration chaos-mesh-validation-auth logger.go:42: 12:59:10 | gr-self-healing-chaos/1-create-cluster | + local resource=validatingwebhookconfiguration logger.go:42: 12:59:10 | gr-self-healing-chaos/1-create-cluster | + local name=chaos-mesh-validation-auth logger.go:42: 12:59:10 | gr-self-healing-chaos/1-create-cluster | + '[' chaos-mesh-validation-auth == --all ']' logger.go:42: 12:59:10 | gr-self-healing-chaos/1-create-cluster | + kubectl delete validatingwebhookconfiguration chaos-mesh-validation-auth --wait=false --ignore-not-found=true logger.go:42: 12:59:10 | gr-self-healing-chaos/1-create-cluster | validatingwebhookconfiguration.admissionregistration.k8s.io "chaos-mesh-validation-auth" deleted logger.go:42: 12:59:10 | gr-self-healing-chaos/1-create-cluster | + kubectl patch validatingwebhookconfiguration chaos-mesh-validation-auth -p '{"metadata":{"finalizers":[]}}' --type=merge logger.go:42: 12:59:11 | gr-self-healing-chaos/1-create-cluster | Error from server (NotFound): validatingwebhookconfigurations.admissionregistration.k8s.io "chaos-mesh-validation-auth" not found logger.go:42: 12:59:11 | gr-self-healing-chaos/1-create-cluster | + : logger.go:42: 12:59:11 | gr-self-healing-chaos/1-create-cluster | + timeout 60 kubectl delete validatingwebhookconfiguration chaos-mesh-validation-auth --ignore-not-found=true logger.go:42: 12:59:12 | gr-self-healing-chaos/1-create-cluster | ++ kubectl get validatingwebhookconfiguration -A logger.go:42: 12:59:12 | gr-self-healing-chaos/1-create-cluster | ++ grep validate-auth logger.go:42: 12:59:12 | gr-self-healing-chaos/1-create-cluster | ++ awk '{print $1}' logger.go:42: 12:59:13 | gr-self-healing-chaos/1-create-cluster | + helm repo add chaos-mesh https://charts.chaos-mesh.org logger.go:42: 12:59:13 | gr-self-healing-chaos/1-create-cluster | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 12:59:13 | gr-self-healing-chaos/1-create-cluster | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 12:59:13 | gr-self-healing-chaos/1-create-cluster | "chaos-mesh" already exists with the same configuration, skipping logger.go:42: 12:59:13 | gr-self-healing-chaos/1-create-cluster | + helm install chaos-mesh chaos-mesh/chaos-mesh --namespace=kuttl-test-alive-corgi --set chaosDaemon.runtime=containerd --set chaosDaemon.socketPath=/run/containerd/containerd.sock --set dashboard.create=false --version 2.5.1 logger.go:42: 12:59:13 | gr-self-healing-chaos/1-create-cluster | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 12:59:13 | gr-self-healing-chaos/1-create-cluster | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-ps-operator_PR-339/kubeconfig logger.go:42: 12:59:42 | gr-self-healing-chaos/1-create-cluster | NAME: chaos-mesh logger.go:42: 12:59:42 | gr-self-healing-chaos/1-create-cluster | LAST DEPLOYED: Wed Aug 9 12:59:28 2023 logger.go:42: 12:59:42 | gr-self-healing-chaos/1-create-cluster | NAMESPACE: kuttl-test-alive-corgi logger.go:42: 12:59:42 | gr-self-healing-chaos/1-create-cluster | STATUS: deployed logger.go:42: 12:59:42 | gr-self-healing-chaos/1-create-cluster | REVISION: 1 logger.go:42: 12:59:42 | gr-self-healing-chaos/1-create-cluster | TEST SUITE: None logger.go:42: 12:59:42 | gr-self-healing-chaos/1-create-cluster | NOTES: logger.go:42: 12:59:42 | gr-self-healing-chaos/1-create-cluster | 1. Make sure chaos-mesh components are running logger.go:42: 12:59:42 | gr-self-healing-chaos/1-create-cluster | kubectl get pods --namespace kuttl-test-alive-corgi -l app.kubernetes.io/instance=chaos-mesh logger.go:42: 12:59:42 | gr-self-healing-chaos/1-create-cluster | + sleep 10 logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | + get_cr logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | + local name_suffix= logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | + yq eval '.spec.mysql.clusterType="group-replication"' - logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | + kubectl -n kuttl-test-alive-corgi apply -f - logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | ++ printf '.metadata.name="%s"' gr-self-healing-chaos logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | + yq eval '.metadata.name="gr-self-healing-chaos"' /mnt/jenkins/workspace/cloud-ps-operator_PR-339/deploy/cr.yaml logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | + yq eval '.spec.sslSecretName="test-ssl"' - logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | ++ printf '.spec.mysql.image="%s"' perconalab/percona-server-mysql-operator:main-psmysql logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | + yq eval '.spec.mysql.image="perconalab/percona-server-mysql-operator:main-psmysql"' - logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | + '[' -n '' ']' logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | + yq eval - logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | ++ printf '.spec.pmm.image="%s"' perconalab/pmm-client:dev-latest logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | + yq eval '.spec.pmm.image="perconalab/pmm-client:dev-latest"' - logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | + yq eval '.spec.secretsName="test-secrets"' - logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | ++ printf '.spec.initImage="%s"' perconalab/percona-server-mysql-operator:PR-339-3a8483b logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | + yq eval '.spec.initImage="perconalab/percona-server-mysql-operator:PR-339-3a8483b"' - logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | + yq eval '.spec.upgradeOptions.apply="disabled"' - logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | ++ printf '.spec.proxy.router.image="%s"' perconalab/percona-server-mysql-operator:main-router logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | + yq eval '.spec.proxy.router.image="perconalab/percona-server-mysql-operator:main-router"' - logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | ++ printf '.spec.backup.image="%s"' perconalab/percona-server-mysql-operator:main-backup logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | ++ printf '.spec.orchestrator.image="%s"' perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | + yq eval '.spec.orchestrator.image="perconalab/percona-server-mysql-operator:main-orchestrator"' - logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | + yq eval '.spec.backup.image="perconalab/percona-server-mysql-operator:main-backup"' - logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | ++ printf '.spec.toolkit.image="%s"' perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | + yq eval '.spec.toolkit.image="perconalab/percona-server-mysql-operator:main-toolkit"' - logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | ++ printf '.spec.proxy.haproxy.image="%s"' perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 12:59:52 | gr-self-healing-chaos/1-create-cluster | + yq eval '.spec.proxy.haproxy.image="perconalab/percona-server-mysql-operator:main-haproxy"' - logger.go:42: 12:59:54 | gr-self-healing-chaos/1-create-cluster | perconaservermysql.ps.percona.com/gr-self-healing-chaos created logger.go:42: 12:59:54 | gr-self-healing-chaos/1-create-cluster | + sleep 60 logger.go:42: 13:06:57 | gr-self-healing-chaos/1-create-cluster | test step failed 1-create-cluster case.go:364: failed in step 1-create-cluster case.go:366: deployments.apps "gr-self-healing-chaos-router" not found case.go:366: --- PerconaServerMySQL:kuttl-test-alive-corgi/gr-self-healing-chaos +++ PerconaServerMySQL:kuttl-test-alive-corgi/gr-self-healing-chaos @@ -1,21 +1,292 @@ apiVersion: ps.percona.com/v1alpha1 kind: PerconaServerMySQL metadata: + annotations: + kubectl.kubernetes.io/last-applied-configuration: | + {"apiVersion":"ps.percona.com/v1alpha1","kind":"PerconaServerMySQL","metadata":{"annotations":{},"finalizers":["delete-mysql-pods-in-order"],"name":"gr-self-healing-chaos","namespace":"kuttl-test-alive-corgi"},"spec":{"allowUnsafeConfigurations":false,"backup":{"enabled":true,"image":"perconalab/percona-server-mysql-operator:main-backup","imagePullPolicy":"Always","storages":{"s3-us-west":{"s3":{"bucket":"S3-BACKUP-BUCKET-NAME-HERE","credentialsSecret":"cluster1-s3-credentials","region":"us-west-2"},"type":"s3","verifyTLS":true}}},"crVersion":"0.6.0","initImage":"perconalab/percona-server-mysql-operator:PR-339-3a8483b","mysql":{"affinity":{"antiAffinityTopologyKey":"kubernetes.io/hostname"},"autoRecovery":true,"clusterType":"group-replication","image":"perconalab/percona-server-mysql-operator:main-psmysql","imagePullPolicy":"Always","resources":{"limits":{"memory":"2G"},"requests":{"memory":"1G"}},"size":3,"volumeSpec":{"persistentVolumeClaim":{"resources":{"requests":{"storage":"2G"}}}}},"orchestrator":{"affinity":{"antiAffinityTopologyKey":"kubernetes.io/hostname"},"enabled":false,"image":"perconalab/percona-server-mysql-operator:main-orchestrator","imagePullPolicy":"Always","resources":{"limits":{"memory":"256M"},"requests":{"memory":"128M"}},"size":3,"volumeSpec":{"persistentVolumeClaim":{"resources":{"requests":{"storage":"1G"}}}}},"pmm":{"enabled":false,"image":"perconalab/pmm-client:dev-latest","imagePullPolicy":"Always","resources":{"requests":{"cpu":"300m","memory":"150M"}},"serverHost":"monitoring-service","serverUser":"admin"},"proxy":{"haproxy":{"affinity":{"antiAffinityTopologyKey":"kubernetes.io/hostname"},"enabled":true,"image":"perconalab/percona-server-mysql-operator:main-haproxy","imagePullPolicy":"Always","resources":{"requests":{"cpu":"600m","memory":"1G"}},"size":3},"router":{"affinity":{"antiAffinityTopologyKey":"kubernetes.io/hostname"},"enabled":false,"image":"perconalab/percona-server-mysql-operator:main-router","imagePullPolicy":"Always","resources":{"limits":{"memory":"256M"},"requests":{"memory":"256M"}},"size":3}},"secretsName":"test-secrets","sslSecretName":"test-ssl","toolkit":{"image":"perconalab/percona-server-mysql-operator:main-toolkit","imagePullPolicy":"Always"},"upgradeOptions":{"apply":"disabled","versionServiceEndpoint":"https://check.percona.com"}}} + finalizers: + - delete-mysql-pods-in-order + managedFields: + - apiVersion: ps.percona.com/v1alpha1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:kubectl.kubernetes.io/last-applied-configuration: {} + f:finalizers: + .: {} + v:"delete-mysql-pods-in-order": {} + f:spec: + .: {} + f:allowUnsafeConfigurations: {} + f:backup: + .: {} + f:enabled: {} + f:image: {} + f:imagePullPolicy: {} + f:storages: + .: {} + f:s3-us-west: + .: {} + f:s3: + .: {} + f:bucket: {} + f:credentialsSecret: {} + f:region: {} + f:type: {} + f:verifyTLS: {} + f:crVersion: {} + f:initImage: {} + f:mysql: + .: {} + f:affinity: + .: {} + f:antiAffinityTopologyKey: {} + f:autoRecovery: {} + f:clusterType: {} + f:image: {} + f:imagePullPolicy: {} + f:resources: + .: {} + f:limits: + .: {} + f:memory: {} + f:requests: + .: {} + f:memory: {} + f:size: {} + f:volumeSpec: + .: {} + f:persistentVolumeClaim: + .: {} + f:resources: + .: {} + f:requests: + .: {} + f:storage: {} + f:orchestrator: + .: {} + f:affinity: + .: {} + f:antiAffinityTopologyKey: {} + f:enabled: {} + f:image: {} + f:imagePullPolicy: {} + f:resources: + .: {} + f:limits: + .: {} + f:memory: {} + f:requests: + .: {} + f:memory: {} + f:size: {} + f:volumeSpec: + .: {} + f:persistentVolumeClaim: + .: {} + f:resources: + .: {} + f:requests: + .: {} + f:storage: {} + f:pmm: + .: {} + f:enabled: {} + f:image: {} + f:imagePullPolicy: {} + f:resources: + .: {} + f:requests: + .: {} + f:cpu: {} + f:memory: {} + f:serverHost: {} + f:serverUser: {} + f:proxy: + .: {} + f:haproxy: + .: {} + f:affinity: + .: {} + f:antiAffinityTopologyKey: {} + f:enabled: {} + f:image: {} + f:imagePullPolicy: {} + f:resources: + .: {} + f:requests: + .: {} + f:cpu: {} + f:memory: {} + f:size: {} + f:router: + .: {} + f:affinity: + .: {} + f:antiAffinityTopologyKey: {} + f:enabled: {} + f:image: {} + f:imagePullPolicy: {} + f:resources: + .: {} + f:limits: + .: {} + f:memory: {} + f:requests: + .: {} + f:memory: {} + f:size: {} + f:secretsName: {} + f:sslSecretName: {} + f:toolkit: + .: {} + f:image: {} + f:imagePullPolicy: {} + f:upgradeOptions: + .: {} + f:apply: {} + f:versionServiceEndpoint: {} + manager: kubectl-client-side-apply + operation: Update + time: "2023-08-09T12:59:54Z" + - apiVersion: ps.percona.com/v1alpha1 + fieldsType: FieldsV1 + fieldsV1: + f:status: + .: {} + f:conditions: {} + f:haproxy: + .: {} + f:ready: {} + f:size: {} + f:state: {} + f:host: {} + f:mysql: + .: {} + f:ready: {} + f:size: {} + f:state: {} + f:orchestrator: {} + f:router: {} + f:state: {} + manager: Go-http-client + operation: Update + subresource: status + time: "2023-08-09T13:03:42Z" name: gr-self-healing-chaos namespace: kuttl-test-alive-corgi +spec: + allowUnsafeConfigurations: false + backup: + enabled: true + image: perconalab/percona-server-mysql-operator:main-backup + imagePullPolicy: Always + storages: + s3-us-west: + s3: + bucket: S3-BACKUP-BUCKET-NAME-HERE + credentialsSecret: cluster1-s3-credentials + region: us-west-2 + type: s3 + verifyTLS: true + crVersion: 0.6.0 + initImage: perconalab/percona-server-mysql-operator:PR-339-3a8483b + mysql: + affinity: + antiAffinityTopologyKey: kubernetes.io/hostname + autoRecovery: true + clusterType: group-replication + image: perconalab/percona-server-mysql-operator:main-psmysql + imagePullPolicy: Always + resources: + limits: + memory: 2G + requests: + memory: 1G + size: 3 + volumeSpec: + persistentVolumeClaim: + resources: + requests: + storage: 2G + orchestrator: + affinity: + antiAffinityTopologyKey: kubernetes.io/hostname + enabled: false + image: perconalab/percona-server-mysql-operator:main-orchestrator + imagePullPolicy: Always + resources: + limits: + memory: 256M + requests: + memory: 128M + size: 3 + volumeSpec: + persistentVolumeClaim: + resources: + requests: + storage: 1G + pmm: + enabled: false + image: perconalab/pmm-client:dev-latest + imagePullPolicy: Always + resources: + requests: + cpu: 300m + memory: 150M + serverHost: monitoring-service + serverUser: admin + proxy: + haproxy: + affinity: + antiAffinityTopologyKey: kubernetes.io/hostname + enabled: true + image: perconalab/percona-server-mysql-operator:main-haproxy + imagePullPolicy: Always + resources: + requests: + cpu: 600m + memory: 1G + size: 3 + router: + affinity: + antiAffinityTopologyKey: kubernetes.io/hostname + enabled: false + image: perconalab/percona-server-mysql-operator:main-router + imagePullPolicy: Always + resources: + limits: + memory: 256M + requests: + memory: 256M + size: 3 + secretsName: test-secrets + sslSecretName: test-ssl + toolkit: + image: perconalab/percona-server-mysql-operator:main-toolkit + imagePullPolicy: Always + upgradeOptions: + apply: disabled + versionServiceEndpoint: https://check.percona.com status: conditions: - - message: InnoDB cluster successfully bootstrapped with 3 nodes + - lastTransitionTime: "2023-08-09T13:03:47Z" + message: InnoDB cluster successfully bootstrapped with 3 nodes reason: InnoDBClusterBootstrapped status: "True" type: InnoDBClusterBootstrapped + haproxy: + ready: 3 + size: 3 + state: ready + host: gr-self-healing-chaos-haproxy.kuttl-test-alive-corgi mysql: ready: 3 size: 3 state: ready - router: - ready: 3 - size: 3 - state: ready + orchestrator: {} + router: {} state: ready case.go:366: resource PerconaServerMySQL:kuttl-test-alive-corgi/gr-self-healing-chaos: .status.router.ready: key is missing from map logger.go:42: 13:06:57 | gr-self-healing-chaos | gr-self-healing-chaos events from ns kuttl-test-alive-corgi: logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:57:01 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-hfqqc Scheduled Successfully assigned kuttl-test-alive-corgi/percona-server-mysql-operator-6dcd78d857-hfqqc to gke-jen-ps-339-3a8483b-1-default-pool-7f97af72-klx5 logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:57:01 +0000 UTC Normal ReplicaSet.apps percona-server-mysql-operator-6dcd78d857 SuccessfulCreate Created pod: percona-server-mysql-operator-6dcd78d857-hfqqc logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:57:01 +0000 UTC Normal Deployment.apps percona-server-mysql-operator ScalingReplicaSet Scaled up replica set percona-server-mysql-operator-6dcd78d857 to 1 logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:57:04 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-hfqqc.spec.containers{manager} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:57:04 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-hfqqc.spec.containers{manager} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 424.815889ms logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:57:04 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-hfqqc.spec.containers{manager} Created Created container manager logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:57:05 +0000 UTC Normal Lease.coordination.k8s.io 08db2feb.percona.com LeaderElection percona-server-mysql-operator-6dcd78d857-hfqqc_f8d9244a-2bb5-4972-81dd-f8614f46adfb became leader logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:57:05 +0000 UTC Normal Pod mysql-client Scheduled Successfully assigned kuttl-test-alive-corgi/mysql-client to gke-jen-ps-339-3a8483b-1-default-pool-7f97af72-klx5 logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:57:05 +0000 UTC Normal Pod percona-server-mysql-operator-6dcd78d857-hfqqc.spec.containers{manager} Started Started container manager logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:57:06 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Pulled Container image "percona/percona-server:8.0.25" already present on machine logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:57:06 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Created Created container mysql-client logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:57:06 +0000 UTC Normal Pod mysql-client.spec.containers{mysql-client} Started Started container mysql-client logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:40 +0000 UTC Normal Pod chaos-daemon-6gq5q Scheduled Successfully assigned kuttl-test-alive-corgi/chaos-daemon-6gq5q to gke-jen-ps-339-3a8483b-1-default-pool-7f97af72-xrzt logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:40 +0000 UTC Normal Pod chaos-daemon-tkbl6 Scheduled Successfully assigned kuttl-test-alive-corgi/chaos-daemon-tkbl6 to gke-jen-ps-339-3a8483b-1-default-pool-7f97af72-zbqv logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:40 +0000 UTC Normal DaemonSet.apps chaos-daemon SuccessfulCreate Created pod: chaos-daemon-6gq5q logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:40 +0000 UTC Normal DaemonSet.apps chaos-daemon SuccessfulCreate Created pod: chaos-daemon-tkbl6 logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:40 +0000 UTC Normal DaemonSet.apps chaos-daemon SuccessfulCreate Created pod: chaos-daemon-rzlwh logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:41 +0000 UTC Normal Pod chaos-controller-manager-f5d59d778-6xlm5 Scheduled Successfully assigned kuttl-test-alive-corgi/chaos-controller-manager-f5d59d778-6xlm5 to gke-jen-ps-339-3a8483b-1-default-pool-7f97af72-zbqv logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:41 +0000 UTC Normal Pod chaos-controller-manager-f5d59d778-97v9m Scheduled Successfully assigned kuttl-test-alive-corgi/chaos-controller-manager-f5d59d778-97v9m to gke-jen-ps-339-3a8483b-1-default-pool-7f97af72-xrzt logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:41 +0000 UTC Normal Pod chaos-controller-manager-f5d59d778-qwpnr Scheduled Successfully assigned kuttl-test-alive-corgi/chaos-controller-manager-f5d59d778-qwpnr to gke-jen-ps-339-3a8483b-1-default-pool-7f97af72-klx5 logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:41 +0000 UTC Normal ReplicaSet.apps chaos-controller-manager-f5d59d778 SuccessfulCreate Created pod: chaos-controller-manager-f5d59d778-97v9m logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:41 +0000 UTC Normal ReplicaSet.apps chaos-controller-manager-f5d59d778 SuccessfulCreate Created pod: chaos-controller-manager-f5d59d778-6xlm5 logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:41 +0000 UTC Normal ReplicaSet.apps chaos-controller-manager-f5d59d778 SuccessfulCreate Created pod: chaos-controller-manager-f5d59d778-qwpnr logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:41 +0000 UTC Normal Deployment.apps chaos-controller-manager ScalingReplicaSet Scaled up replica set chaos-controller-manager-f5d59d778 to 3 logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:41 +0000 UTC Normal Pod chaos-daemon-6gq5q.spec.containers{chaos-daemon} Pulled Container image "ghcr.io/chaos-mesh/chaos-daemon:v2.5.1" already present on machine logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:41 +0000 UTC Normal Pod chaos-daemon-6gq5q.spec.containers{chaos-daemon} Created Created container chaos-daemon logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:41 +0000 UTC Normal Pod chaos-daemon-6gq5q.spec.containers{chaos-daemon} Started Started container chaos-daemon logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:41 +0000 UTC Normal Pod chaos-daemon-rzlwh Scheduled Successfully assigned kuttl-test-alive-corgi/chaos-daemon-rzlwh to gke-jen-ps-339-3a8483b-1-default-pool-7f97af72-klx5 logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:41 +0000 UTC Normal Pod chaos-daemon-tkbl6.spec.containers{chaos-daemon} Pulled Container image "ghcr.io/chaos-mesh/chaos-daemon:v2.5.1" already present on machine logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:41 +0000 UTC Normal Pod chaos-daemon-tkbl6.spec.containers{chaos-daemon} Created Created container chaos-daemon logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:41 +0000 UTC Normal Pod chaos-daemon-tkbl6.spec.containers{chaos-daemon} Started Started container chaos-daemon logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:42 +0000 UTC Warning Pod chaos-controller-manager-f5d59d778-6xlm5 FailedMount MountVolume.SetUp failed for volume "webhook-certs" : failed to sync secret cache: timed out waiting for the condition logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:42 +0000 UTC Warning Pod chaos-controller-manager-f5d59d778-6xlm5 FailedMount MountVolume.SetUp failed for volume "chaosd-client-cert" : failed to sync secret cache: timed out waiting for the condition logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:42 +0000 UTC Normal Pod chaos-controller-manager-f5d59d778-97v9m.spec.containers{chaos-mesh} Pulled Container image "ghcr.io/chaos-mesh/chaos-mesh:v2.5.1" already present on machine logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:42 +0000 UTC Normal Pod chaos-controller-manager-f5d59d778-97v9m.spec.containers{chaos-mesh} Created Created container chaos-mesh logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:42 +0000 UTC Normal Pod chaos-controller-manager-f5d59d778-97v9m.spec.containers{chaos-mesh} Started Started container chaos-mesh logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:42 +0000 UTC Normal Pod chaos-controller-manager-f5d59d778-qwpnr.spec.containers{chaos-mesh} Pulled Container image "ghcr.io/chaos-mesh/chaos-mesh:v2.5.1" already present on machine logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:42 +0000 UTC Normal Pod chaos-controller-manager-f5d59d778-qwpnr.spec.containers{chaos-mesh} Created Created container chaos-mesh logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:42 +0000 UTC Normal Pod chaos-controller-manager-f5d59d778-qwpnr.spec.containers{chaos-mesh} Started Started container chaos-mesh logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:42 +0000 UTC Normal Pod chaos-daemon-rzlwh.spec.containers{chaos-daemon} Pulled Container image "ghcr.io/chaos-mesh/chaos-daemon:v2.5.1" already present on machine logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:42 +0000 UTC Normal Pod chaos-daemon-rzlwh.spec.containers{chaos-daemon} Created Created container chaos-daemon logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:42 +0000 UTC Normal Pod chaos-daemon-rzlwh.spec.containers{chaos-daemon} Started Started container chaos-daemon logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:42 +0000 UTC Normal ConfigMap chaos-mesh LeaderElection chaos-controller-manager-f5d59d778-qwpnr_f3368a38-8213-4562-b3ca-578167c52818 became leader logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:42 +0000 UTC Normal Lease.coordination.k8s.io chaos-mesh LeaderElection chaos-controller-manager-f5d59d778-qwpnr_f3368a38-8213-4562-b3ca-578167c52818 became leader logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:43 +0000 UTC Normal Pod chaos-controller-manager-f5d59d778-6xlm5.spec.containers{chaos-mesh} Pulled Container image "ghcr.io/chaos-mesh/chaos-mesh:v2.5.1" already present on machine logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:43 +0000 UTC Normal Pod chaos-controller-manager-f5d59d778-6xlm5.spec.containers{chaos-mesh} Created Created container chaos-mesh logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:43 +0000 UTC Normal Pod chaos-controller-manager-f5d59d778-6xlm5.spec.containers{chaos-mesh} Started Started container chaos-mesh logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:57 +0000 UTC Normal PersistentVolumeClaim datadir-gr-self-healing-chaos-mysql-0 WaitForFirstConsumer waiting for first consumer to be created before binding logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:57 +0000 UTC Normal PersistentVolumeClaim datadir-gr-self-healing-chaos-mysql-0 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:57 +0000 UTC Normal PersistentVolumeClaim datadir-gr-self-healing-chaos-mysql-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-alive-corgi/datadir-gr-self-healing-chaos-mysql-0" logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:57 +0000 UTC Normal StatefulSet.apps gr-self-healing-chaos-mysql SuccessfulCreate create Claim datadir-gr-self-healing-chaos-mysql-0 Pod gr-self-healing-chaos-mysql-0 in StatefulSet gr-self-healing-chaos-mysql success logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 12:59:57 +0000 UTC Normal StatefulSet.apps gr-self-healing-chaos-mysql SuccessfulCreate create Pod gr-self-healing-chaos-mysql-0 in StatefulSet gr-self-healing-chaos-mysql successful logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:01 +0000 UTC Normal PersistentVolumeClaim datadir-gr-self-healing-chaos-mysql-0 ProvisioningSucceeded Successfully provisioned volume pvc-8dc1b235-25b4-4d48-b66f-d1851945720a logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:02 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-0 Scheduled Successfully assigned kuttl-test-alive-corgi/gr-self-healing-chaos-mysql-0 to gke-jen-ps-339-3a8483b-1-default-pool-7f97af72-zbqv logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:07 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-8dc1b235-25b4-4d48-b66f-d1851945720a" logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:10 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:11 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 1.207458899s logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:11 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-0.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:11 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:13 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:16 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 3.289331396s logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:16 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-0.spec.containers{mysql} Created Created container mysql logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:16 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-0.spec.containers{mysql} Started Started container mysql logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:16 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:16 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 446.420004ms logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:16 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-0.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:17 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:49 +0000 UTC Normal PersistentVolumeClaim datadir-gr-self-healing-chaos-mysql-1 WaitForFirstConsumer waiting for first consumer to be created before binding logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:49 +0000 UTC Normal StatefulSet.apps gr-self-healing-chaos-mysql SuccessfulCreate create Claim datadir-gr-self-healing-chaos-mysql-1 Pod gr-self-healing-chaos-mysql-1 in StatefulSet gr-self-healing-chaos-mysql success logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:49 +0000 UTC Normal StatefulSet.apps gr-self-healing-chaos-mysql SuccessfulCreate create Pod gr-self-healing-chaos-mysql-1 in StatefulSet gr-self-healing-chaos-mysql successful logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:50 +0000 UTC Normal PersistentVolumeClaim datadir-gr-self-healing-chaos-mysql-1 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:50 +0000 UTC Normal PersistentVolumeClaim datadir-gr-self-healing-chaos-mysql-1 Provisioning External provisioner is provisioning volume for claim "kuttl-test-alive-corgi/datadir-gr-self-healing-chaos-mysql-1" logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:50 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-0 Scheduled Successfully assigned kuttl-test-alive-corgi/gr-self-healing-chaos-haproxy-0 to gke-jen-ps-339-3a8483b-1-default-pool-7f97af72-klx5 logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:50 +0000 UTC Normal StatefulSet.apps gr-self-healing-chaos-haproxy SuccessfulCreate create Pod gr-self-healing-chaos-haproxy-0 in StatefulSet gr-self-healing-chaos-haproxy successful logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:51 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-0.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:52 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-0.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 431.262688ms logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:52 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-0.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:52 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-0.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:53 +0000 UTC Normal PersistentVolumeClaim datadir-gr-self-healing-chaos-mysql-1 ProvisioningSucceeded Successfully provisioned volume pvc-9ff5d9d2-a538-4c42-963c-6a9689bd4105 logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:54 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-0.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:54 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-1 Scheduled Successfully assigned kuttl-test-alive-corgi/gr-self-healing-chaos-mysql-1 to gke-jen-ps-339-3a8483b-1-default-pool-7f97af72-xrzt logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:55 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-0.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 375.954345ms logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:55 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-0.spec.containers{haproxy} Created Created container haproxy logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:55 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-0.spec.containers{haproxy} Started Started container haproxy logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:55 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:55 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 388.983606ms logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:55 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-0.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:55 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-0.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:55 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-1 Scheduled Successfully assigned kuttl-test-alive-corgi/gr-self-healing-chaos-haproxy-1 to gke-jen-ps-339-3a8483b-1-default-pool-7f97af72-xrzt logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:55 +0000 UTC Normal StatefulSet.apps gr-self-healing-chaos-haproxy SuccessfulCreate create Pod gr-self-healing-chaos-haproxy-1 in StatefulSet gr-self-healing-chaos-haproxy successful logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:56 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-1.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:57 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-1.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 383.009958ms logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:57 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-1.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:57 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-1.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:59 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-1.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:59 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-1.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 374.987605ms logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:59 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-1.spec.containers{haproxy} Created Created container haproxy logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:59 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-1.spec.containers{haproxy} Started Started container haproxy logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:00:59 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-1.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:00 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-1.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 376.489162ms logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:00 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-1.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:00 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-1.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:00 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-2 Scheduled Successfully assigned kuttl-test-alive-corgi/gr-self-healing-chaos-haproxy-2 to gke-jen-ps-339-3a8483b-1-default-pool-7f97af72-zbqv logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:00 +0000 UTC Normal StatefulSet.apps gr-self-healing-chaos-haproxy SuccessfulCreate create Pod gr-self-healing-chaos-haproxy-2 in StatefulSet gr-self-healing-chaos-haproxy successful logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:01 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-2.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:01 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-2.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 422.628718ms logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:01 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-2.spec.initContainers{haproxy-init} Created Created container haproxy-init logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:01 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-2.spec.initContainers{haproxy-init} Started Started container haproxy-init logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:01 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-1 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-9ff5d9d2-a538-4c42-963c-6a9689bd4105" logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:02 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-1.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:03 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-2.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:03 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-2.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 372.344051ms logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:03 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-2.spec.containers{haproxy} Created Created container haproxy logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:03 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-2.spec.containers{haproxy} Started Started container haproxy logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:03 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-2.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:03 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-1.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 370.11823ms logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:03 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-1.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:03 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-1.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:04 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-2.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 352.694011ms logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:04 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-2.spec.containers{mysql-monit} Created Created container mysql-monit logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:04 +0000 UTC Normal Pod gr-self-healing-chaos-haproxy-2.spec.containers{mysql-monit} Started Started container mysql-monit logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:05 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-1.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:05 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 402.182418ms logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:05 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-1.spec.containers{mysql} Created Created container mysql logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:05 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-1.spec.containers{mysql} Started Started container mysql logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:05 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-1.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:06 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-1.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 455.657464ms logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:06 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-1.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:06 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-1.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:35 +0000 UTC Warning Pod gr-self-healing-chaos-mysql-1.spec.containers{mysql} Unhealthy Startup probe failed: logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:35 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-1.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:01:36 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-1.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 389.796448ms logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:02:07 +0000 UTC Normal PersistentVolumeClaim datadir-gr-self-healing-chaos-mysql-2 WaitForFirstConsumer waiting for first consumer to be created before binding logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:02:07 +0000 UTC Normal PersistentVolumeClaim datadir-gr-self-healing-chaos-mysql-2 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:02:07 +0000 UTC Normal PersistentVolumeClaim datadir-gr-self-healing-chaos-mysql-2 Provisioning External provisioner is provisioning volume for claim "kuttl-test-alive-corgi/datadir-gr-self-healing-chaos-mysql-2" logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:02:07 +0000 UTC Normal StatefulSet.apps gr-self-healing-chaos-mysql SuccessfulCreate create Claim datadir-gr-self-healing-chaos-mysql-2 Pod gr-self-healing-chaos-mysql-2 in StatefulSet gr-self-healing-chaos-mysql success logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:02:07 +0000 UTC Normal StatefulSet.apps gr-self-healing-chaos-mysql SuccessfulCreate create Pod gr-self-healing-chaos-mysql-2 in StatefulSet gr-self-healing-chaos-mysql successful logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:02:11 +0000 UTC Normal PersistentVolumeClaim datadir-gr-self-healing-chaos-mysql-2 ProvisioningSucceeded Successfully provisioned volume pvc-ddbbc6ef-785f-4d67-a3bb-e0931cfc85d6 logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:02:12 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-2 Scheduled Successfully assigned kuttl-test-alive-corgi/gr-self-healing-chaos-mysql-2 to gke-jen-ps-339-3a8483b-1-default-pool-7f97af72-klx5 logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:02:17 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-2 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-ddbbc6ef-785f-4d67-a3bb-e0931cfc85d6" logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:02:19 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-2.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:02:20 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-2.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-339-3a8483b" in 470.875226ms logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:02:20 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-2.spec.initContainers{mysql-init} Created Created container mysql-init logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:02:20 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-2.spec.initContainers{mysql-init} Started Started container mysql-init logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:02:21 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-2.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql" logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:02:22 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 422.177678ms logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:02:22 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-2.spec.containers{mysql} Created Created container mysql logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:02:22 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-2.spec.containers{mysql} Started Started container mysql logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:02:22 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-2.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup" logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:02:22 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-2.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup" in 349.748932ms logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:02:22 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-2.spec.containers{xtrabackup} Created Created container xtrabackup logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:02:22 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-2.spec.containers{xtrabackup} Started Started container xtrabackup logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:02:53 +0000 UTC Warning Pod gr-self-healing-chaos-mysql-2.spec.containers{mysql} Unhealthy Startup probe failed: logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:02:53 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-2.spec.containers{mysql} Killing Container mysql failed startup probe, will be restarted logger.go:42: 13:06:57 | gr-self-healing-chaos | 2023-08-09 13:02:53 +0000 UTC Normal Pod gr-self-healing-chaos-mysql-2.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql" in 400.140516ms logger.go:42: 13:06:57 | gr-self-healing-chaos | Deleting namespace: kuttl-test-alive-corgi case.go:114: context deadline exceeded === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (793.84s) --- FAIL: kuttl/harness (0.00s) --- FAIL: kuttl/harness/gr-self-healing-chaos (790.30s) FAIL