++ echo 'Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/logs/operator-self-healing-chaos.log' Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/logs/operator-self-healing-chaos.log ++ '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/cloud-secret.yml ']' ++ SKIP_BACKUPS_TO_AWS_GCP= ++ oc get projects error: No Auth Provider found for name "gcp" +++ kubectl version -o json +++ grep '\-eks\-' +++ jq -r .serverVersion.gitVersion ++ '[' ']' ++ EKS=0 +++ kubectl version -o json +++ jq -r '.serverVersion.major + "." + .serverVersion.minor' +++ /usr/bin/sed -r 's/[^0-9.]+//g' ++ KUBE_VERSION=1.20 +++ helm version -c +++ /usr/bin/sed -re 's/.*SemVer:"([^"]+)".*/\1/; s/.*\bVersion:"([^"]+)".*/\1/' ++ HELM_VERSION=v3.8.2 ++ '[' v3 == v2 ']' + cluster=operator-chaos + main + create_infra operator-self-healing-chaos-21589 + local ns=operator-self-healing-chaos-21589 + '[' -n pxc-operator ']' + kubectl get pxc --all-namespaces -o wide + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + grep -v NAMESPACE No resources found + kubectl patch pxc -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: resource(s) were provided, but no name, label selector, or --all flag specified + : + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.QjSVaIk3hu ++ mktemp + local LAST_ERR=/tmp/tmp.rbtXPGcXij + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl delete pxc --all --all-namespaces + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.QjSVaIk3hu No resources found + cat /tmp/tmp.rbtXPGcXij + rm /tmp/tmp.QjSVaIk3hu /tmp/tmp.rbtXPGcXij + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.eQH9WwFkYN ++ mktemp + local LAST_ERR=/tmp/tmp.ooEtl60G80 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.eQH9WwFkYN No resources found + cat /tmp/tmp.ooEtl60G80 + rm /tmp/tmp.eQH9WwFkYN /tmp/tmp.ooEtl60G80 + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.6zpCIHKqjh ++ mktemp + local LAST_ERR=/tmp/tmp.ZuV3JPmESg + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.6zpCIHKqjh No resources found + cat /tmp/tmp.ZuV3JPmESg + rm /tmp/tmp.6zpCIHKqjh /tmp/tmp.ZuV3JPmESg + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + kubectl_bin get ns + '[' '!' -z '' ']' + kubectl_bin delete namespace pxc-operator + egrep -v '^kube-|^default|Terminating|pxc-operator|openshift|^NAME' ++ mktemp + xargs kubectl delete ns ++ mktemp + awk '{print$1}' + local LAST_OUT=/tmp/tmp.SBZyGeBYbo ++ mktemp + local LAST_OUT=/tmp/tmp.CeMXKuD7sJ + local LAST_ERR=/tmp/tmp.nyIaxxHWw0 + local exit_status=0 ++ mktemp ++ seq 0 2 + local LAST_ERR=/tmp/tmp.YyOBn7pEHM + local exit_status=0 + for i in '$(seq 0 2)' + kubectl get ns ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl delete namespace pxc-operator + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.SBZyGeBYbo + cat /tmp/tmp.nyIaxxHWw0 + rm /tmp/tmp.SBZyGeBYbo /tmp/tmp.nyIaxxHWw0 + return 0 namespace "operator-self-healing-chaos-21885" deleted + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.CeMXKuD7sJ namespace "pxc-operator" deleted + cat /tmp/tmp.YyOBn7pEHM + rm /tmp/tmp.CeMXKuD7sJ /tmp/tmp.YyOBn7pEHM + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + set +o xtrace namespace/pxc-operator - Error from server (NotFound): namespaces "pxc-operator" not found + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.wm6uHl31XY ++ mktemp + local LAST_ERR=/tmp/tmp.AhyoC5TX6R + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl create namespace pxc-operator + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.wm6uHl31XY namespace/pxc-operator created + cat /tmp/tmp.AhyoC5TX6R + rm /tmp/tmp.wm6uHl31XY /tmp/tmp.AhyoC5TX6R + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.h1hgKKMTYQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.7shm3Gseuo ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl config current-context ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.h1hgKKMTYQ ++ cat /tmp/tmp.7shm3Gseuo ++ rm /tmp/tmp.h1hgKKMTYQ /tmp/tmp.7shm3Gseuo ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jenkins-pxc-706f792a-selfhealing --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.vv7CXqHkLR ++ mktemp + local LAST_ERR=/tmp/tmp.uco2ozlzbZ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jenkins-pxc-706f792a-selfhealing --namespace=pxc-operator + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.vv7CXqHkLR Context "gke_cloud-dev-112233_us-central1-a_jenkins-pxc-706f792a-selfhealing" modified. + cat /tmp/tmp.uco2ozlzbZ + rm /tmp/tmp.vv7CXqHkLR /tmp/tmp.uco2ozlzbZ + return 0 + deploy_operator + desc 'start operator' + set +o xtrace ----------------------------------------------------------------------------------- start operator ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.E0Ta905HR2 ++ mktemp + local LAST_ERR=/tmp/tmp.IlxBWW2vlb + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/deploy/crd.yaml + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.E0Ta905HR2 customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com unchanged customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com unchanged customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com unchanged customresourcedefinition.apiextensions.k8s.io/perconaxtradbbackups.pxc.percona.com configured + cat /tmp/tmp.IlxBWW2vlb + rm /tmp/tmp.E0Ta905HR2 /tmp/tmp.IlxBWW2vlb + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/deploy/cw-rbac.yaml + kubectl_bin apply -f - + sed -e 's^namespace: .*^namespace: pxc-operator^' ++ mktemp + local LAST_OUT=/tmp/tmp.ZU2GhKiQVt ++ mktemp + local LAST_ERR=/tmp/tmp.9ZjHNA9XLR + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.ZU2GhKiQVt clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator configured serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.9ZjHNA9XLR + rm /tmp/tmp.ZU2GhKiQVt /tmp/tmp.9ZjHNA9XLR + return 0 + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/deploy/cw-operator.yaml + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-1125-706f792a^' + sed -e 's^failureThreshold: .*^failureThreshold: 10^' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.3thzACgS2i ++ mktemp + local LAST_ERR=/tmp/tmp.JeSqmCjOkI + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.3thzACgS2i deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.JeSqmCjOkI + rm /tmp/tmp.3thzACgS2i /tmp/tmp.JeSqmCjOkI + return 0 + sleep 10 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.SK6LX4BOcA +++ mktemp ++ local LAST_ERR=/tmp/tmp.H40dEH9aiL ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.SK6LX4BOcA ++ cat /tmp/tmp.H40dEH9aiL ++ rm /tmp/tmp.SK6LX4BOcA /tmp/tmp.H40dEH9aiL ++ return 0 + wait_pod percona-xtradb-cluster-operator-5699d7755d-dgnwl 480 pxc-operator + local pod=percona-xtradb-cluster-operator-5699d7755d-dgnwl + local max_retry=480 + local ns=pxc-operator ++ echo percona-xtradb-cluster-operator-5699d7755d-dgnwl ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace percona-xtradb-cluster-operator-5699d7755d-dgnwl.Ok + sleep 3 + create_namespace operator-self-healing-chaos-21589 + local namespace=operator-self-healing-chaos-21589 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + egrep -v '^kube-|^default|Terminating|pxc-operator|openshift|^NAME' + '[' '!' -z '' ']' + kubectl_bin delete namespace operator-self-healing-chaos-21589 + kubectl_bin get ns ++ mktemp + local LAST_OUT=/tmp/tmp.NLTAES3SS8 + awk '{print$1}' + xargs kubectl delete ns ++ mktemp + local LAST_OUT=/tmp/tmp.v86QzQlNZS ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.BVJ71ydXkz + local exit_status=0 + local LAST_ERR=/tmp/tmp.cnuLaF9H0j + local exit_status=0 ++ seq 0 2 ++ seq 0 2 + for i in '$(seq 0 2)' + for i in '$(seq 0 2)' + kubectl get ns + kubectl delete namespace operator-self-healing-chaos-21589 + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + kubectl delete namespace operator-self-healing-chaos-21589 + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.NLTAES3SS8 + cat /tmp/tmp.BVJ71ydXkz + rm /tmp/tmp.NLTAES3SS8 /tmp/tmp.BVJ71ydXkz + return 0 error: resource(s) were provided, but no name, label selector, or --all flag specified + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + kubectl delete namespace operator-self-healing-chaos-21589 + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + cat /tmp/tmp.v86QzQlNZS + cat /tmp/tmp.cnuLaF9H0j Error from server (NotFound): namespaces "operator-self-healing-chaos-21589" not found + rm /tmp/tmp.v86QzQlNZS /tmp/tmp.cnuLaF9H0j + return 1 + : + wait_for_delete namespace/operator-self-healing-chaos-21589 + local res=namespace/operator-self-healing-chaos-21589 + set +o xtrace namespace/operator-self-healing-chaos-21589 - Error from server (NotFound): namespaces "operator-self-healing-chaos-21589" not found + kubectl_bin create namespace operator-self-healing-chaos-21589 ++ mktemp + local LAST_OUT=/tmp/tmp.RJI2MJu0D5 ++ mktemp + local LAST_ERR=/tmp/tmp.U4mM2XxKhP + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl create namespace operator-self-healing-chaos-21589 + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.RJI2MJu0D5 namespace/operator-self-healing-chaos-21589 created + cat /tmp/tmp.U4mM2XxKhP + rm /tmp/tmp.RJI2MJu0D5 /tmp/tmp.U4mM2XxKhP + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.AWlykfJdcG +++ mktemp ++ local LAST_ERR=/tmp/tmp.6fNpgXcZhM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl config current-context ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.AWlykfJdcG ++ cat /tmp/tmp.6fNpgXcZhM ++ rm /tmp/tmp.AWlykfJdcG /tmp/tmp.6fNpgXcZhM ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jenkins-pxc-706f792a-selfhealing --namespace=operator-self-healing-chaos-21589 ++ mktemp + local LAST_OUT=/tmp/tmp.4QX7wD0105 ++ mktemp + local LAST_ERR=/tmp/tmp.moguAYdIzp + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jenkins-pxc-706f792a-selfhealing --namespace=operator-self-healing-chaos-21589 + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.4QX7wD0105 Context "gke_cloud-dev-112233_us-central1-a_jenkins-pxc-706f792a-selfhealing" modified. + cat /tmp/tmp.moguAYdIzp + rm /tmp/tmp.4QX7wD0105 /tmp/tmp.moguAYdIzp + return 0 + apply_secrets + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.iEuLtgE9Mx ++ mktemp + local LAST_ERR=/tmp/tmp.FApPVye1ys + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/cloud-secret.yml + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.iEuLtgE9Mx secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.FApPVye1ys + rm /tmp/tmp.iEuLtgE9Mx /tmp/tmp.FApPVye1ys + return 0 + test_namespace=operator-self-healing-chaos-21589 + '[' -n pxc-operator ']' + kubectl_bin patch clusterrole percona-xtradb-cluster-operator --type=json -p '[{"op":"remove","path":"/rules/1"}]' ++ mktemp + local LAST_OUT=/tmp/tmp.PulXARsSPX ++ mktemp + local LAST_ERR=/tmp/tmp.zmPnGaGFmL + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl patch clusterrole percona-xtradb-cluster-operator --type=json -p '[{"op":"remove","path":"/rules/1"}]' + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.PulXARsSPX clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator patched + cat /tmp/tmp.zmPnGaGFmL + rm /tmp/tmp.PulXARsSPX /tmp/tmp.zmPnGaGFmL + return 0 + kubectl_bin delete validatingwebhookconfigurations.admissionregistration.k8s.io percona-xtradbcluster-webhook ++ mktemp + local LAST_OUT=/tmp/tmp.OAo3HUfvz4 ++ mktemp + local LAST_ERR=/tmp/tmp.4PIu7hFR3N + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl delete validatingwebhookconfigurations.admissionregistration.k8s.io percona-xtradbcluster-webhook + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.OAo3HUfvz4 validatingwebhookconfiguration.admissionregistration.k8s.io "percona-xtradbcluster-webhook" deleted + cat /tmp/tmp.4PIu7hFR3N + rm /tmp/tmp.OAo3HUfvz4 /tmp/tmp.4PIu7hFR3N + return 0 + test_namespace=pxc-operator + deploy_chaos_mesh pxc-operator + local chaos_mesh_ns=pxc-operator + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + timeout 30 kubectl delete podchaos --all --all-namespaces podchaos.chaos-mesh.org "chaos-pod-failure" deleted podchaos.chaos-mesh.org "chaos-pod-kill" deleted + : + timeout 30 kubectl delete networkchaos --all --all-namespaces networkchaos.chaos-mesh.org "chaos-pod-network-loss" deleted + : + '[' -n '' ']' + timeout 30 kubectl delete crd awschaos.chaos-mesh.org dnschaos.chaos-mesh.org gcpchaos.chaos-mesh.org httpchaos.chaos-mesh.org iochaos.chaos-mesh.org jvmchaos.chaos-mesh.org kernelchaos.chaos-mesh.org networkchaos.chaos-mesh.org podchaos.chaos-mesh.org podhttpchaos.chaos-mesh.org podiochaos.chaos-mesh.org podnetworkchaos.chaos-mesh.org schedules.chaos-mesh.org stresschaos.chaos-mesh.org timechaos.chaos-mesh.org workflownodes.chaos-mesh.org workflows.chaos-mesh.org customresourcedefinition.apiextensions.k8s.io "awschaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "dnschaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "gcpchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "httpchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "iochaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "jvmchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "kernelchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "networkchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "podchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "podhttpchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "podiochaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "podnetworkchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "schedules.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "stresschaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "timechaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "workflownodes.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "workflows.chaos-mesh.org" deleted + : + timeout 30 kubectl delete clusterrolebinding chaos-mesh-chaos-controller-manager-cluster-level Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "chaos-mesh-chaos-controller-manager-cluster-level" not found + : + timeout 30 kubectl delete clusterrole chaos-mesh-chaos-controller-manager-cluster-level chaos-mesh-chaos-controller-manager-target-namespace Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "chaos-mesh-chaos-controller-manager-cluster-level" not found Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "chaos-mesh-chaos-controller-manager-target-namespace" not found + : + desc 'install chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- install chaos-mesh ----------------------------------------------------------------------------------- + helm repo add chaos-mesh https://charts.chaos-mesh.org "chaos-mesh" already exists with the same configuration, skipping + version_gt 1.19 ++ echo '1.20 >= 1.19' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + helm install chaos-mesh chaos-mesh/chaos-mesh --namespace=pxc-operator --set chaosDaemon.runtime=containerd --set chaosDaemon.socketPath=/run/containerd/containerd.sock --set dashboard.create=false --version 2.0.4 --set clusterScoped=false --set controllerManager.targetNamespace=pxc-operator Error: INSTALLATION FAILED: rendered manifests contain a resource that already exists. Unable to continue with install: MutatingWebhookConfiguration "chaos-mesh-mutation" in namespace "" exists and cannot be imported into the current release: invalid ownership metadata; annotation validation error: key "meta.helm.sh/release-namespace" must equal "pxc-operator": current value is "self-healing-chaos-27971"