Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2476/e2e-tests/logs/haproxy-5-7.log Warning: version difference between client (1.36) and server (1.33) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.36) and server (1.33) exceeds the supported minor version skew of +/-1 No resources found + kubectl patch pxc -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: resource(s) were provided, but no name was specified No resources found No resources found No resources found error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- error: resource(s) were provided, but no name was specified namespace "pxc-operator" deleted waiting for namespace/pxc-operator to be deletedError from server (NotFound): namespaces "pxc-operator" not found ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- namespace/pxc-operator created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2476-a8b01a39-5-cluster6" modified. ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created pod/percona-xtradb-cluster-operator-8548fd5788-lbk2b condition met E0516 19:09:42.877649 23209 reflector.go:227] "Failed to watch" err="Get \"https://136.116.224.45/api/v1/namespaces/pxc-operator/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpercona-xtradb-cluster-operator-8548fd5788-lbk2b&resourceVersion=1778958582428853000&timeoutSeconds=474&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" pod/percona-xtradb-cluster-operator-8548fd5788-lbk2b condition met E0516 19:09:47.851802 24212 reflector.go:227] "Failed to watch" err="Get \"https://136.116.224.45/api/v1/namespaces/pxc-operator/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpercona-xtradb-cluster-operator-8548fd5788-lbk2b&resourceVersion=1778958586843915000&timeoutSeconds=593&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/percona-xtradb-cluster-operator-8548fd5788-lbk2b to become Ready.Ok error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces haproxy-16121 ----------------------------------------------------------------------------------- Error from server (NotFound): namespaces "haproxy-16121" not found waiting for namespace/haproxy-16121 to be deletederror: resource(s) were provided, but no name was specified Error from server (NotFound): namespaces "haproxy-16121" not found ----------------------------------------------------------------------------------- create namespace haproxy-16121 ----------------------------------------------------------------------------------- namespace/haproxy-16121 created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2476-a8b01a39-5-cluster6" modified. ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- secret/minio-secret created secret/aws-s3-secret created secret/do-spaces-secret created secret/gcp-cs-secret created secret/azure-secret created error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- install chaos-mesh ----------------------------------------------------------------------------------- "chaos-mesh" has been added to your repositories NAME: chaos-mesh LAST DEPLOYED: Sat May 16 19:11:04 2026 NAMESPACE: haproxy-16121 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: 1. Make sure chaos-mesh components are running kubectl get pods --namespace haproxy-16121 -l app.kubernetes.io/instance=chaos-mesh Waiting for DaemonSet chaos-daemon... runtimeclass.node.k8s.io/docker-rc created ----------------------------------------------------------------------------------- create first PXC cluster with HAProxy ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created deployment.apps/pxc-client created perconaxtradbcluster.pxc.percona.com/haproxy created ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- error: no matching resources found ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/haproxy-haproxy-0 condition met waiting for pod/haproxy-haproxy-0 to become Ready.Ok ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/haproxy-pxc-0 condition met E0516 19:13:20.755521 22148 reflector.go:227] "Failed to watch" err="Get \"https://136.116.224.45/api/v1/namespaces/haproxy-16121/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-pxc-0&resourceVersion=1778958796575551003&timeoutSeconds=589&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-pxc-0 to become Ready.Ok pod/haproxy-pxc-1 condition met waiting for pod/haproxy-pxc-1 to become Ready.Ok pod/haproxy-pxc-2 condition met waiting for pod/haproxy-pxc-2 to become Ready.Ok ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- pod/pxc-client-97cb9c68b-c55pg condition met E0516 19:16:11.651776 15428 reflector.go:227] "Failed to watch" err="Get \"https://136.116.224.45/api/v1/namespaces/haproxy-16121/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-c55pg&resourceVersion=1778958969724673000&timeoutSeconds=517&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-c55pg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-c55pg condition met E0516 19:16:28.976633 18020 reflector.go:227] "Failed to watch" err="Get \"https://136.116.224.45/api/v1/namespaces/haproxy-16121/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-c55pg&resourceVersion=1778958985433007000&timeoutSeconds=472&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-c55pg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-c55pg condition met E0516 19:17:12.856408 24532 reflector.go:227] "Failed to watch" err="Get \"https://136.116.224.45/api/v1/namespaces/haproxy-16121/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-c55pg&resourceVersion=1778959030501350000&timeoutSeconds=306&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-c55pg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-c55pg condition met E0516 19:17:27.231921 26486 reflector.go:227] "Failed to watch" err="Get \"https://136.116.224.45/api/v1/namespaces/haproxy-16121/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-c55pg&resourceVersion=1778959044467822000&timeoutSeconds=561&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-c55pg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-c55pg condition met E0516 19:17:39.170399 28455 reflector.go:227] "Failed to watch" err="Get \"https://136.116.224.45/api/v1/namespaces/haproxy-16121/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-c55pg&resourceVersion=1778959058104823000&timeoutSeconds=460&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-c55pg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- checking all haproxy pods point to the same writer ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/haproxy-pxc-0 condition met E0516 19:17:55.292856 30420 reflector.go:227] "Failed to watch" err="Get \"https://136.116.224.45/api/v1/namespaces/haproxy-16121/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-pxc-0&resourceVersion=1778959073105477000&timeoutSeconds=379&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-pxc-0 to become Ready.Ok pod/haproxy-pxc-1 condition met E0516 19:18:05.523431 31876 reflector.go:227] "Failed to watch" err="Get \"https://136.116.224.45/api/v1/namespaces/haproxy-16121/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-pxc-1&resourceVersion=1778959083105503000&timeoutSeconds=461&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-pxc-1 to become Ready.Ok pod/haproxy-pxc-2 condition met E0516 19:18:15.099580 727 reflector.go:227] "Failed to watch" err="Get \"https://136.116.224.45/api/v1/namespaces/haproxy-16121/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-pxc-2&resourceVersion=1778959093106348000&timeoutSeconds=379&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-pxc-2 to become Ready.Ok ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- waiting for pxc/haproxy to be ready pod/pxc-client-97cb9c68b-c55pg condition met E0516 19:18:49.597261 5602 reflector.go:227] "Failed to watch" err="Get \"https://136.116.224.45/api/v1/namespaces/haproxy-16121/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-c55pg&resourceVersion=1778959125635837000&timeoutSeconds=357&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-c55pg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-c55pg condition met E0516 19:19:11.141237 8704 reflector.go:227] "Failed to watch" err="Get \"https://136.116.224.45/api/v1/namespaces/haproxy-16121/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-c55pg&resourceVersion=1778959147637300000&timeoutSeconds=393&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-c55pg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-c55pg condition met E0516 19:19:27.608057 11087 reflector.go:227] "Failed to watch" err="Get \"https://136.116.224.45/api/v1/namespaces/haproxy-16121/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-c55pg&resourceVersion=1778959164171084000&timeoutSeconds=348&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-c55pg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- delete active writer and checking all haproxy pods still point to the same writer ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- fail pxc-pod-0 pod for 60s ----------------------------------------------------------------------------------- podchaos.chaos-mesh.org/chaos-pod-failure created NAME READY STATUS RESTARTS AGE chaos-controller-manager-7d99f769dc-b66tj 1/1 Running 0 8m47s chaos-controller-manager-7d99f769dc-rr6z5 1/1 Running 0 8m47s chaos-controller-manager-7d99f769dc-xttcp 1/1 Running 0 8m47s chaos-daemon-5ptvz 1/1 Running 0 8m47s chaos-daemon-rrhsc 1/1 Running 0 8m47s chaos-daemon-v489z 1/1 Running 0 8m47s chaos-dns-server-6648bb7956-srjjb 1/1 Running 0 8m47s haproxy-haproxy-0 3/3 Running 0 8m16s haproxy-haproxy-1 3/3 Running 0 6m51s haproxy-haproxy-2 3/3 Running 0 6m30s haproxy-pxc-0 0/1 CrashLoopBackOff 2 (8s ago) 8m16s haproxy-pxc-1 1/1 Running 0 6m54s haproxy-pxc-2 1/1 Running 0 5m38s pxc-client-97cb9c68b-c55pg 2/2 Running 0 8m21s pod/pxc-client-97cb9c68b-c55pg condition met E0516 19:20:07.087744 16729 reflector.go:227] "Failed to watch" err="Get \"https://136.116.224.45/api/v1/namespaces/haproxy-16121/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-c55pg&resourceVersion=1778959205306037000&timeoutSeconds=433&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-c55pg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-c55pg condition met E0516 19:20:22.640842 18918 reflector.go:227] "Failed to watch" err="Get \"https://136.116.224.45/api/v1/namespaces/haproxy-16121/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-c55pg&resourceVersion=1778959220074104000&timeoutSeconds=505&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-c55pg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-c55pg condition met E0516 19:20:39.828956 21417 reflector.go:227] "Failed to watch" err="Get \"https://136.116.224.45/api/v1/namespaces/haproxy-16121/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-c55pg&resourceVersion=1778959237430432000&timeoutSeconds=367&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-c55pg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/haproxy-pxc-0 condition met waiting for pod/haproxy-pxc-0 to become Ready.Ok pod/haproxy-pxc-1 condition met E0516 19:21:16.931622 26161 reflector.go:227] "Failed to watch" err="Get \"https://136.116.224.45/api/v1/namespaces/haproxy-16121/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-pxc-1&resourceVersion=1778959272432607021&timeoutSeconds=385&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-pxc-1 to become Ready.Ok pod/haproxy-pxc-2 condition met E0516 19:21:21.942639 26805 reflector.go:227] "Failed to watch" err="Get \"https://136.116.224.45/api/v1/namespaces/haproxy-16121/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-pxc-2&resourceVersion=1778959277435016000&timeoutSeconds=483&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-pxc-2 to become Ready.Ok ----------------------------------------------------------------------------------- check advanced options are enabled in haproxy statefulset ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- compare pdb/haproxy-haproxy- ----------------------------------------------------------------------------------- [2026-05-16T19:21:27+0000] compare_kubectl: pdb/haproxy-haproxy OK ----------------------------------------------------------------------------------- compare statefulset/haproxy-haproxy- ----------------------------------------------------------------------------------- [2026-05-16T19:21:29+0000] compare_kubectl: statefulset/haproxy-haproxy OK ----------------------------------------------------------------------------------- default haproxy-replicas service ----------------------------------------------------------------------------------- The request is invalid: the server rejected our request due to an error in our request ----------------------------------------------------------------------------------- compare service/haproxy-haproxy-replicas- ----------------------------------------------------------------------------------- [2026-05-16T19:21:33+0000] compare_kubectl: service/haproxy-haproxy-replicas OK ----------------------------------------------------------------------------------- disable haproxy-replicas service ----------------------------------------------------------------------------------- waiting for svc/haproxy-haproxy-replicas to be deletedError from server (NotFound): services "haproxy-haproxy-replicas" not found ----------------------------------------------------------------------------------- enable haproxy-replicas service ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- compare service/haproxy-haproxy-replicas- ----------------------------------------------------------------------------------- [2026-05-16T19:22:19+0000] compare_kubectl: service/haproxy-haproxy-replicas OK ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/haproxy-haproxy-0 condition met E0516 19:22:21.169717 2431 reflector.go:227] "Failed to watch" err="Get \"https://136.116.224.45/api/v1/namespaces/haproxy-16121/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-haproxy-0&resourceVersion=1778959337436022000&timeoutSeconds=479&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-haproxy-0 to become Ready.Ok pod/haproxy-haproxy-1 condition met E0516 19:22:24.840631 2802 reflector.go:227] "Failed to watch" err="Get \"https://136.116.224.45/api/v1/namespaces/haproxy-16121/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-haproxy-1&resourceVersion=1778959342436063000&timeoutSeconds=321&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-haproxy-1 to become Ready.Ok pod/haproxy-haproxy-2 condition met E0516 19:22:30.095772 3531 reflector.go:227] "Failed to watch" err="Get \"https://136.116.224.45/api/v1/namespaces/haproxy-16121/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-haproxy-2&resourceVersion=1778959347436181000&timeoutSeconds=530&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-haproxy-2 to become Ready.Ok Unable to use a TTY - input is not a terminal or the right kind of file secret/haproxy-haproxy created ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- waiting for pxc/haproxy to be ready.......... ----------------------------------------------------------------------------------- compare statefulset/haproxy-haproxy--secret ----------------------------------------------------------------------------------- [2026-05-16T19:24:01+0000] compare_kubectl: statefulset/haproxy-haproxy OK Unable to use a TTY - input is not a terminal or the right kind of file ----------------------------------------------------------------------------------- clean up ----------------------------------------------------------------------------------- release "chaos-mesh" uninstalled error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified No resources found No resources found No resources found No resources found No resources found No resources found No resources found No resources found No resources found No resources found No resources found No resources found podchaos.chaos-mesh.org/chaos-pod-failure patched podchaos.chaos-mesh.org "chaos-pod-failure" deleted from haproxy-16121 namespace No resources found No resources found No resources found No resources found No resources found No resources found No resources found No resources found No resources found No resources found customresourcedefinition.apiextensions.k8s.io "awschaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "azurechaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "blockchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "dnschaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "gcpchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "httpchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "iochaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "jvmchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "kernelchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "networkchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "physicalmachinechaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "physicalmachines.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "podchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "podhttpchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "podiochaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "podnetworkchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "remoteclusters.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "schedules.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "statuschecks.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "stresschaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "timechaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "workflownodes.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "workflows.chaos-mesh.org" deleted error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + kubectl patch pxc -n haproxy-16121 haproxy --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/haproxy patched perconaxtradbcluster.pxc.percona.com "haproxy" deleted from haproxy-16121 namespace No resources found No resources found validatingwebhookconfiguration.admissionregistration.k8s.io "percona-xtradbcluster-webhook" deleted ----------------------------------------------------------------------------------- test passed -----------------------------------------------------------------------------------