Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/logs/haproxy-5-7.log Warning: version difference between client (1.36) and server (1.33) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.36) and server (1.33) exceeds the supported minor version skew of +/-1 No resources found + kubectl patch pxc -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: resource(s) were provided, but no name was specified No resources found No resources found No resources found error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- error: resource(s) were provided, but no name was specified namespace "pxc-operator" deleted waiting for namespace/pxc-operator to be deletedError from server (NotFound): namespaces "pxc-operator" not found ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- namespace/pxc-operator created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2473-6d392bea-4-cluster2" modified. ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created pod/percona-xtradb-cluster-operator-55d95dc9d8-rhsjg condition met E0516 23:37:42.043747 7088 reflector.go:227] "Failed to watch" err="Get \"https://34.61.233.25/api/v1/namespaces/pxc-operator/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpercona-xtradb-cluster-operator-55d95dc9d8-rhsjg&resourceVersion=1778974661682114000&timeoutSeconds=496&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" pod/percona-xtradb-cluster-operator-55d95dc9d8-rhsjg condition met E0516 23:37:49.073144 8103 reflector.go:227] "Failed to watch" err="Get \"https://34.61.233.25/api/v1/namespaces/pxc-operator/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpercona-xtradb-cluster-operator-55d95dc9d8-rhsjg&resourceVersion=1778974665934265000&timeoutSeconds=566&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/percona-xtradb-cluster-operator-55d95dc9d8-rhsjg to become Ready.Ok error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces haproxy-12465 ----------------------------------------------------------------------------------- error: resource(s) were provided, but no name was specified Error from server (NotFound): namespaces "haproxy-12465" not found waiting for namespace/haproxy-12465 to be deletedError from server (NotFound): namespaces "haproxy-12465" not found ----------------------------------------------------------------------------------- create namespace haproxy-12465 ----------------------------------------------------------------------------------- namespace/haproxy-12465 created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2473-6d392bea-4-cluster2" modified. ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- secret/minio-secret created secret/aws-s3-secret created secret/do-spaces-secret created secret/gcp-cs-secret created secret/azure-secret created error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- install chaos-mesh ----------------------------------------------------------------------------------- "chaos-mesh" already exists with the same configuration, skipping NAME: chaos-mesh LAST DEPLOYED: Sat May 16 23:39:22 2026 NAMESPACE: haproxy-12465 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: 1. Make sure chaos-mesh components are running kubectl get pods --namespace haproxy-12465 -l app.kubernetes.io/instance=chaos-mesh Waiting for DaemonSet chaos-daemon... runtimeclass.node.k8s.io/docker-rc created ----------------------------------------------------------------------------------- create first PXC cluster with HAProxy ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created deployment.apps/pxc-client created perconaxtradbcluster.pxc.percona.com/haproxy created ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- error: no matching resources found ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/haproxy-haproxy-0 condition met waiting for pod/haproxy-haproxy-0 to become Ready.Ok ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/haproxy-pxc-0 condition met E0516 23:41:29.766581 3352 reflector.go:227] "Failed to watch" err="Get \"https://34.61.233.25/api/v1/namespaces/haproxy-12465/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-pxc-0&resourceVersion=1778974888414671003&timeoutSeconds=495&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-pxc-0 to become Ready.Ok pod/haproxy-pxc-1 condition met waiting for pod/haproxy-pxc-1 to become Ready.Ok pod/haproxy-pxc-2 condition met waiting for pod/haproxy-pxc-2 to become Ready.Ok ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- pod/pxc-client-97cb9c68b-fk2fl condition met E0516 23:44:11.773674 24503 reflector.go:227] "Failed to watch" err="Get \"https://34.61.233.25/api/v1/namespaces/haproxy-12465/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-fk2fl&resourceVersion=1778975049922496000&timeoutSeconds=530&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-fk2fl to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-fk2fl condition met E0516 23:44:20.872908 25896 reflector.go:227] "Failed to watch" err="Get \"https://34.61.233.25/api/v1/namespaces/haproxy-12465/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-fk2fl&resourceVersion=1778975060069522000&timeoutSeconds=364&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-fk2fl to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-fk2fl condition met E0516 23:45:00.851794 31365 reflector.go:227] "Failed to watch" err="Get \"https://34.61.233.25/api/v1/namespaces/haproxy-12465/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-fk2fl&resourceVersion=1778975098713591000&timeoutSeconds=326&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-fk2fl to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-fk2fl condition met E0516 23:45:07.830329 32527 reflector.go:227] "Failed to watch" err="Get \"https://34.61.233.25/api/v1/namespaces/haproxy-12465/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-fk2fl&resourceVersion=1778975107028504000&timeoutSeconds=456&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-fk2fl to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-fk2fl condition met E0516 23:45:15.583346 1100 reflector.go:227] "Failed to watch" err="Get \"https://34.61.233.25/api/v1/namespaces/haproxy-12465/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-fk2fl&resourceVersion=1778975114775816000&timeoutSeconds=452&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-fk2fl to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- checking all haproxy pods point to the same writer ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/haproxy-pxc-0 condition met E0516 23:45:24.810024 2197 reflector.go:227] "Failed to watch" err="Get \"https://34.61.233.25/api/v1/namespaces/haproxy-12465/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-pxc-0&resourceVersion=1778975119775842000&timeoutSeconds=344&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-pxc-0 to become Ready.Ok pod/haproxy-pxc-1 condition met E0516 23:45:28.512679 2601 reflector.go:227] "Failed to watch" err="Get \"https://34.61.233.25/api/v1/namespaces/haproxy-12465/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-pxc-1&resourceVersion=1778975124775896000&timeoutSeconds=459&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-pxc-1 to become Ready.Ok pod/haproxy-pxc-2 condition met E0516 23:45:32.969413 3138 reflector.go:227] "Failed to watch" err="Get \"https://34.61.233.25/api/v1/namespaces/haproxy-12465/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-pxc-2&resourceVersion=1778975129775969000&timeoutSeconds=349&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-pxc-2 to become Ready.Ok ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- waiting for pxc/haproxy to be ready pod/pxc-client-97cb9c68b-fk2fl condition met E0516 23:45:55.587199 6033 reflector.go:227] "Failed to watch" err="Get \"https://34.61.233.25/api/v1/namespaces/haproxy-12465/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-fk2fl&resourceVersion=1778975153186280000&timeoutSeconds=438&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-fk2fl to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-fk2fl condition met E0516 23:46:07.244121 7845 reflector.go:227] "Failed to watch" err="Get \"https://34.61.233.25/api/v1/namespaces/haproxy-12465/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-fk2fl&resourceVersion=1778975164971605000&timeoutSeconds=553&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-fk2fl to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-fk2fl condition met E0516 23:46:21.903125 9921 reflector.go:227] "Failed to watch" err="Get \"https://34.61.233.25/api/v1/namespaces/haproxy-12465/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-fk2fl&resourceVersion=1778975179808417000&timeoutSeconds=369&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-fk2fl to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- delete active writer and checking all haproxy pods still point to the same writer ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- fail pxc-pod-0 pod for 60s ----------------------------------------------------------------------------------- podchaos.chaos-mesh.org/chaos-pod-failure created NAME READY STATUS RESTARTS AGE chaos-controller-manager-7bbb9bc8bc-9mdn6 1/1 Running 0 7m18s chaos-controller-manager-7bbb9bc8bc-fwzbt 1/1 Running 0 7m19s chaos-controller-manager-7bbb9bc8bc-qmr56 1/1 Running 0 7m18s chaos-daemon-47dpd 1/1 Running 0 7m18s chaos-daemon-5hglr 1/1 Running 0 7m19s chaos-daemon-fvdc9 1/1 Running 0 7m18s chaos-dns-server-6648bb7956-hvr6j 1/1 Running 0 7m19s haproxy-haproxy-0 3/3 Running 0 6m54s haproxy-haproxy-1 3/3 Running 0 5m29s haproxy-haproxy-2 3/3 Running 0 5m9s haproxy-pxc-0 0/1 CrashLoopBackOff 2 (7s ago) 6m54s haproxy-pxc-1 1/1 Running 0 5m32s haproxy-pxc-2 1/1 Running 0 4m19s pxc-client-97cb9c68b-fk2fl 2/2 Running 0 6m58s pod/pxc-client-97cb9c68b-fk2fl condition met E0516 23:46:57.349943 14120 reflector.go:227] "Failed to watch" err="Get \"https://34.61.233.25/api/v1/namespaces/haproxy-12465/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-fk2fl&resourceVersion=1778975216604427000&timeoutSeconds=538&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-fk2fl to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-fk2fl condition met E0516 23:47:04.349069 14867 reflector.go:227] "Failed to watch" err="Get \"https://34.61.233.25/api/v1/namespaces/haproxy-12465/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-fk2fl&resourceVersion=1778975222630583000&timeoutSeconds=402&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-fk2fl to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-fk2fl condition met E0516 23:47:13.327494 16181 reflector.go:227] "Failed to watch" err="Get \"https://34.61.233.25/api/v1/namespaces/haproxy-12465/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-fk2fl&resourceVersion=1778975231901175000&timeoutSeconds=536&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-fk2fl to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/haproxy-pxc-0 condition met waiting for pod/haproxy-pxc-0 to become Ready.Ok pod/haproxy-pxc-1 condition met E0516 23:48:12.763104 24496 reflector.go:227] "Failed to watch" err="Get \"https://34.61.233.25/api/v1/namespaces/haproxy-12465/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-pxc-1&resourceVersion=1778975288875103000&timeoutSeconds=386&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-pxc-1 to become Ready.Ok pod/haproxy-pxc-2 condition met E0516 23:48:20.550741 25664 reflector.go:227] "Failed to watch" err="Get \"https://34.61.233.25/api/v1/namespaces/haproxy-12465/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-pxc-2&resourceVersion=1778975298875161000&timeoutSeconds=450&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-pxc-2 to become Ready.Ok ----------------------------------------------------------------------------------- check advanced options are enabled in haproxy statefulset ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- compare pdb/haproxy-haproxy- ----------------------------------------------------------------------------------- [2026-05-16T23:48:27+0000] compare_kubectl: pdb/haproxy-haproxy OK ----------------------------------------------------------------------------------- compare statefulset/haproxy-haproxy- ----------------------------------------------------------------------------------- [2026-05-16T23:48:29+0000] compare_kubectl: statefulset/haproxy-haproxy OK ----------------------------------------------------------------------------------- default haproxy-replicas service ----------------------------------------------------------------------------------- The request is invalid: the server rejected our request due to an error in our request ----------------------------------------------------------------------------------- compare service/haproxy-haproxy-replicas- ----------------------------------------------------------------------------------- [2026-05-16T23:48:33+0000] compare_kubectl: service/haproxy-haproxy-replicas OK ----------------------------------------------------------------------------------- disable haproxy-replicas service ----------------------------------------------------------------------------------- waiting for svc/haproxy-haproxy-replicas to be deletedError from server (NotFound): services "haproxy-haproxy-replicas" not found ----------------------------------------------------------------------------------- enable haproxy-replicas service ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- compare service/haproxy-haproxy-replicas- ----------------------------------------------------------------------------------- [2026-05-16T23:49:18+0000] compare_kubectl: service/haproxy-haproxy-replicas OK ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/haproxy-haproxy-0 condition met E0516 23:49:20.500086 2177 reflector.go:227] "Failed to watch" err="Get \"https://34.61.233.25/api/v1/namespaces/haproxy-12465/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-haproxy-0&resourceVersion=1778975358877515000&timeoutSeconds=570&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-haproxy-0 to become Ready.Ok pod/haproxy-haproxy-1 condition met E0516 23:49:29.509855 3503 reflector.go:227] "Failed to watch" err="Get \"https://34.61.233.25/api/v1/namespaces/haproxy-12465/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-haproxy-1&resourceVersion=1778975368877938000&timeoutSeconds=466&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-haproxy-1 to become Ready.Ok pod/haproxy-haproxy-2 condition met E0516 23:49:37.651773 4643 reflector.go:227] "Failed to watch" err="Get \"https://34.61.233.25/api/v1/namespaces/haproxy-12465/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-haproxy-2&resourceVersion=1778975373877969000&timeoutSeconds=418&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-haproxy-2 to become Ready.Ok Unable to use a TTY - input is not a terminal or the right kind of file secret/haproxy-haproxy created ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- waiting for pxc/haproxy to be ready......... ----------------------------------------------------------------------------------- compare statefulset/haproxy-haproxy--secret ----------------------------------------------------------------------------------- [2026-05-16T23:51:20+0000] compare_kubectl: statefulset/haproxy-haproxy OK Unable to use a TTY - input is not a terminal or the right kind of file ----------------------------------------------------------------------------------- clean up ----------------------------------------------------------------------------------- release "chaos-mesh" uninstalled error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified No resources found No resources found No resources found No resources found No resources found No resources found No resources found No resources found No resources found No resources found No resources found No resources found podchaos.chaos-mesh.org/chaos-pod-failure patched podchaos.chaos-mesh.org "chaos-pod-failure" deleted from haproxy-12465 namespace No resources found No resources found No resources found No resources found No resources found No resources found No resources found No resources found No resources found No resources found customresourcedefinition.apiextensions.k8s.io "awschaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "azurechaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "blockchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "dnschaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "gcpchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "httpchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "iochaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "jvmchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "kernelchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "networkchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "physicalmachinechaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "physicalmachines.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "podchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "podhttpchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "podiochaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "podnetworkchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "remoteclusters.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "schedules.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "statuschecks.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "stresschaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "timechaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "workflownodes.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "workflows.chaos-mesh.org" deleted error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + kubectl patch pxc -n haproxy-12465 haproxy --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/haproxy patched perconaxtradbcluster.pxc.percona.com "haproxy" deleted from haproxy-12465 namespace No resources found No resources found validatingwebhookconfiguration.admissionregistration.k8s.io "percona-xtradbcluster-webhook" deleted ----------------------------------------------------------------------------------- test passed -----------------------------------------------------------------------------------