Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/logs/haproxy-8-4.log Warning: version difference between client (1.36) and server (1.33) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.36) and server (1.33) exceeds the supported minor version skew of +/-1 No resources found + kubectl patch pxc -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: resource(s) were provided, but no name was specified No resources found No resources found No resources found error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- Error from server (NotFound): namespaces "pxc-operator" not found waiting for namespace/pxc-operator to be deletederror: resource(s) were provided, but no name was specified Error from server (NotFound): namespaces "pxc-operator" not found ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- namespace/pxc-operator created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2467-3dc7f023-16-cluster7" modified. ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created pod/percona-xtradb-cluster-operator-9d9fbdb5-pb9rq condition met E0516 19:12:09.094642 23075 reflector.go:227] "Failed to watch" err="Get \"https://35.253.94.202/api/v1/namespaces/pxc-operator/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpercona-xtradb-cluster-operator-9d9fbdb5-pb9rq&resourceVersion=1778958728733656000&timeoutSeconds=457&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" pod/percona-xtradb-cluster-operator-9d9fbdb5-pb9rq condition met E0516 19:12:17.290231 24149 reflector.go:227] "Failed to watch" err="Get \"https://35.253.94.202/api/v1/namespaces/pxc-operator/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpercona-xtradb-cluster-operator-9d9fbdb5-pb9rq&resourceVersion=1778958734674392000&timeoutSeconds=569&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/percona-xtradb-cluster-operator-9d9fbdb5-pb9rq to become Ready.Ok error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces haproxy-25475 ----------------------------------------------------------------------------------- Error from server (NotFound): namespaces "haproxy-25475" not found waiting for namespace/haproxy-25475 to be deletederror: resource(s) were provided, but no name was specified Error from server (NotFound): namespaces "haproxy-25475" not found ----------------------------------------------------------------------------------- create namespace haproxy-25475 ----------------------------------------------------------------------------------- namespace/haproxy-25475 created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2467-3dc7f023-16-cluster7" modified. ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- secret/minio-secret created secret/aws-s3-secret created secret/do-spaces-secret created secret/gcp-cs-secret created secret/azure-secret created error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- install chaos-mesh ----------------------------------------------------------------------------------- "chaos-mesh" already exists with the same configuration, skipping NAME: chaos-mesh LAST DEPLOYED: Sat May 16 19:13:36 2026 NAMESPACE: haproxy-25475 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: 1. Make sure chaos-mesh components are running kubectl get pods --namespace haproxy-25475 -l app.kubernetes.io/instance=chaos-mesh Waiting for DaemonSet chaos-daemon... runtimeclass.node.k8s.io/docker-rc created ----------------------------------------------------------------------------------- create first PXC cluster with HAProxy ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created deployment.apps/pxc-client created perconaxtradbcluster.pxc.percona.com/haproxy created ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- error: no matching resources found ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/haproxy-haproxy-0 condition met waiting for pod/haproxy-haproxy-0 to become Ready.Ok ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/haproxy-pxc-0 condition met E0516 19:15:12.395050 16221 reflector.go:227] "Failed to watch" err="Get \"https://35.253.94.202/api/v1/namespaces/haproxy-25475/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-pxc-0&resourceVersion=1778958908216831023&timeoutSeconds=358&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-pxc-0 to become Ready.Ok pod/haproxy-pxc-1 condition met waiting for pod/haproxy-pxc-1 to become Ready.Ok pod/haproxy-pxc-2 condition met waiting for pod/haproxy-pxc-2 to become Ready.Ok ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- pod/pxc-client-56fd5498cd-4dj8b condition met E0516 19:17:25.764055 2832 reflector.go:227] "Failed to watch" err="Get \"https://35.253.94.202/api/v1/namespaces/haproxy-25475/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-56fd5498cd-4dj8b&resourceVersion=1778959043525167000&timeoutSeconds=399&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-56fd5498cd-4dj8b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-56fd5498cd-4dj8b condition met E0516 19:17:36.968330 4436 reflector.go:227] "Failed to watch" err="Get \"https://35.253.94.202/api/v1/namespaces/haproxy-25475/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-56fd5498cd-4dj8b&resourceVersion=1778959054377036000&timeoutSeconds=313&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-56fd5498cd-4dj8b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-56fd5498cd-4dj8b condition met E0516 19:18:23.551340 10884 reflector.go:227] "Failed to watch" err="Get \"https://35.253.94.202/api/v1/namespaces/haproxy-25475/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-56fd5498cd-4dj8b&resourceVersion=1778959100350183000&timeoutSeconds=449&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-56fd5498cd-4dj8b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-56fd5498cd-4dj8b condition met E0516 19:18:39.434698 13133 reflector.go:227] "Failed to watch" err="Get \"https://35.253.94.202/api/v1/namespaces/haproxy-25475/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-56fd5498cd-4dj8b&resourceVersion=1778959116710854000&timeoutSeconds=520&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-56fd5498cd-4dj8b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-56fd5498cd-4dj8b condition met E0516 19:18:55.857849 15154 reflector.go:227] "Failed to watch" err="Get \"https://35.253.94.202/api/v1/namespaces/haproxy-25475/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-56fd5498cd-4dj8b&resourceVersion=1778959132079641000&timeoutSeconds=593&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-56fd5498cd-4dj8b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- checking all haproxy pods point to the same writer ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/haproxy-pxc-0 condition met E0516 19:19:13.201247 17436 reflector.go:227] "Failed to watch" err="Get \"https://35.253.94.202/api/v1/namespaces/haproxy-25475/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-pxc-0&resourceVersion=1778959152080136000&timeoutSeconds=442&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-pxc-0 to become Ready.Ok pod/haproxy-pxc-1 condition met E0516 19:19:26.794783 19344 reflector.go:227] "Failed to watch" err="Get \"https://35.253.94.202/api/v1/namespaces/haproxy-25475/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-pxc-1&resourceVersion=1778959162080247000&timeoutSeconds=455&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-pxc-1 to become Ready.Ok pod/haproxy-pxc-2 condition met E0516 19:19:38.156749 21281 reflector.go:227] "Failed to watch" err="Get \"https://35.253.94.202/api/v1/namespaces/haproxy-25475/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-pxc-2&resourceVersion=1778959177080323000&timeoutSeconds=308&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-pxc-2 to become Ready.Ok ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- waiting for pxc/haproxy to be ready pod/pxc-client-56fd5498cd-4dj8b condition met E0516 19:20:17.890173 26202 reflector.go:227] "Failed to watch" err="Get \"https://35.253.94.202/api/v1/namespaces/haproxy-25475/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-56fd5498cd-4dj8b&resourceVersion=1778959213649166000&timeoutSeconds=420&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-56fd5498cd-4dj8b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-56fd5498cd-4dj8b condition met E0516 19:20:35.496793 28812 reflector.go:227] "Failed to watch" err="Get \"https://35.253.94.202/api/v1/namespaces/haproxy-25475/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-56fd5498cd-4dj8b&resourceVersion=1778959232651467000&timeoutSeconds=554&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-56fd5498cd-4dj8b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-56fd5498cd-4dj8b condition met E0516 19:20:51.022140 31370 reflector.go:227] "Failed to watch" err="Get \"https://35.253.94.202/api/v1/namespaces/haproxy-25475/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-56fd5498cd-4dj8b&resourceVersion=1778959249953854000&timeoutSeconds=408&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-56fd5498cd-4dj8b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- delete active writer and checking all haproxy pods still point to the same writer ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- fail pxc-pod-0 pod for 60s ----------------------------------------------------------------------------------- podchaos.chaos-mesh.org/chaos-pod-failure created NAME READY STATUS RESTARTS AGE chaos-controller-manager-8ffdfcd58-2hsq4 1/1 Running 0 7m43s chaos-controller-manager-8ffdfcd58-92wk4 1/1 Running 0 7m42s chaos-controller-manager-8ffdfcd58-vrzj2 1/1 Running 0 7m42s chaos-daemon-j825m 1/1 Running 0 7m43s chaos-daemon-kd7p2 1/1 Running 0 7m43s chaos-daemon-wsqb9 1/1 Running 0 7m43s chaos-dns-server-6648bb7956-lmsxp 1/1 Running 0 7m43s haproxy-haproxy-0 3/3 Running 0 7m12s haproxy-haproxy-1 3/3 Running 0 6m27s haproxy-haproxy-2 3/3 Running 0 6m6s haproxy-pxc-0 0/1 CrashLoopBackOff 2 (9s ago) 7m12s haproxy-pxc-1 1/1 Running 0 6m28s haproxy-pxc-2 1/1 Running 0 5m43s pxc-client-56fd5498cd-4dj8b 2/2 Running 0 7m18s pod/pxc-client-56fd5498cd-4dj8b condition met E0516 19:21:35.526576 5146 reflector.go:227] "Failed to watch" err="Get \"https://35.253.94.202/api/v1/namespaces/haproxy-25475/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-56fd5498cd-4dj8b&resourceVersion=1778959294523625000&timeoutSeconds=579&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-56fd5498cd-4dj8b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-56fd5498cd-4dj8b condition met E0516 19:21:47.534036 6546 reflector.go:227] "Failed to watch" err="Get \"https://35.253.94.202/api/v1/namespaces/haproxy-25475/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-56fd5498cd-4dj8b&resourceVersion=1778959305231295000&timeoutSeconds=522&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-56fd5498cd-4dj8b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-56fd5498cd-4dj8b condition met E0516 19:22:07.312220 9166 reflector.go:227] "Failed to watch" err="Get \"https://35.253.94.202/api/v1/namespaces/haproxy-25475/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-56fd5498cd-4dj8b&resourceVersion=1778959323936060000&timeoutSeconds=321&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-56fd5498cd-4dj8b to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/haproxy-pxc-0 condition met waiting for pod/haproxy-pxc-0 to become Ready.Ok pod/haproxy-pxc-1 condition met E0516 19:22:46.262841 14656 reflector.go:227] "Failed to watch" err="Get \"https://35.253.94.202/api/v1/namespaces/haproxy-25475/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-pxc-1&resourceVersion=1778959365035090000&timeoutSeconds=484&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-pxc-1 to become Ready.Ok pod/haproxy-pxc-2 condition met E0516 19:22:50.248458 15117 reflector.go:227] "Failed to watch" err="Get \"https://35.253.94.202/api/v1/namespaces/haproxy-25475/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-pxc-2&resourceVersion=1778959370035141000&timeoutSeconds=443&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-pxc-2 to become Ready.Ok ----------------------------------------------------------------------------------- check advanced options are enabled in haproxy statefulset ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- compare pdb/haproxy-haproxy- ----------------------------------------------------------------------------------- [2026-05-16T19:22:55+0000] compare_kubectl: pdb/haproxy-haproxy OK ----------------------------------------------------------------------------------- compare statefulset/haproxy-haproxy- ----------------------------------------------------------------------------------- [2026-05-16T19:22:57+0000] compare_kubectl: statefulset/haproxy-haproxy OK ----------------------------------------------------------------------------------- default haproxy-replicas service ----------------------------------------------------------------------------------- The request is invalid: the server rejected our request due to an error in our request ----------------------------------------------------------------------------------- compare service/haproxy-haproxy-replicas- ----------------------------------------------------------------------------------- [2026-05-16T19:23:00+0000] compare_kubectl: service/haproxy-haproxy-replicas OK ----------------------------------------------------------------------------------- disable haproxy-replicas service ----------------------------------------------------------------------------------- waiting for svc/haproxy-haproxy-replicas to be deletedError from server (NotFound): services "haproxy-haproxy-replicas" not found ----------------------------------------------------------------------------------- enable haproxy-replicas service ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- compare service/haproxy-haproxy-replicas- ----------------------------------------------------------------------------------- [2026-05-16T19:23:41+0000] compare_kubectl: service/haproxy-haproxy-replicas OK ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/haproxy-haproxy-0 condition met E0516 19:23:42.980282 21724 reflector.go:227] "Failed to watch" err="Get \"https://35.253.94.202/api/v1/namespaces/haproxy-25475/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-haproxy-0&resourceVersion=1778959420036297000&timeoutSeconds=576&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-haproxy-0 to become Ready.Ok pod/haproxy-haproxy-1 condition met E0516 19:23:50.275907 22705 reflector.go:227] "Failed to watch" err="Get \"https://35.253.94.202/api/v1/namespaces/haproxy-25475/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-haproxy-1&resourceVersion=1778959430037002000&timeoutSeconds=553&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-haproxy-1 to become Ready.Ok pod/haproxy-haproxy-2 condition met E0516 19:23:57.122669 23612 reflector.go:227] "Failed to watch" err="Get \"https://35.253.94.202/api/v1/namespaces/haproxy-25475/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dhaproxy-haproxy-2&resourceVersion=1778959435037050000&timeoutSeconds=396&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/haproxy-haproxy-2 to become Ready.Ok Unable to use a TTY - input is not a terminal or the right kind of file secret/haproxy-haproxy created ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- waiting for pxc/haproxy to be ready......... ----------------------------------------------------------------------------------- compare statefulset/haproxy-haproxy--secret ----------------------------------------------------------------------------------- [2026-05-16T19:25:28+0000] compare_kubectl: statefulset/haproxy-haproxy OK Unable to use a TTY - input is not a terminal or the right kind of file ----------------------------------------------------------------------------------- clean up ----------------------------------------------------------------------------------- release "chaos-mesh" uninstalled error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified No resources found No resources found No resources found No resources found No resources found No resources found No resources found No resources found No resources found No resources found No resources found No resources found podchaos.chaos-mesh.org/chaos-pod-failure patched podchaos.chaos-mesh.org "chaos-pod-failure" deleted from haproxy-25475 namespace No resources found No resources found No resources found No resources found No resources found No resources found No resources found No resources found No resources found No resources found customresourcedefinition.apiextensions.k8s.io "awschaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "azurechaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "blockchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "dnschaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "gcpchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "httpchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "iochaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "jvmchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "kernelchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "networkchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "physicalmachinechaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "physicalmachines.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "podchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "podhttpchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "podiochaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "podnetworkchaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "remoteclusters.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "schedules.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "statuschecks.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "stresschaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "timechaos.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "workflownodes.chaos-mesh.org" deleted customresourcedefinition.apiextensions.k8s.io "workflows.chaos-mesh.org" deleted error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + kubectl patch pxc -n haproxy-25475 haproxy --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/haproxy patched perconaxtradbcluster.pxc.percona.com "haproxy" deleted from haproxy-25475 namespace No resources found No resources found validatingwebhookconfiguration.admissionregistration.k8s.io "percona-xtradbcluster-webhook" deleted ----------------------------------------------------------------------------------- test passed -----------------------------------------------------------------------------------