Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2476/e2e-tests/logs/monitoring-2-0-8-0.log Warning: version difference between client (1.36) and server (1.33) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.36) and server (1.33) exceeds the supported minor version skew of +/-1 No resources found + kubectl patch pxc -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: resource(s) were provided, but no name was specified No resources found No resources found No resources found error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- Error from server (NotFound): namespaces "pxc-operator" not found waiting for namespace/pxc-operator to be deletederror: resource(s) were provided, but no name was specified Error from server (NotFound): namespaces "pxc-operator" not found ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- namespace/pxc-operator created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2476-a8b01a39-5-cluster3" modified. ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created pod/percona-xtradb-cluster-operator-8548fd5788-2bwdn condition met E0516 19:20:14.139746 17738 reflector.go:227] "Failed to watch" err="Get \"https://35.232.100.53/api/v1/namespaces/pxc-operator/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpercona-xtradb-cluster-operator-8548fd5788-2bwdn&resourceVersion=1778959213779988000&timeoutSeconds=404&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" pod/percona-xtradb-cluster-operator-8548fd5788-2bwdn condition met E0516 19:20:21.108688 18725 reflector.go:227] "Failed to watch" err="Get \"https://35.232.100.53/api/v1/namespaces/pxc-operator/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpercona-xtradb-cluster-operator-8548fd5788-2bwdn&resourceVersion=1778959218479420000&timeoutSeconds=399&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/percona-xtradb-cluster-operator-8548fd5788-2bwdn to become Ready.Ok error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces monitoring-2-0-4388 ----------------------------------------------------------------------------------- Error from server (NotFound): namespaces "monitoring-2-0-4388" not found waiting for namespace/monitoring-2-0-4388 to be deletederror: resource(s) were provided, but no name was specified Error from server (NotFound): namespaces "monitoring-2-0-4388" not found ----------------------------------------------------------------------------------- create namespace monitoring-2-0-4388 ----------------------------------------------------------------------------------- namespace/monitoring-2-0-4388 created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2476-a8b01a39-5-cluster3" modified. ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- secret/minio-secret created secret/aws-s3-secret created secret/do-spaces-secret created secret/gcp-cs-secret created secret/azure-secret created "hashicorp" already exists with the same configuration, skipping "minio" already exists with the same configuration, skipping Hang tight while we grab the latest from your chart repositories... ...Successfully got an update from the "minio" chart repository ...Successfully got an update from the "chaos-mesh" chart repository ...Successfully got an update from the "hashicorp" chart repository Update Complete. ⎈Happy Helming!⎈ ----------------------------------------------------------------------------------- install PMM Server ----------------------------------------------------------------------------------- "percona" has been added to your repositories Hang tight while we grab the latest from your chart repositories... ...Successfully got an update from the "minio" chart repository ...Successfully got an update from the "chaos-mesh" chart repository ...Successfully got an update from the "hashicorp" chart repository ...Successfully got an update from the "percona" chart repository Update Complete. ⎈Happy Helming!⎈ Error: uninstall: Release not loaded: monitoring: release: not found NAME: monitoring LAST DEPLOYED: Sat May 16 19:21:00 2026 NAMESPACE: monitoring-2-0-4388 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: PMM server can be accessed via HTTPS (port 443) on the following DNS name from within your cluster: endpoint: https://monitoring-service.monitoring-2-0-4388.svc.cluster.local:443 login: admin password: admin statefulset.apps/monitoring condition met E0516 19:21:41.495637 29667 reflector.go:227] "Failed to watch" err="Get \"https://35.232.100.53/apis/apps/v1/namespaces/monitoring-2-0-4388/statefulsets?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dmonitoring&resourceVersion=1778959300000033000&timeoutSeconds=532&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" logger=settings t=2026-05-16T19:21:51.642860855Z level=info msg="Starting Grafana" version= commit= branch= compiled=1970-01-01T00:00:00Z logger=settings t=2026-05-16T19:21:51.642995835Z level=info msg="Config loaded from" file=/usr/share/grafana/conf/defaults.ini logger=settings t=2026-05-16T19:21:51.643008335Z level=info msg="Config loaded from" file=/etc/grafana/grafana.ini logger=settings t=2026-05-16T19:21:51.643014145Z level=info msg="Path Home" path=/usr/share/grafana logger=settings t=2026-05-16T19:21:51.643019015Z level=info msg="Path Data" path=/srv/grafana logger=settings t=2026-05-16T19:21:51.643023375Z level=info msg="Path Logs" path=/srv/logs logger=settings t=2026-05-16T19:21:51.643027645Z level=info msg="Path Plugins" path=/srv/grafana/plugins logger=settings t=2026-05-16T19:21:51.643032535Z level=info msg="Path Provisioning" path=/usr/share/grafana/conf/provisioning logger=settings t=2026-05-16T19:21:51.643037415Z level=info msg="App mode production" logger=sqlstore t=2026-05-16T19:21:51.643095955Z level=info msg="Connecting to DB" dbtype=postgres logger=migrator t=2026-05-16T19:21:51.657691893Z level=info msg="Starting DB migrations" logger=migrator t=2026-05-16T19:21:51.661062523Z level=info msg="migrations completed" performed=0 skipped=452 duration=306.21µs logger=secrets t=2026-05-16T19:21:51.662339613Z level=info msg="Envelope encryption state" enabled=true currentprovider=secretKey.v1 logger=plugin.finder t=2026-05-16T19:21:51.69131596Z level=warn msg="Skipping finding plugins as directory does not exist" path=/usr/share/grafana/plugins-bundled logger=plugin.signature.validator t=2026-05-16T19:21:51.814775666Z level=warn msg="Permitting unsigned plugin. This is not recommended" pluginID=pmm-qan-app-panel pluginDir=/srv/grafana/plugins/pmm-app/dist/pmm-qan logger=plugin.signature.validator t=2026-05-16T19:21:51.814795506Z level=warn msg="Permitting unsigned plugin. This is not recommended" pluginID=grafana-polystat-panel pluginDir=/srv/grafana/plugins/grafana-polystat-panel logger=plugin.signature.validator t=2026-05-16T19:21:51.814824246Z level=warn msg="Permitting unsigned plugin. This is not recommended" pluginID=pmm-app pluginDir=/srv/grafana/plugins/pmm-app/dist logger=plugin.pmm-app t=2026-05-16T19:21:51.820157456Z level=warn msg="Included dashboard is missing a UID field" logger=plugin.signature.validator t=2026-05-16T19:21:51.820326596Z level=warn msg="Permitting unsigned plugin. This is not recommended" pluginID=pmm-pt-summary-panel pluginDir=/srv/grafana/plugins/pmm-app/dist/pmm-pt-summary/panel logger=plugin.signature.validator t=2026-05-16T19:21:51.820376816Z level=warn msg="Permitting unsigned plugin. This is not recommended" pluginID=pmm-pt-summary-datasource pluginDir=/srv/grafana/plugins/pmm-app/dist/pmm-pt-summary/datasource logger=plugin.loader t=2026-05-16T19:21:51.820441256Z level=info msg="Plugin registered" pluginID=camptocamp-prometheus-alertmanager-datasource logger=plugin.loader t=2026-05-16T19:21:51.820460256Z level=info msg="Plugin registered" pluginID=jdbranham-diagram-panel logger=plugin.loader t=2026-05-16T19:21:51.820467496Z level=info msg="Plugin registered" pluginID=pmm-qan-app-panel logger=plugin.loader t=2026-05-16T19:21:51.820473725Z level=info msg="Plugin registered" pluginID=grafana-polystat-panel logger=plugin.loader t=2026-05-16T19:21:51.820478976Z level=info msg="Plugin registered" pluginID=grafana-worldmap-panel logger=plugin.loader t=2026-05-16T19:21:51.820484496Z level=info msg="Plugin registered" pluginID=natel-discrete-panel logger=plugin.loader t=2026-05-16T19:21:51.820490056Z level=info msg="Plugin registered" pluginID=pmm-app logger=plugin.loader t=2026-05-16T19:21:51.820495216Z level=info msg="Plugin registered" pluginID=pmm-pt-summary-panel logger=plugin.loader t=2026-05-16T19:21:51.820500476Z level=info msg="Plugin registered" pluginID=grafana-clickhouse-datasource logger=plugin.grafana-clickhouse-datasource t=2026-05-16T19:21:51.831837114Z level=warn msg="Plugin process is running with elevated privileges. This is not recommended" logger=plugin.loader t=2026-05-16T19:21:51.831956094Z level=info msg="Plugin registered" pluginID=grafana-piechart-panel logger=plugin.loader t=2026-05-16T19:21:51.831980244Z level=info msg="Plugin registered" pluginID=petrslavotinek-carpetplot-panel logger=plugin.loader t=2026-05-16T19:21:51.831987044Z level=info msg="Plugin registered" pluginID=pmm-pt-summary-datasource Admin password changed successfully ✔ ----------------------------------------------------------------------------------- create PXC cluster ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- secret/my-cluster-secrets created deployment.apps/pxc-client created perconaxtradbcluster.pxc.percona.com/monitoring created ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- error: no matching resources found ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- Error from server (NotFound): pods "monitoring-haproxy-0" not found waiting for pod/monitoring-haproxy-0 to become Ready..........................Ok ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/monitoring-pxc-0 condition met E0516 19:23:08.676652 8841 reflector.go:227] "Failed to watch" err="Get \"https://35.232.100.53/api/v1/namespaces/monitoring-2-0-4388/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dmonitoring-pxc-0&resourceVersion=1778959386157583020&timeoutSeconds=389&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/monitoring-pxc-0 to become Ready.Ok pod/monitoring-pxc-1 condition met waiting for pod/monitoring-pxc-1 to become Ready.Ok pod/monitoring-pxc-2 condition met waiting for pod/monitoring-pxc-2 to become Ready.Ok ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- pod/pxc-client-67fc4995bb-rvk2g condition met E0516 19:27:52.919383 9500 reflector.go:227] "Failed to watch" err="Get \"https://35.232.100.53/api/v1/namespaces/monitoring-2-0-4388/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-rvk2g&resourceVersion=1778959671504550000&timeoutSeconds=339&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-rvk2g to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-67fc4995bb-rvk2g condition met E0516 19:27:59.691909 10712 reflector.go:227] "Failed to watch" err="Get \"https://35.232.100.53/api/v1/namespaces/monitoring-2-0-4388/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-rvk2g&resourceVersion=1778959678366762000&timeoutSeconds=557&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-rvk2g to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-67fc4995bb-rvk2g condition met E0516 19:28:36.185184 14812 reflector.go:227] "Failed to watch" err="Get \"https://35.232.100.53/api/v1/namespaces/monitoring-2-0-4388/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-rvk2g&resourceVersion=1778959714617661000&timeoutSeconds=331&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-rvk2g to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-67fc4995bb-rvk2g condition met E0516 19:28:45.438665 16258 reflector.go:227] "Failed to watch" err="Get \"https://35.232.100.53/api/v1/namespaces/monitoring-2-0-4388/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-rvk2g&resourceVersion=1778959724059250000&timeoutSeconds=396&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-rvk2g to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-67fc4995bb-rvk2g condition met E0516 19:28:53.413348 17273 reflector.go:227] "Failed to watch" err="Get \"https://35.232.100.53/api/v1/namespaces/monitoring-2-0-4388/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-rvk2g&resourceVersion=1778959731943001000&timeoutSeconds=518&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-rvk2g to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- add PMM API key to secret ----------------------------------------------------------------------------------- % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 155 100 119 100 36 274 82 --:--:-- --:--:-- --:--:-- 357 secret/my-cluster-secrets patched Waiting for sts/monitoring-pxc to reach generation 2... Resource sts/monitoring-pxc is at generation 1. Waiting... Resource sts/monitoring-pxc has reached generation 2. Waiting for sts/monitoring-haproxy to reach generation 2... Resource sts/monitoring-haproxy has reached generation 2. E0516 19:29:24.572539 21088 reflector.go:227] "Failed to watch" err="Get \"https://35.232.100.53/api/v1/namespaces/monitoring-2-0-4388/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dmonitoring-haproxy-0&resourceVersion=1778959763877473000&timeoutSeconds=415&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" pod/monitoring-haproxy-0 condition met pod/monitoring-haproxy-1 condition met E0516 19:29:36.287402 21088 reflector.go:227] "Failed to watch" err="Get \"https://35.232.100.53/api/v1/namespaces/monitoring-2-0-4388/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dmonitoring-pxc-0&resourceVersion=1778959775972367021&timeoutSeconds=368&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" pod/monitoring-pxc-0 condition met E0516 19:29:36.588414 21088 reflector.go:227] "Failed to watch" err="Get \"https://35.232.100.53/api/v1/namespaces/monitoring-2-0-4388/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dmonitoring-pxc-1&resourceVersion=1778959775972367021&timeoutSeconds=342&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" pod/monitoring-pxc-1 condition met pod/monitoring-pxc-2 condition met ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- waiting for pxc/monitoring to be ready................. ----------------------------------------------------------------------------------- compare statefulset/monitoring-pxc--no-prefix ----------------------------------------------------------------------------------- [2026-05-16T19:32:04+0000] compare_kubectl: statefulset/monitoring-pxc OK ----------------------------------------------------------------------------------- compare statefulset/monitoring-haproxy--no-prefix ----------------------------------------------------------------------------------- [2026-05-16T19:32:05+0000] compare_kubectl: statefulset/monitoring-haproxy OK ----------------------------------------------------------------------------------- apply my-env-var-secrets to add PMM_PREFIX ----------------------------------------------------------------------------------- secret/my-env-var-secrets created Waiting for sts/monitoring-pxc to reach generation 3... Resource sts/monitoring-pxc has reached generation 3. Waiting for sts/monitoring-haproxy to reach generation 3... Resource sts/monitoring-haproxy has reached generation 3. ----------------------------------------------------------------------------------- add new PMM API key to secret ----------------------------------------------------------------------------------- % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 167 100 127 100 40 291 91 --:--:-- --:--:-- --:--:-- 383 secret/my-cluster-secrets patched ----------------------------------------------------------------------------------- delete old PMM key ----------------------------------------------------------------------------------- % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 597 100 597 0 0 1379 0 --:--:-- --:--:-- --:--:-- 1378 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 29 100 29 0 0 68 0 --:--:-- --:--:-- --:--:-- 67 {"message":"API key deleted"}Waiting for sts/monitoring-pxc to reach generation 4... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc has reached generation 4. Waiting for sts/monitoring-haproxy to reach generation 4... Resource sts/monitoring-haproxy has reached generation 4. E0516 19:35:10.198072 1674 reflector.go:227] "Failed to watch" err="Get \"https://35.232.100.53/api/v1/namespaces/monitoring-2-0-4388/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dmonitoring-haproxy-0&resourceVersion=1778960109514647000&timeoutSeconds=531&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" pod/monitoring-haproxy-0 condition met pod/monitoring-haproxy-1 condition met E0516 19:35:18.904349 1674 reflector.go:227] "Failed to watch" err="Get \"https://35.232.100.53/api/v1/namespaces/monitoring-2-0-4388/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dmonitoring-pxc-0&resourceVersion=1778960118630703021&timeoutSeconds=408&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" pod/monitoring-pxc-0 condition met E0516 19:35:19.205548 1674 reflector.go:227] "Failed to watch" err="Get \"https://35.232.100.53/api/v1/namespaces/monitoring-2-0-4388/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dmonitoring-pxc-1&resourceVersion=1778960118910175021&timeoutSeconds=476&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" pod/monitoring-pxc-1 condition met pod/monitoring-pxc-2 condition met ----------------------------------------------------------------------------------- check if pmm-client container enabled ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- compare statefulset/monitoring-pxc- ----------------------------------------------------------------------------------- [2026-05-16T19:35:47+0000] compare_kubectl: statefulset/monitoring-pxc OK ----------------------------------------------------------------------------------- compare statefulset/monitoring-haproxy- ----------------------------------------------------------------------------------- [2026-05-16T19:35:49+0000] compare_kubectl: statefulset/monitoring-haproxy OK ----------------------------------------------------------------------------------- verify clients agents statuses ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- check mysql metrics ----------------------------------------------------------------------------------- "1778954877" "1778954877" "156" "216" ----------------------------------------------------------------------------------- check haproxy metrics ----------------------------------------------------------------------------------- "0" "0" "1" "1" ----------------------------------------------------------------------------------- check QAN data ----------------------------------------------------------------------------------- null perconaxtradbcluster.pxc.percona.com/monitoring patched waiting for pod/monitoring-pxc-0 to be deleted...................Error from server (NotFound): pods "monitoring-pxc-0" not found release "monitoring" uninstalled ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + kubectl patch pxc -n monitoring-2-0-4388 monitoring --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/monitoring patched (no change) perconaxtradbcluster.pxc.percona.com "monitoring" deleted from monitoring-2-0-4388 namespace No resources found No resources found validatingwebhookconfiguration.admissionregistration.k8s.io "percona-xtradbcluster-webhook" deleted ----------------------------------------------------------------------------------- test passed -----------------------------------------------------------------------------------