Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/logs/monitoring-2-0-8-0.log Warning: version difference between client (1.36) and server (1.33) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.36) and server (1.33) exceeds the supported minor version skew of +/-1 No resources found + kubectl patch pxc -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: resource(s) were provided, but no name was specified No resources found No resources found No resources found error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- Error from server (NotFound): namespaces "pxc-operator" not found waiting for namespace/pxc-operator to be deletederror: resource(s) were provided, but no name was specified Error from server (NotFound): namespaces "pxc-operator" not found ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- namespace/pxc-operator created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2473-6d392bea-4-cluster10" modified. ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created pod/percona-xtradb-cluster-operator-55d95dc9d8-kbptc condition met E0516 23:53:59.596775 9510 reflector.go:227] "Failed to watch" err="Get \"https://34.45.33.33/api/v1/namespaces/pxc-operator/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpercona-xtradb-cluster-operator-55d95dc9d8-kbptc&resourceVersion=1778975639235750000&timeoutSeconds=554&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" pod/percona-xtradb-cluster-operator-55d95dc9d8-kbptc condition met E0516 23:54:06.887122 10567 reflector.go:227] "Failed to watch" err="Get \"https://34.45.33.33/api/v1/namespaces/pxc-operator/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpercona-xtradb-cluster-operator-55d95dc9d8-kbptc&resourceVersion=1778975644220123000&timeoutSeconds=442&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/percona-xtradb-cluster-operator-55d95dc9d8-kbptc to become Ready.Ok error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces monitoring-2-0-29896 ----------------------------------------------------------------------------------- Error from server (NotFound): namespaces "monitoring-2-0-29896" not found waiting for namespace/monitoring-2-0-29896 to be deletederror: resource(s) were provided, but no name was specified Error from server (NotFound): namespaces "monitoring-2-0-29896" not found ----------------------------------------------------------------------------------- create namespace monitoring-2-0-29896 ----------------------------------------------------------------------------------- namespace/monitoring-2-0-29896 created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2473-6d392bea-4-cluster10" modified. ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- secret/minio-secret created secret/aws-s3-secret created secret/do-spaces-secret created secret/gcp-cs-secret created secret/azure-secret created "hashicorp" already exists with the same configuration, skipping "minio" already exists with the same configuration, skipping Hang tight while we grab the latest from your chart repositories... ...Successfully got an update from the "minio" chart repository ...Successfully got an update from the "chaos-mesh" chart repository ...Successfully got an update from the "percona" chart repository ...Successfully got an update from the "hashicorp" chart repository Update Complete. ⎈Happy Helming!⎈ ----------------------------------------------------------------------------------- install PMM Server ----------------------------------------------------------------------------------- "percona" already exists with the same configuration, skipping Hang tight while we grab the latest from your chart repositories... ...Successfully got an update from the "chaos-mesh" chart repository ...Successfully got an update from the "minio" chart repository ...Successfully got an update from the "hashicorp" chart repository ...Successfully got an update from the "percona" chart repository Update Complete. ⎈Happy Helming!⎈ Error: uninstall: Release not loaded: monitoring: release: not found NAME: monitoring LAST DEPLOYED: Sat May 16 23:54:46 2026 NAMESPACE: monitoring-2-0-29896 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: PMM server can be accessed via HTTPS (port 443) on the following DNS name from within your cluster: endpoint: https://monitoring-service.monitoring-2-0-29896.svc.cluster.local:443 login: admin password: admin statefulset.apps/monitoring condition met E0516 23:55:31.931848 22995 reflector.go:227] "Failed to watch" err="Get \"https://34.45.33.33/apis/apps/v1/namespaces/monitoring-2-0-29896/statefulsets?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dmonitoring&resourceVersion=1778975731190185000&timeoutSeconds=458&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" logger=settings t=2026-05-16T23:55:41.181709884Z level=info msg="Starting Grafana" version= commit= branch= compiled=1970-01-01T00:00:00Z logger=settings t=2026-05-16T23:55:41.181831364Z level=info msg="Config loaded from" file=/usr/share/grafana/conf/defaults.ini logger=settings t=2026-05-16T23:55:41.181842194Z level=info msg="Config loaded from" file=/etc/grafana/grafana.ini logger=settings t=2026-05-16T23:55:41.181847724Z level=info msg="Path Home" path=/usr/share/grafana logger=settings t=2026-05-16T23:55:41.181852564Z level=info msg="Path Data" path=/srv/grafana logger=settings t=2026-05-16T23:55:41.181857334Z level=info msg="Path Logs" path=/srv/logs logger=settings t=2026-05-16T23:55:41.181862164Z level=info msg="Path Plugins" path=/srv/grafana/plugins logger=settings t=2026-05-16T23:55:41.181866954Z level=info msg="Path Provisioning" path=/usr/share/grafana/conf/provisioning logger=settings t=2026-05-16T23:55:41.181871664Z level=info msg="App mode production" logger=sqlstore t=2026-05-16T23:55:41.181929464Z level=info msg="Connecting to DB" dbtype=postgres logger=migrator t=2026-05-16T23:55:41.194011165Z level=info msg="Starting DB migrations" logger=migrator t=2026-05-16T23:55:41.198486191Z level=info msg="migrations completed" performed=0 skipped=452 duration=1.00985ms logger=secrets t=2026-05-16T23:55:41.2000095Z level=info msg="Envelope encryption state" enabled=true currentprovider=secretKey.v1 logger=plugin.finder t=2026-05-16T23:55:41.22729844Z level=warn msg="Skipping finding plugins as directory does not exist" path=/usr/share/grafana/plugins-bundled logger=plugin.signature.validator t=2026-05-16T23:55:41.344255991Z level=warn msg="Permitting unsigned plugin. This is not recommended" pluginID=grafana-polystat-panel pluginDir=/srv/grafana/plugins/grafana-polystat-panel logger=plugin.signature.validator t=2026-05-16T23:55:41.344282561Z level=warn msg="Permitting unsigned plugin. This is not recommended" pluginID=pmm-app pluginDir=/srv/grafana/plugins/pmm-app/dist logger=plugin.pmm-app t=2026-05-16T23:55:41.349658727Z level=warn msg="Included dashboard is missing a UID field" logger=plugin.signature.validator t=2026-05-16T23:55:41.349877657Z level=warn msg="Permitting unsigned plugin. This is not recommended" pluginID=pmm-pt-summary-panel pluginDir=/srv/grafana/plugins/pmm-app/dist/pmm-pt-summary/panel logger=plugin.signature.validator t=2026-05-16T23:55:41.349938647Z level=warn msg="Permitting unsigned plugin. This is not recommended" pluginID=pmm-pt-summary-datasource pluginDir=/srv/grafana/plugins/pmm-app/dist/pmm-pt-summary/datasource logger=plugin.signature.validator t=2026-05-16T23:55:41.349959307Z level=warn msg="Permitting unsigned plugin. This is not recommended" pluginID=pmm-qan-app-panel pluginDir=/srv/grafana/plugins/pmm-app/dist/pmm-qan logger=plugin.loader t=2026-05-16T23:55:41.350046307Z level=info msg="Plugin registered" pluginID=jdbranham-diagram-panel logger=plugin.loader t=2026-05-16T23:55:41.350060317Z level=info msg="Plugin registered" pluginID=petrslavotinek-carpetplot-panel logger=plugin.loader t=2026-05-16T23:55:41.350067117Z level=info msg="Plugin registered" pluginID=grafana-polystat-panel logger=plugin.loader t=2026-05-16T23:55:41.350072947Z level=info msg="Plugin registered" pluginID=pmm-app logger=plugin.loader t=2026-05-16T23:55:41.350079207Z level=info msg="Plugin registered" pluginID=camptocamp-prometheus-alertmanager-datasource logger=plugin.loader t=2026-05-16T23:55:41.350084587Z level=info msg="Plugin registered" pluginID=grafana-worldmap-panel logger=plugin.loader t=2026-05-16T23:55:41.350088487Z level=info msg="Plugin registered" pluginID=pmm-pt-summary-panel logger=plugin.loader t=2026-05-16T23:55:41.350092067Z level=info msg="Plugin registered" pluginID=grafana-clickhouse-datasource logger=plugin.grafana-clickhouse-datasource t=2026-05-16T23:55:41.361437168Z level=warn msg="Plugin process is running with elevated privileges. This is not recommended" logger=plugin.loader t=2026-05-16T23:55:41.361685998Z level=info msg="Plugin registered" pluginID=natel-discrete-panel logger=plugin.loader t=2026-05-16T23:55:41.361839328Z level=info msg="Plugin registered" pluginID=pmm-pt-summary-datasource logger=plugin.loader t=2026-05-16T23:55:41.361997128Z level=info msg="Plugin registered" pluginID=pmm-qan-app-panel logger=plugin.loader t=2026-05-16T23:55:41.362009558Z level=info msg="Plugin registered" pluginID=grafana-piechart-panel Admin password changed successfully ✔ ----------------------------------------------------------------------------------- create PXC cluster ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- secret/my-cluster-secrets created deployment.apps/pxc-client created perconaxtradbcluster.pxc.percona.com/monitoring created ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- pod/monitoring-haproxy-0 condition met pod/monitoring-pxc-0 condition met E0516 23:56:55.227841 26325 reflector.go:227] "Failed to watch" err="Get \"https://34.45.33.33/api/v1/namespaces/monitoring-2-0-29896/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dmonitoring-pxc-0&resourceVersion=1778975815014959010&timeoutSeconds=378&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/monitoring-haproxy-0 condition met E0516 23:56:57.585175 2538 reflector.go:227] "Failed to watch" err="Get \"https://34.45.33.33/api/v1/namespaces/monitoring-2-0-29896/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dmonitoring-haproxy-0&resourceVersion=1778975815859295010&timeoutSeconds=528&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/monitoring-haproxy-0 to become Ready.Ok ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/monitoring-pxc-0 condition met E0516 23:57:06.846122 3810 reflector.go:227] "Failed to watch" err="Get \"https://34.45.33.33/api/v1/namespaces/monitoring-2-0-29896/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dmonitoring-pxc-0&resourceVersion=1778975822881743010&timeoutSeconds=533&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/monitoring-pxc-0 to become Ready.Ok pod/monitoring-pxc-1 condition met waiting for pod/monitoring-pxc-1 to become Ready.Ok pod/monitoring-pxc-2 condition met waiting for pod/monitoring-pxc-2 to become Ready.Ok ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- pod/pxc-client-67fc4995bb-5grkm condition met E0517 00:01:48.859966 31176 reflector.go:227] "Failed to watch" err="Get \"https://34.45.33.33/api/v1/namespaces/monitoring-2-0-29896/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-5grkm&resourceVersion=1778976107011742000&timeoutSeconds=586&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-5grkm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-67fc4995bb-5grkm condition met E0517 00:02:01.630896 441 reflector.go:227] "Failed to watch" err="Get \"https://34.45.33.33/api/v1/namespaces/monitoring-2-0-29896/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-5grkm&resourceVersion=1778976118828170000&timeoutSeconds=453&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-5grkm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-67fc4995bb-5grkm condition met E0517 00:02:42.927895 6367 reflector.go:227] "Failed to watch" err="Get \"https://34.45.33.33/api/v1/namespaces/monitoring-2-0-29896/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-5grkm&resourceVersion=1778976160361459000&timeoutSeconds=441&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-5grkm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-67fc4995bb-5grkm condition met E0517 00:02:54.825990 8085 reflector.go:227] "Failed to watch" err="Get \"https://34.45.33.33/api/v1/namespaces/monitoring-2-0-29896/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-5grkm&resourceVersion=1778976171972333000&timeoutSeconds=488&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-5grkm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-67fc4995bb-5grkm condition met E0517 00:03:07.009515 9800 reflector.go:227] "Failed to watch" err="Get \"https://34.45.33.33/api/v1/namespaces/monitoring-2-0-29896/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-5grkm&resourceVersion=1778976184897583000&timeoutSeconds=379&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-5grkm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- add PMM API key to secret ----------------------------------------------------------------------------------- % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 155 100 119 100 36 266 80 --:--:-- --:--:-- --:--:-- 346 secret/my-cluster-secrets patched Waiting for sts/monitoring-pxc to reach generation 2... Resource sts/monitoring-pxc is at generation 1. Waiting... Resource sts/monitoring-pxc has reached generation 2. Waiting for sts/monitoring-haproxy to reach generation 2... Resource sts/monitoring-haproxy has reached generation 2. E0517 00:03:37.938199 13716 reflector.go:227] "Failed to watch" err="Get \"https://34.45.33.33/api/v1/namespaces/monitoring-2-0-29896/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dmonitoring-haproxy-0&resourceVersion=1778976217248595000&timeoutSeconds=489&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" pod/monitoring-haproxy-0 condition met pod/monitoring-haproxy-1 condition met E0517 00:03:49.019210 13716 reflector.go:227] "Failed to watch" err="Get \"https://34.45.33.33/api/v1/namespaces/monitoring-2-0-29896/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dmonitoring-pxc-0&resourceVersion=1778976228704015000&timeoutSeconds=386&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" pod/monitoring-pxc-0 condition met E0517 00:03:49.320362 13716 reflector.go:227] "Failed to watch" err="Get \"https://34.45.33.33/api/v1/namespaces/monitoring-2-0-29896/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dmonitoring-pxc-1&resourceVersion=1778976229027791000&timeoutSeconds=436&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" pod/monitoring-pxc-1 condition met pod/monitoring-pxc-2 condition met ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- waiting for pxc/monitoring to be ready.................. ----------------------------------------------------------------------------------- compare statefulset/monitoring-pxc--no-prefix ----------------------------------------------------------------------------------- [2026-05-17T00:06:29+0000] compare_kubectl: statefulset/monitoring-pxc OK ----------------------------------------------------------------------------------- compare statefulset/monitoring-haproxy--no-prefix ----------------------------------------------------------------------------------- [2026-05-17T00:06:32+0000] compare_kubectl: statefulset/monitoring-haproxy OK ----------------------------------------------------------------------------------- apply my-env-var-secrets to add PMM_PREFIX ----------------------------------------------------------------------------------- secret/my-env-var-secrets created Waiting for sts/monitoring-pxc to reach generation 3... Resource sts/monitoring-pxc has reached generation 3. Waiting for sts/monitoring-haproxy to reach generation 3... Resource sts/monitoring-haproxy has reached generation 3. ----------------------------------------------------------------------------------- add new PMM API key to secret ----------------------------------------------------------------------------------- % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 40 0 0 100 40 0 93 --:--:-- --:--:-- --:--:-- 93 100 167 100 127 100 40 286 90 --:--:-- --:--:-- --:--:-- 376 secret/my-cluster-secrets patched ----------------------------------------------------------------------------------- delete old PMM key ----------------------------------------------------------------------------------- % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 602 100 602 0 0 1398 0 --:--:-- --:--:-- --:--:-- 1400 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 29 100 29 0 0 64 0 --:--:-- --:--:-- --:--:-- 64 {"message":"API key deleted"}Waiting for sts/monitoring-pxc to reach generation 4... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc has reached generation 4. Waiting for sts/monitoring-haproxy to reach generation 4... Resource sts/monitoring-haproxy has reached generation 4. E0517 00:09:34.336091 21030 reflector.go:227] "Failed to watch" err="Get \"https://34.45.33.33/api/v1/namespaces/monitoring-2-0-29896/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dmonitoring-haproxy-0&resourceVersion=1778976573643218000&timeoutSeconds=469&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" pod/monitoring-haproxy-0 condition met pod/monitoring-haproxy-1 condition met E0517 00:09:42.648589 21030 reflector.go:227] "Failed to watch" err="Get \"https://34.45.33.33/api/v1/namespaces/monitoring-2-0-29896/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dmonitoring-pxc-0&resourceVersion=1778976582335327000&timeoutSeconds=524&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" pod/monitoring-pxc-0 condition met E0517 00:09:42.950094 21030 reflector.go:227] "Failed to watch" err="Get \"https://34.45.33.33/api/v1/namespaces/monitoring-2-0-29896/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dmonitoring-pxc-1&resourceVersion=1778976582335327000&timeoutSeconds=308&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" pod/monitoring-pxc-1 condition met pod/monitoring-pxc-2 condition met ----------------------------------------------------------------------------------- check if pmm-client container enabled ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- compare statefulset/monitoring-pxc- ----------------------------------------------------------------------------------- [2026-05-17T00:10:09+0000] compare_kubectl: statefulset/monitoring-pxc OK ----------------------------------------------------------------------------------- compare statefulset/monitoring-haproxy- ----------------------------------------------------------------------------------- [2026-05-17T00:10:11+0000] compare_kubectl: statefulset/monitoring-haproxy OK ----------------------------------------------------------------------------------- verify clients agents statuses ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- check mysql metrics ----------------------------------------------------------------------------------- "1778970328" "1778970328" "156" "216" ----------------------------------------------------------------------------------- check haproxy metrics ----------------------------------------------------------------------------------- "0" "0" "1" "1" ----------------------------------------------------------------------------------- check QAN data ----------------------------------------------------------------------------------- null perconaxtradbcluster.pxc.percona.com/monitoring patched waiting for pod/monitoring-pxc-0 to be deleted...............Error from server (NotFound): pods "monitoring-pxc-0" not found release "monitoring" uninstalled ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + kubectl patch pxc -n monitoring-2-0-29896 monitoring --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/monitoring patched (no change) perconaxtradbcluster.pxc.percona.com "monitoring" deleted from monitoring-2-0-29896 namespace No resources found No resources found validatingwebhookconfiguration.admissionregistration.k8s.io "percona-xtradbcluster-webhook" deleted ----------------------------------------------------------------------------------- test passed -----------------------------------------------------------------------------------