Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2431/e2e-tests/logs/monitoring-2-0-8-0.log Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 No resources found + kubectl patch pxc -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: resource(s) were provided, but no name was specified No resources found No resources found No resources found error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- error: resource(s) were provided, but no name was specified Error from server (NotFound): namespaces "pxc-operator" not found waiting for namespace/pxc-operator to be deletedError from server (NotFound): namespaces "pxc-operator" not found ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- namespace/pxc-operator created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2431-1e1f76de-1-cluster2" modified. ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created pod/percona-xtradb-cluster-operator-6894467547-5lffg condition met pod/percona-xtradb-cluster-operator-6894467547-5lffg condition met waiting for pod/percona-xtradb-cluster-operator-6894467547-5lffg to become Ready.Ok error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces monitoring-2-0-7711 ----------------------------------------------------------------------------------- error: resource(s) were provided, but no name was specified Error from server (NotFound): namespaces "monitoring-2-0-7711" not found waiting for namespace/monitoring-2-0-7711 to be deletedError from server (NotFound): namespaces "monitoring-2-0-7711" not found ----------------------------------------------------------------------------------- create namespace monitoring-2-0-7711 ----------------------------------------------------------------------------------- namespace/monitoring-2-0-7711 created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2431-1e1f76de-1-cluster2" modified. ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- secret/minio-secret created secret/aws-s3-secret created secret/do-spaces-secret created secret/gcp-cs-secret created secret/azure-secret created "hashicorp" already exists with the same configuration, skipping "minio" already exists with the same configuration, skipping Hang tight while we grab the latest from your chart repositories... ...Successfully got an update from the "minio" chart repository ...Successfully got an update from the "chaos-mesh" chart repository ...Successfully got an update from the "hashicorp" chart repository ...Successfully got an update from the "percona" chart repository Update Complete. ⎈Happy Helming!⎈ ----------------------------------------------------------------------------------- install PMM Server ----------------------------------------------------------------------------------- "percona" already exists with the same configuration, skipping Hang tight while we grab the latest from your chart repositories... ...Successfully got an update from the "minio" chart repository ...Successfully got an update from the "chaos-mesh" chart repository ...Successfully got an update from the "hashicorp" chart repository ...Successfully got an update from the "percona" chart repository Update Complete. ⎈Happy Helming!⎈ Error: uninstall: Release not loaded: monitoring: release: not found NAME: monitoring LAST DEPLOYED: Thu Apr 9 02:43:33 2026 NAMESPACE: monitoring-2-0-7711 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: PMM server can be accessed via HTTPS (port 443) on the following DNS name from within your cluster: endpoint: https://monitoring-service.monitoring-2-0-7711.svc.cluster.local:443 login: admin password: admin statefulset.apps/monitoring condition met logger=settings t=2026-04-09T02:44:21.22578195Z level=info msg="Starting Grafana" version= commit= branch= compiled=1970-01-01T00:00:00Z logger=settings t=2026-04-09T02:44:21.22590311Z level=info msg="Config loaded from" file=/usr/share/grafana/conf/defaults.ini logger=settings t=2026-04-09T02:44:21.22591524Z level=info msg="Config loaded from" file=/etc/grafana/grafana.ini logger=settings t=2026-04-09T02:44:21.22592081Z level=info msg="Path Home" path=/usr/share/grafana logger=settings t=2026-04-09T02:44:21.22592565Z level=info msg="Path Data" path=/srv/grafana logger=settings t=2026-04-09T02:44:21.22593059Z level=info msg="Path Logs" path=/srv/logs logger=settings t=2026-04-09T02:44:21.2259352Z level=info msg="Path Plugins" path=/srv/grafana/plugins logger=settings t=2026-04-09T02:44:21.2259415Z level=info msg="Path Provisioning" path=/usr/share/grafana/conf/provisioning logger=settings t=2026-04-09T02:44:21.22594597Z level=info msg="App mode production" logger=sqlstore t=2026-04-09T02:44:21.22601819Z level=info msg="Connecting to DB" dbtype=postgres logger=migrator t=2026-04-09T02:44:21.2461004Z level=info msg="Starting DB migrations" logger=migrator t=2026-04-09T02:44:21.258392059Z level=info msg="migrations completed" performed=0 skipped=452 duration=582.73µs logger=secrets t=2026-04-09T02:44:21.260305169Z level=info msg="Envelope encryption state" enabled=true currentprovider=secretKey.v1 logger=plugin.finder t=2026-04-09T02:44:21.299820368Z level=warn msg="Skipping finding plugins as directory does not exist" path=/usr/share/grafana/plugins-bundled logger=plugin.signature.validator t=2026-04-09T02:44:21.415590096Z level=warn msg="Permitting unsigned plugin. This is not recommended" pluginID=pmm-pt-summary-panel pluginDir=/srv/grafana/plugins/pmm-app/dist/pmm-pt-summary/panel logger=plugin.signature.validator t=2026-04-09T02:44:21.415614436Z level=warn msg="Permitting unsigned plugin. This is not recommended" pluginID=pmm-qan-app-panel pluginDir=/srv/grafana/plugins/pmm-app/dist/pmm-qan logger=plugin.signature.validator t=2026-04-09T02:44:21.415635706Z level=warn msg="Permitting unsigned plugin. This is not recommended" pluginID=pmm-app pluginDir=/srv/grafana/plugins/pmm-app/dist logger=plugin.pmm-app t=2026-04-09T02:44:21.421082306Z level=warn msg="Included dashboard is missing a UID field" logger=plugin.signature.validator t=2026-04-09T02:44:21.421257626Z level=warn msg="Permitting unsigned plugin. This is not recommended" pluginID=grafana-polystat-panel pluginDir=/srv/grafana/plugins/grafana-polystat-panel logger=plugin.signature.validator t=2026-04-09T02:44:21.421281896Z level=warn msg="Permitting unsigned plugin. This is not recommended" pluginID=pmm-pt-summary-datasource pluginDir=/srv/grafana/plugins/pmm-app/dist/pmm-pt-summary/datasource logger=plugin.loader t=2026-04-09T02:44:21.421331176Z level=info msg="Plugin registered" pluginID=grafana-piechart-panel logger=plugin.loader t=2026-04-09T02:44:21.421341696Z level=info msg="Plugin registered" pluginID=jdbranham-diagram-panel logger=plugin.loader t=2026-04-09T02:44:21.421348136Z level=info msg="Plugin registered" pluginID=petrslavotinek-carpetplot-panel logger=plugin.loader t=2026-04-09T02:44:21.421353826Z level=info msg="Plugin registered" pluginID=pmm-pt-summary-panel logger=plugin.loader t=2026-04-09T02:44:21.421358806Z level=info msg="Plugin registered" pluginID=pmm-qan-app-panel logger=plugin.loader t=2026-04-09T02:44:21.421363936Z level=info msg="Plugin registered" pluginID=grafana-clickhouse-datasource logger=plugin.grafana-clickhouse-datasource t=2026-04-09T02:44:21.430871946Z level=warn msg="Plugin process is running with elevated privileges. This is not recommended" logger=plugin.loader t=2026-04-09T02:44:21.430886816Z level=info msg="Plugin registered" pluginID=pmm-app logger=plugin.loader t=2026-04-09T02:44:21.430891966Z level=info msg="Plugin registered" pluginID=camptocamp-prometheus-alertmanager-datasource logger=plugin.loader t=2026-04-09T02:44:21.430895806Z level=info msg="Plugin registered" pluginID=grafana-polystat-panel logger=plugin.loader t=2026-04-09T02:44:21.430901296Z level=info msg="Plugin registered" pluginID=grafana-worldmap-panel logger=plugin.loader t=2026-04-09T02:44:21.430907066Z level=info msg="Plugin registered" pluginID=pmm-pt-summary-datasource logger=plugin.loader t=2026-04-09T02:44:21.430913416Z level=info msg="Plugin registered" pluginID=natel-discrete-panel Admin password changed successfully ✔ ----------------------------------------------------------------------------------- create PXC cluster ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- secret/my-cluster-secrets created deployment.apps/pxc-client created perconaxtradbcluster.pxc.percona.com/monitoring created ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- pod/monitoring-haproxy-0 condition met pod/monitoring-pxc-0 condition met ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/monitoring-haproxy-0 condition met waiting for pod/monitoring-haproxy-0 to become Ready.Ok ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/monitoring-pxc-0 condition met waiting for pod/monitoring-pxc-0 to become Ready.Ok pod/monitoring-pxc-1 condition met waiting for pod/monitoring-pxc-1 to become Ready.Ok pod/monitoring-pxc-2 condition met waiting for pod/monitoring-pxc-2 to become Ready.Ok ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- pod/pxc-client-67fc4995bb-z2rgk condition met waiting for pod/pxc-client-67fc4995bb-z2rgk to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-67fc4995bb-z2rgk condition met waiting for pod/pxc-client-67fc4995bb-z2rgk to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-67fc4995bb-z2rgk condition met waiting for pod/pxc-client-67fc4995bb-z2rgk to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-67fc4995bb-z2rgk condition met waiting for pod/pxc-client-67fc4995bb-z2rgk to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-67fc4995bb-z2rgk condition met waiting for pod/pxc-client-67fc4995bb-z2rgk to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- add PMM API key to secret ----------------------------------------------------------------------------------- % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 36 0 0 100 36 0 81 --:--:-- --:--:-- --:--:-- 81 100 155 100 119 100 36 260 78 --:--:-- --:--:-- --:--:-- 338 secret/my-cluster-secrets patched Waiting for sts/monitoring-pxc to reach generation 2... Resource sts/monitoring-pxc is at generation 1. Waiting... Resource sts/monitoring-pxc has reached generation 2. Waiting for sts/monitoring-haproxy to reach generation 2... Resource sts/monitoring-haproxy has reached generation 2. pod/monitoring-haproxy-0 condition met pod/monitoring-haproxy-1 condition met pod/monitoring-pxc-0 condition met pod/monitoring-pxc-1 condition met pod/monitoring-pxc-2 condition met ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- waiting for pxc/monitoring to be ready................. ----------------------------------------------------------------------------------- compare statefulset/monitoring-pxc--no-prefix ----------------------------------------------------------------------------------- [2026-04-09T02:55:04+0000] compare_kubectl: statefulset/monitoring-pxc OK ----------------------------------------------------------------------------------- compare statefulset/monitoring-haproxy--no-prefix ----------------------------------------------------------------------------------- [2026-04-09T02:55:06+0000] compare_kubectl: statefulset/monitoring-haproxy OK ----------------------------------------------------------------------------------- apply my-env-var-secrets to add PMM_PREFIX ----------------------------------------------------------------------------------- secret/my-env-var-secrets created Waiting for sts/monitoring-pxc to reach generation 3... Resource sts/monitoring-pxc is at generation 2. Waiting... Resource sts/monitoring-pxc has reached generation 3. Waiting for sts/monitoring-haproxy to reach generation 3... Resource sts/monitoring-haproxy has reached generation 3. ----------------------------------------------------------------------------------- add new PMM API key to secret ----------------------------------------------------------------------------------- % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 167 100 127 100 40 277 87 --:--:-- --:--:-- --:--:-- 364 secret/my-cluster-secrets patched ----------------------------------------------------------------------------------- delete old PMM key ----------------------------------------------------------------------------------- % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 597 100 597 0 0 1327 0 --:--:-- --:--:-- --:--:-- 1329 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 29 100 29 0 0 65 0 --:--:-- --:--:-- --:--:-- 65 {"message":"API key deleted"}Waiting for sts/monitoring-pxc to reach generation 4... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc has reached generation 4. Waiting for sts/monitoring-haproxy to reach generation 4... Resource sts/monitoring-haproxy has reached generation 4. pod/monitoring-haproxy-0 condition met pod/monitoring-haproxy-1 condition met pod/monitoring-pxc-0 condition met pod/monitoring-pxc-1 condition met pod/monitoring-pxc-2 condition met ----------------------------------------------------------------------------------- check if pmm-client container enabled ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- compare statefulset/monitoring-pxc- ----------------------------------------------------------------------------------- [2026-04-09T02:59:05+0000] compare_kubectl: statefulset/monitoring-pxc OK ----------------------------------------------------------------------------------- compare statefulset/monitoring-haproxy- ----------------------------------------------------------------------------------- [2026-04-09T02:59:07+0000] compare_kubectl: statefulset/monitoring-haproxy OK ----------------------------------------------------------------------------------- verify clients agents statuses ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- check mysql metrics ----------------------------------------------------------------------------------- "1775697607" "1775697607" "158" "218" ----------------------------------------------------------------------------------- check haproxy metrics ----------------------------------------------------------------------------------- "0" "0" "1" "1" ----------------------------------------------------------------------------------- check QAN data ----------------------------------------------------------------------------------- null perconaxtradbcluster.pxc.percona.com/monitoring patched waiting for pod/monitoring-pxc-0 to be deleted...................Error from server (NotFound): pods "monitoring-pxc-0" not found release "monitoring" uninstalled ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + kubectl patch pxc -n monitoring-2-0-7711 monitoring --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/monitoring patched (no change) perconaxtradbcluster.pxc.percona.com "monitoring" deleted from monitoring-2-0-7711 namespace No resources found No resources found validatingwebhookconfiguration.admissionregistration.k8s.io "percona-xtradbcluster-webhook" deleted ----------------------------------------------------------------------------------- test passed -----------------------------------------------------------------------------------