Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-1948/e2e-tests/logs/monitoring-2-0-8-0.log WARNING: version difference between client (1.32) and server (1.28) exceeds the supported minor version skew of +/-1 WARNING: version difference between client (1.32) and server (1.28) exceeds the supported minor version skew of +/-1 No resources found + kubectl patch pxc -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: resource(s) were provided, but no name was specified No resources found No resources found No resources found error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- Error from server (NotFound): namespaces "pxc-operator" not found waiting for namespace/pxc-operator to be deletednamespace "gke-managed-system" deleted Error from server (NotFound): namespaces "pxc-operator" not found ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- namespace "gmp-public" deleted namespace "gmp-system" deleted Error from server (Forbidden): namespaces "default" is forbidden: this namespace may not be deleted namespace/pxc-operator created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-1948-70c11cfc-1-cluster9" modified. ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created pod/percona-xtradb-cluster-operator-5f5f9cbb4-vmrdt condition met pod/percona-xtradb-cluster-operator-5f5f9cbb4-vmrdt condition met waiting for pod/percona-xtradb-cluster-operator-5f5f9cbb4-vmrdt to become Ready.Ok error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces monitoring-2-0-31193 ----------------------------------------------------------------------------------- Error from server (NotFound): namespaces "monitoring-2-0-31193" not found waiting for namespace/monitoring-2-0-31193 to be deletedError from server (NotFound): namespaces "monitoring-2-0-31193" not found ----------------------------------------------------------------------------------- create namespace monitoring-2-0-31193 ----------------------------------------------------------------------------------- namespace "gke-managed-system" deleted namespace "gmp-public" deleted namespace/monitoring-2-0-31193 created namespace "gmp-system" deleted Error from server (Forbidden): namespaces "default" is forbidden: this namespace may not be deleted Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-1948-70c11cfc-1-cluster9" modified. ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created "hashicorp" already exists with the same configuration, skipping "minio" already exists with the same configuration, skipping Hang tight while we grab the latest from your chart repositories... ...Successfully got an update from the "chaos-mesh" chart repository ...Successfully got an update from the "minio" chart repository ...Successfully got an update from the "hashicorp" chart repository Update Complete. ⎈Happy Helming!⎈ ----------------------------------------------------------------------------------- install PMM Server ----------------------------------------------------------------------------------- "percona" has been added to your repositories Hang tight while we grab the latest from your chart repositories... ...Successfully got an update from the "minio" chart repository ...Successfully got an update from the "chaos-mesh" chart repository ...Successfully got an update from the "hashicorp" chart repository ...Successfully got an update from the "percona" chart repository Update Complete. ⎈Happy Helming!⎈ Error: uninstall: Release not loaded: monitoring: release: not found NAME: monitoring LAST DEPLOYED: Wed Jan 22 12:35:24 2025 NAMESPACE: monitoring-2-0-31193 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: PMM server can be accessed via HTTPS (port 443) on the following DNS name from within your cluster: endpoint: https://monitoring-service.monitoring-2-0-31193.svc.cluster.local:443 login: admin password: admin pod/monitoring-0 condition met command terminated with exit code 1 Retry 0 logger=settings t=2025-01-22T12:36:42.695375261Z level=info msg="Starting Grafana" version= commit= branch= compiled=1970-01-01T00:00:00Z logger=settings t=2025-01-22T12:36:42.695648303Z level=info msg="Config loaded from" file=/usr/share/grafana/conf/defaults.ini logger=settings t=2025-01-22T12:36:42.695678921Z level=info msg="Config loaded from" file=/etc/grafana/grafana.ini logger=settings t=2025-01-22T12:36:42.695687366Z level=info msg="Path Home" path=/usr/share/grafana logger=settings t=2025-01-22T12:36:42.695695694Z level=info msg="Path Data" path=/srv/grafana logger=settings t=2025-01-22T12:36:42.695703347Z level=info msg="Path Logs" path=/srv/logs logger=settings t=2025-01-22T12:36:42.69570905Z level=info msg="Path Plugins" path=/srv/grafana/plugins logger=settings t=2025-01-22T12:36:42.695715874Z level=info msg="Path Provisioning" path=/usr/share/grafana/conf/provisioning logger=settings t=2025-01-22T12:36:42.695722818Z level=info msg="App mode production" logger=sqlstore t=2025-01-22T12:36:42.695830219Z level=info msg="Connecting to DB" dbtype=postgres logger=migrator t=2025-01-22T12:36:42.718226407Z level=info msg="Starting DB migrations" logger=migrator t=2025-01-22T12:36:42.72508328Z level=info msg="migrations completed" performed=0 skipped=452 duration=803.798µs logger=secrets t=2025-01-22T12:36:42.727694969Z level=info msg="Envelope encryption state" enabled=true currentprovider=secretKey.v1 logger=plugin.finder t=2025-01-22T12:36:42.771324109Z level=warn msg="Skipping finding plugins as directory does not exist" path=/usr/share/grafana/plugins-bundled logger=plugin.signature.validator t=2025-01-22T12:36:43.033568917Z level=warn msg="Permitting unsigned plugin. This is not recommended" pluginID=pmm-pt-summary-datasource pluginDir=/srv/grafana/plugins/pmm-app/dist/pmm-pt-summary/datasource logger=plugin.signature.validator t=2025-01-22T12:36:43.033621171Z level=warn msg="Permitting unsigned plugin. This is not recommended" pluginID=pmm-qan-app-panel pluginDir=/srv/grafana/plugins/pmm-app/dist/pmm-qan logger=plugin.signature.validator t=2025-01-22T12:36:43.033642768Z level=warn msg="Permitting unsigned plugin. This is not recommended" pluginID=grafana-polystat-panel pluginDir=/srv/grafana/plugins/grafana-polystat-panel logger=plugin.signature.validator t=2025-01-22T12:36:43.033679073Z level=warn msg="Permitting unsigned plugin. This is not recommended" pluginID=pmm-app pluginDir=/srv/grafana/plugins/pmm-app/dist logger=plugin.pmm-app t=2025-01-22T12:36:43.043404452Z level=warn msg="Included dashboard is missing a UID field" logger=plugin.signature.validator t=2025-01-22T12:36:43.043723225Z level=warn msg="Permitting unsigned plugin. This is not recommended" pluginID=pmm-pt-summary-panel pluginDir=/srv/grafana/plugins/pmm-app/dist/pmm-pt-summary/panel logger=plugin.loader t=2025-01-22T12:36:43.043883547Z level=info msg="Plugin registered" pluginID=pmm-pt-summary-datasource logger=plugin.loader t=2025-01-22T12:36:43.043906799Z level=info msg="Plugin registered" pluginID=pmm-qan-app-panel logger=plugin.loader t=2025-01-22T12:36:43.043915679Z level=info msg="Plugin registered" pluginID=camptocamp-prometheus-alertmanager-datasource logger=plugin.loader t=2025-01-22T12:36:43.043923877Z level=info msg="Plugin registered" pluginID=grafana-polystat-panel logger=plugin.loader t=2025-01-22T12:36:43.043931439Z level=info msg="Plugin registered" pluginID=grafana-worldmap-panel logger=plugin.loader t=2025-01-22T12:36:43.043938429Z level=info msg="Plugin registered" pluginID=grafana-clickhouse-datasource logger=plugin.grafana-clickhouse-datasource t=2025-01-22T12:36:43.060781513Z level=warn msg="Plugin process is running with elevated privileges. This is not recommended" logger=plugin.loader t=2025-01-22T12:36:43.060821157Z level=info msg="Plugin registered" pluginID=jdbranham-diagram-panel logger=plugin.loader t=2025-01-22T12:36:43.060831319Z level=info msg="Plugin registered" pluginID=petrslavotinek-carpetplot-panel logger=plugin.loader t=2025-01-22T12:36:43.060837236Z level=info msg="Plugin registered" pluginID=pmm-app logger=plugin.loader t=2025-01-22T12:36:43.060845713Z level=info msg="Plugin registered" pluginID=pmm-pt-summary-panel logger=plugin.loader t=2025-01-22T12:36:43.060851399Z level=info msg="Plugin registered" pluginID=grafana-piechart-panel logger=plugin.loader t=2025-01-22T12:36:43.060856182Z level=info msg="Plugin registered" pluginID=natel-discrete-panel Admin password changed successfully ✔ ----------------------------------------------------------------------------------- create PXC cluster ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- secret/my-cluster-secrets created deployment.apps/pxc-client created perconaxtradbcluster.pxc.percona.com/monitoring created ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- error: no matching resources found ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- Error from server (NotFound): pods "monitoring-haproxy-0" not found waiting for pod/monitoring-haproxy-0 to become Ready........................................Ok ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/monitoring-pxc-0 condition met waiting for pod/monitoring-pxc-0 to become Ready.Ok pod/monitoring-pxc-1 condition met waiting for pod/monitoring-pxc-1 to become Ready.Ok pod/monitoring-pxc-2 condition met waiting for pod/monitoring-pxc-2 to become Ready.Ok ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- pod/pxc-client-69f9bd646c-qdzfz condition met waiting for pod/pxc-client-69f9bd646c-qdzfz to become Ready.Ok pod/pxc-client-69f9bd646c-qdzfz condition met waiting for pod/pxc-client-69f9bd646c-qdzfz to become Ready.Ok pod/pxc-client-69f9bd646c-qdzfz condition met waiting for pod/pxc-client-69f9bd646c-qdzfz to become Ready.Ok pod/pxc-client-69f9bd646c-qdzfz condition met waiting for pod/pxc-client-69f9bd646c-qdzfz to become Ready.Ok pod/pxc-client-69f9bd646c-qdzfz condition met waiting for pod/pxc-client-69f9bd646c-qdzfz to become Ready.Ok Unable to use a TTY - input is not a terminal or the right kind of file ----------------------------------------------------------------------------------- add PMM API key to secret ----------------------------------------------------------------------------------- % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 155 100 119 100 36 256 77 --:--:-- --:--:-- --:--:-- 334 secret/my-cluster-secrets patched Waiting for sts/monitoring-pxc to reach generation 2... Resource sts/monitoring-pxc is at generation 1. Waiting... Resource sts/monitoring-pxc has reached generation 2. Waiting for sts/monitoring-haproxy to reach generation 2... Resource sts/monitoring-haproxy has reached generation 2. pod/monitoring-haproxy-0 condition met pod/monitoring-haproxy-1 condition met pod/monitoring-pxc-0 condition met pod/monitoring-pxc-1 condition met pod/monitoring-pxc-2 condition met ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- waiting for cluster readyness waiting for cluster readyness waiting for cluster readyness waiting for cluster readyness waiting for cluster readyness ----------------------------------------------------------------------------------- compare statefulset/monitoring-pxc--no-prefix ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- compare statefulset/monitoring-haproxy--no-prefix ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- apply my-env-var-secrets to add PMM_PREFIX ----------------------------------------------------------------------------------- secret/my-env-var-secrets created Waiting for sts/monitoring-pxc to reach generation 3... Resource sts/monitoring-pxc is at generation 2. Waiting... Resource sts/monitoring-pxc has reached generation 3. Waiting for sts/monitoring-haproxy to reach generation 3... Resource sts/monitoring-haproxy has reached generation 3. ----------------------------------------------------------------------------------- add new PMM API key to secret ----------------------------------------------------------------------------------- % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 167 100 127 100 40 272 85 --:--:-- --:--:-- --:--:-- 359 secret/my-cluster-secrets patched ----------------------------------------------------------------------------------- delete old PMM key ----------------------------------------------------------------------------------- % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 602 100 602 0 0 1359 0 --:--:-- --:--:-- --:--:-- 1358 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 29 100 29 0 0 65 0 --:--:-- --:--:-- --:--:-- 65 {"message":"API key deleted"}Waiting for sts/monitoring-pxc to reach generation 4... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc is at generation 3. Waiting... Resource sts/monitoring-pxc has reached generation 4. Waiting for sts/monitoring-haproxy to reach generation 4... Resource sts/monitoring-haproxy has reached generation 4. pod/monitoring-haproxy-0 condition met pod/monitoring-haproxy-1 condition met pod/monitoring-pxc-0 condition met pod/monitoring-pxc-1 condition met pod/monitoring-pxc-2 condition met ----------------------------------------------------------------------------------- check if pmm-client container enabled ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- compare statefulset/monitoring-pxc- ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- compare statefulset/monitoring-haproxy- ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- verify clients agents statuses ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- check mysql metrics ----------------------------------------------------------------------------------- "1737547951" "1737547951" "165" "225" ----------------------------------------------------------------------------------- check haproxy metrics ----------------------------------------------------------------------------------- "0" "0" "1" "1" ----------------------------------------------------------------------------------- check QAN data ----------------------------------------------------------------------------------- null perconaxtradbcluster.pxc.percona.com/monitoring patched waiting for pod/monitoring-pxc-0 to be deleted.....................Error from server (NotFound): pods "monitoring-pxc-0" not found ----------------------------------------------------------------------------------- check for passwords leak ----------------------------------------------------------------------------------- secrets=WEo0K0hITFVQMDJJN0daTm14RWhESDJGT0liTW9sZ0pYU1d0cmw2Vw== K1N6aDhFYjBJWjBvMjV0SXZBei9yN1NzZG1EeWN0M09ENUNMMzIzblZvbFpZdlorWGV3ZFkxM2ZYRW9HM2pHc1o3L05uSWovdTl2ZStBU3R5TEtRa2c9PQo= VEtVS0N0dkFZRC9uR1p3dEYxa0hBOFFuTXpoNlFHMFlrUlZJK0ZLSQ== ZjErZUEzNTNvUFcsOTxiST8hNw== b3BlcmF0b3JhZG1pbg== YWRtaW4= ZXlKcklqb2lXbEY2VTNNeE5UTlNSbU53Y1dKWFduTTBhbXRWZUU1dGFqSmxUVFF3WmxVaUxDSnVJam9pYjNCbGNtRjBiM0l0Ym1WM0lpd2lhV1FpT2pGOQ== YWRtaW5fcGFzc3dvcmQ= cmVwbF9wYXNzd29yZA== cm9vdF9wYXNzd29yZA== YmFja3VwX3Bhc3N3b3Jk c29tZS1zZWNyZXQta2V5 ZjErZUEzNTNvUFcsOTxiST8hNw== b3BlcmF0b3JhZG1pbg== YWRtaW4= ZXlKcklqb2lXbEY2VTNNeE5UTlNSbU53Y1dKWFduTTBhbXRWZUU1dGFqSmxUVFF3WmxVaUxDSnVJam9pYjNCbGNtRjBiM0l0Ym1WM0lpd2lhV1FpT2pGOQ== YWRtaW5fcGFzc3dvcmQ= cmVwbF9wYXNzd29yZA== cm9vdF9wYXNzd29yZA== YmFja3VwX3Bhc3N3b3Jk cHhjLXByZWZpeC0= passwords=XJ4+HHLUP02I7GZNmxEhDH2FOIbMolgJXSWtrl6W +Szh8Eb0IZ0o25tIvAz/r7SsdmDyct3OD5CL323nVolZYvZ+XewdY13fXEoG3jGsZ7/NnIj/u9ve+AStyLKQkg== TKUKCtvAYD/nGZwtF1kHA8QnMzh6QG0YkRVI+FKI f1+eA353oPW,9