=== RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://35.188.195.213 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 180 seconds for each step harness.go:372: testsuite: e2e-tests/tests has 8 tests === RUN kuttl/harness === RUN kuttl/harness/monitoring === PAUSE kuttl/harness/monitoring === CONT kuttl/harness/monitoring logger.go:42: 15:56:54 | monitoring | Creating namespace: kuttl-test-internal-boxer logger.go:42: 15:56:54 | monitoring/0-deploy-operator | starting test step 0-deploy-operator logger.go:42: 15:56:54 | monitoring/0-deploy-operator | running command: [sh -c set -o errexit set -o xtrace source ../../functions deploy_operator deploy_client] logger.go:42: 15:56:54 | monitoring/0-deploy-operator | + source ../../functions logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ realpath ../../.. logger.go:42: 15:56:54 | monitoring/0-deploy-operator | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-687 logger.go:42: 15:56:54 | monitoring/0-deploy-operator | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-687/e2e-tests/vars.sh logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-687 logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-687 logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-687/deploy logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-687/deploy logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-687/e2e-tests logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-687/e2e-tests logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-687/e2e-tests/conf logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-687/e2e-tests/conf logger.go:42: 15:56:54 | monitoring/0-deploy-operator | ++++ mktemp -d logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ export TEMP_DIR=/tmp/tmp.hR99v4wV1O logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ TEMP_DIR=/tmp/tmp.hR99v4wV1O logger.go:42: 15:56:54 | monitoring/0-deploy-operator | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ export GIT_BRANCH=PR-687 logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ GIT_BRANCH=PR-687 logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ export VERSION=PR-687-895efcafe logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ VERSION=PR-687-895efcafe logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-687-895efcafe logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ IMAGE=perconalab/percona-postgresql-operator:PR-687-895efcafe logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ export PG_VER=15 logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ PG_VER=15 logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ export IMAGE_PGBOUNCER=percona/percona-postgresql-operator:2.2.0-ppg15-pgbouncer logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ IMAGE_PGBOUNCER=percona/percona-postgresql-operator:2.2.0-ppg15-pgbouncer logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ export IMAGE_POSTGRESQL=percona/percona-postgresql-operator:2.2.0-ppg15-postgres logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ IMAGE_POSTGRESQL=percona/percona-postgresql-operator:2.2.0-ppg15-postgres logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ export IMAGE_BACKREST=percona/percona-postgresql-operator:2.2.0-ppg15-pgbackrest logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ IMAGE_BACKREST=percona/percona-postgresql-operator:2.2.0-ppg15-pgbackrest logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ export IMAGE_PGBADGER=perconalab/percona-postgresql-operator:main-ppg15-pgbadger logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ IMAGE_PGBADGER=perconalab/percona-postgresql-operator:main-ppg15-pgbadger logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ export BUCKET=pg-operator-testing logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ BUCKET=pg-operator-testing logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 15:56:54 | monitoring/0-deploy-operator | ++++ which gdate logger.go:42: 15:56:54 | monitoring/0-deploy-operator | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-687/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 15:56:54 | monitoring/0-deploy-operator | ++++ which date logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ date=/usr/bin/date logger.go:42: 15:56:54 | monitoring/0-deploy-operator | ++++ which gsed logger.go:42: 15:56:54 | monitoring/0-deploy-operator | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-687/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 15:56:54 | monitoring/0-deploy-operator | ++++ which sed logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ sed=/usr/bin/sed logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ command -v oc logger.go:42: 15:56:54 | monitoring/0-deploy-operator | ++++ pwd logger.go:42: 15:56:54 | monitoring/0-deploy-operator | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-687/e2e-tests/tests/monitoring logger.go:42: 15:56:54 | monitoring/0-deploy-operator | ++ test_name=monitoring logger.go:42: 15:56:54 | monitoring/0-deploy-operator | + deploy_operator logger.go:42: 15:56:54 | monitoring/0-deploy-operator | + '[' -n pg-operator ']' logger.go:42: 15:56:54 | monitoring/0-deploy-operator | + create_namespace pg-operator logger.go:42: 15:56:54 | monitoring/0-deploy-operator | + local namespace=pg-operator logger.go:42: 15:56:54 | monitoring/0-deploy-operator | + [[ 1 == 1 ]] logger.go:42: 15:56:54 | monitoring/0-deploy-operator | + kubectl delete namespace pg-operator --ignore-not-found logger.go:42: 15:56:55 | monitoring/0-deploy-operator | namespace "pg-operator" deleted logger.go:42: 15:57:06 | monitoring/0-deploy-operator | + kubectl wait --for=delete namespace pg-operator logger.go:42: 15:57:06 | monitoring/0-deploy-operator | + kubectl create namespace pg-operator logger.go:42: 15:57:07 | monitoring/0-deploy-operator | namespace/pg-operator created logger.go:42: 15:57:07 | monitoring/0-deploy-operator | + kubectl -n pg-operator apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pg-operator_PR-687/deploy/crd.yaml logger.go:42: 15:57:07 | monitoring/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconapgbackups.pgv2.percona.com serverside-applied logger.go:42: 15:57:09 | monitoring/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconapgclusters.pgv2.percona.com serverside-applied logger.go:42: 15:57:09 | monitoring/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconapgrestores.pgv2.percona.com serverside-applied logger.go:42: 15:57:10 | monitoring/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/postgresclusters.postgres-operator.crunchydata.com serverside-applied logger.go:42: 15:57:10 | monitoring/0-deploy-operator | + kubectl -n pg-operator apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pg-operator_PR-687/deploy/cw-rbac.yaml logger.go:42: 15:57:11 | monitoring/0-deploy-operator | serviceaccount/percona-postgresql-operator serverside-applied logger.go:42: 15:57:11 | monitoring/0-deploy-operator | clusterrole.rbac.authorization.k8s.io/percona-postgresql-operator serverside-applied logger.go:42: 15:57:11 | monitoring/0-deploy-operator | clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-postgresql-operator serverside-applied logger.go:42: 15:57:11 | monitoring/0-deploy-operator | + kubectl -n pg-operator apply -f - logger.go:42: 15:57:11 | monitoring/0-deploy-operator | ++ printf 'select(documentIndex==0).spec.template.spec.containers[0].image="%s"' perconalab/percona-postgresql-operator:PR-687-895efcafe logger.go:42: 15:57:11 | monitoring/0-deploy-operator | + yq eval 'select(documentIndex==0).spec.template.spec.containers[0].image="perconalab/percona-postgresql-operator:PR-687-895efcafe"' /mnt/jenkins/workspace/cloud-pg-operator_PR-687/deploy/cw-operator.yaml logger.go:42: 15:57:12 | monitoring/0-deploy-operator | deployment.apps/percona-postgresql-operator created logger.go:42: 15:57:12 | monitoring/0-deploy-operator | + deploy_client logger.go:42: 15:57:12 | monitoring/0-deploy-operator | + kubectl -n kuttl-test-internal-boxer apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-687/e2e-tests/conf/client.yaml logger.go:42: 15:57:13 | monitoring/0-deploy-operator | deployment.apps/pg-client created logger.go:42: 15:57:14 | monitoring/0-deploy-operator | running command: [sh -c kubectl assert exist-enhanced deployment percona-postgresql-operator -n ${OPERATOR_NS:-$NAMESPACE} --field-selector status.readyReplicas=1] logger.go:42: 15:57:14 | monitoring/0-deploy-operator | ASSERT deployment percona-postgresql-operator matching field criteria 'status.readyReplicas=1' should exist. logger.go:42: 15:57:14 | monitoring/0-deploy-operator | INFO Found 1 resource(s). logger.go:42: 15:57:14 | monitoring/0-deploy-operator | NAME NAMESPACE COL0 logger.go:42: 15:57:14 | monitoring/0-deploy-operator | percona-postgresql-operator pg-operator 1 logger.go:42: 15:57:14 | monitoring/0-deploy-operator | ASSERT PASS logger.go:42: 15:57:14 | monitoring/0-deploy-operator | test step completed 0-deploy-operator logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | starting test step 1-deploy-pmm-server logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | running command: [sh -c set -o errexit set -o xtrace source ../../functions deploy_pmm_server] logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | + source ../../functions logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ realpath ../../.. logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-687 logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-687/e2e-tests/vars.sh logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-687 logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-687 logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-687/deploy logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-687/deploy logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-687/e2e-tests logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-687/e2e-tests logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-687/e2e-tests/conf logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-687/e2e-tests/conf logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | ++++ mktemp -d logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ export TEMP_DIR=/tmp/tmp.19qEOvRg7j logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ TEMP_DIR=/tmp/tmp.19qEOvRg7j logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ export GIT_BRANCH=PR-687 logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ GIT_BRANCH=PR-687 logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ export VERSION=PR-687-895efcafe logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ VERSION=PR-687-895efcafe logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-687-895efcafe logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ IMAGE=perconalab/percona-postgresql-operator:PR-687-895efcafe logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ export PG_VER=15 logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ PG_VER=15 logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ export IMAGE_PGBOUNCER=percona/percona-postgresql-operator:2.2.0-ppg15-pgbouncer logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ IMAGE_PGBOUNCER=percona/percona-postgresql-operator:2.2.0-ppg15-pgbouncer logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ export IMAGE_POSTGRESQL=percona/percona-postgresql-operator:2.2.0-ppg15-postgres logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ IMAGE_POSTGRESQL=percona/percona-postgresql-operator:2.2.0-ppg15-postgres logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ export IMAGE_BACKREST=percona/percona-postgresql-operator:2.2.0-ppg15-pgbackrest logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ IMAGE_BACKREST=percona/percona-postgresql-operator:2.2.0-ppg15-pgbackrest logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ export IMAGE_PGBADGER=perconalab/percona-postgresql-operator:main-ppg15-pgbadger logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ IMAGE_PGBADGER=perconalab/percona-postgresql-operator:main-ppg15-pgbadger logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ export BUCKET=pg-operator-testing logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ BUCKET=pg-operator-testing logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ export IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ IMAGE_PMM=perconalab/pmm-client:dev-latest logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ export IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ IMAGE_PMM_SERVER_REPO=perconalab/pmm-server logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ export IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ IMAGE_PMM_SERVER_TAG=dev-latest logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | ++++ which gdate logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-687/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | ++++ which date logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ date=/usr/bin/date logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | ++++ which gsed logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-687/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | ++++ which sed logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ sed=/usr/bin/sed logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ command -v oc logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | ++++ pwd logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-687/e2e-tests/tests/monitoring logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | ++ test_name=monitoring logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | + deploy_pmm_server logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | + local platform=kubernetes logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | + helm uninstall -n kuttl-test-internal-boxer pmm logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-pg-operator_PR-687/kubeconfig logger.go:42: 15:57:14 | monitoring/1-deploy-pmm-server | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-pg-operator_PR-687/kubeconfig logger.go:42: 15:57:15 | monitoring/1-deploy-pmm-server | Error: uninstall: Release not loaded: pmm: release: not found logger.go:42: 15:57:15 | monitoring/1-deploy-pmm-server | + : logger.go:42: 15:57:15 | monitoring/1-deploy-pmm-server | + helm install monitoring -n kuttl-test-internal-boxer --set imageTag=dev-latest --set service.type=LoadBalancer --set imageRepo=perconalab/pmm-server --set platform=kubernetes https://percona-charts.storage.googleapis.com/pmm-server-9.9.9.tgz logger.go:42: 15:57:15 | monitoring/1-deploy-pmm-server | WARNING: Kubernetes configuration file is group-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-pg-operator_PR-687/kubeconfig logger.go:42: 15:57:15 | monitoring/1-deploy-pmm-server | WARNING: Kubernetes configuration file is world-readable. This is insecure. Location: /mnt/jenkins/workspace/cloud-pg-operator_PR-687/kubeconfig logger.go:42: 15:57:19 | monitoring/1-deploy-pmm-server | NAME: monitoring logger.go:42: 15:57:19 | monitoring/1-deploy-pmm-server | LAST DEPLOYED: Tue Mar 5 15:57:16 2024 logger.go:42: 15:57:19 | monitoring/1-deploy-pmm-server | NAMESPACE: kuttl-test-internal-boxer logger.go:42: 15:57:19 | monitoring/1-deploy-pmm-server | STATUS: deployed logger.go:42: 15:57:19 | monitoring/1-deploy-pmm-server | REVISION: 1 logger.go:42: 15:57:19 | monitoring/1-deploy-pmm-server | TEST SUITE: None logger.go:42: 15:57:19 | monitoring/1-deploy-pmm-server | NOTES: logger.go:42: 15:57:19 | monitoring/1-deploy-pmm-server | PMM server can be accessed via HTTPS (port 443) on the following DNS name from within your cluster: logger.go:42: 15:57:19 | monitoring/1-deploy-pmm-server | logger.go:42: 15:57:19 | monitoring/1-deploy-pmm-server | endpoint: https://monitoring-service.kuttl-test-internal-boxer.svc.cluster.local:443 logger.go:42: 15:57:19 | monitoring/1-deploy-pmm-server | login: admin logger.go:42: 15:57:19 | monitoring/1-deploy-pmm-server | password: admin logger.go:42: 15:59:21 | monitoring/1-deploy-pmm-server | test step failed 1-deploy-pmm-server case.go:364: failed in step 1-deploy-pmm-server case.go:366: --- Event:kuttl-test-internal-boxer/ +++ Event:kuttl-test-internal-boxer/monitoring-0.17b9e979e455f55b @@ -1,17 +1,41 @@ apiVersion: v1 count: 1 +eventTime: null +firstTimestamp: "2024-03-05T15:57:23Z" involvedObject: apiVersion: v1 - kind: Service - name: monitoring-service + kind: Pod + name: monitoring-0 + namespace: kuttl-test-internal-boxer + resourceVersion: "3836" + uid: dacbfee3-eac4-4ca3-ac54-8b80e9c8a721 kind: Event -message: Ensured load balancer +lastTimestamp: "2024-03-05T15:57:23Z" +message: Successfully assigned kuttl-test-internal-boxer/monitoring-0 to gke-jen-pg-687-895efcafe-default-pool-180f5d1b-q1sb metadata: + managedFields: + - apiVersion: v1 + fieldsType: FieldsV1 + fieldsV1: + f:count: {} + f:firstTimestamp: {} + f:involvedObject: {} + f:lastTimestamp: {} + f:message: {} + f:reason: {} + f:reportingComponent: {} + f:source: + f:component: {} + f:type: {} + manager: kube-scheduler + operation: Update + time: "2024-03-05T15:57:23Z" + name: monitoring-0.17b9e979e455f55b namespace: kuttl-test-internal-boxer -reason: EnsuredLoadBalancer -reportingComponent: "" +reason: Scheduled +reportingComponent: default-scheduler reportingInstance: "" source: - component: service-controller + component: default-scheduler type: Normal case.go:366: resource Event:kuttl-test-internal-boxer/: .message: value mismatch, expected: Ensured load balancer != actual: Successfully assigned kuttl-test-internal-boxer/monitoring-0 to gke-jen-pg-687-895efcafe-default-pool-180f5d1b-q1sb case.go:366: --- Event:kuttl-test-internal-boxer/ +++ Event:kuttl-test-internal-boxer/monitoring-0.17b9e97ba756bc7a @@ -1,17 +1,41 @@ apiVersion: v1 count: 1 +eventTime: null +firstTimestamp: "2024-03-05T15:57:30Z" involvedObject: apiVersion: v1 - kind: Service - name: monitoring-service + kind: Pod + name: monitoring-0 + namespace: kuttl-test-internal-boxer + resourceVersion: "3877" + uid: dacbfee3-eac4-4ca3-ac54-8b80e9c8a721 kind: Event -message: Ensured load balancer +lastTimestamp: "2024-03-05T15:57:30Z" +message: 'AttachVolume.Attach succeeded for volume "pvc-bd15bf1b-fe46-4802-942f-532ea3704dd3" ' metadata: + managedFields: + - apiVersion: v1 + fieldsType: FieldsV1 + fieldsV1: + f:count: {} + f:firstTimestamp: {} + f:involvedObject: {} + f:lastTimestamp: {} + f:message: {} + f:reason: {} + f:reportingComponent: {} + f:source: + f:component: {} + f:type: {} + manager: kube-controller-manager + operation: Update + time: "2024-03-05T15:57:30Z" + name: monitoring-0.17b9e97ba756bc7a namespace: kuttl-test-internal-boxer -reason: EnsuredLoadBalancer -reportingComponent: "" +reason: SuccessfulAttachVolume +reportingComponent: attachdetach-controller reportingInstance: "" source: - component: service-controller + component: attachdetach-controller type: Normal case.go:366: resource Event:kuttl-test-internal-boxer/: .reason: value mismatch, expected: EnsuredLoadBalancer != actual: SuccessfulAttachVolume case.go:366: --- Event:kuttl-test-internal-boxer/ +++ Event:kuttl-test-internal-boxer/monitoring-0.17b9e97c0507a433 @@ -1,17 +1,45 @@ apiVersion: v1 count: 1 +eventTime: null +firstTimestamp: "2024-03-05T15:57:32Z" involvedObject: apiVersion: v1 - kind: Service - name: monitoring-service + fieldPath: spec.containers{monitoring} + kind: Pod + name: monitoring-0 + namespace: kuttl-test-internal-boxer + resourceVersion: "3877" + uid: dacbfee3-eac4-4ca3-ac54-8b80e9c8a721 kind: Event -message: Ensured load balancer +lastTimestamp: "2024-03-05T15:57:32Z" +message: Pulling image "perconalab/pmm-server:dev-latest" metadata: + managedFields: + - apiVersion: v1 + fieldsType: FieldsV1 + fieldsV1: + f:count: {} + f:firstTimestamp: {} + f:involvedObject: {} + f:lastTimestamp: {} + f:message: {} + f:reason: {} + f:reportingComponent: {} + f:reportingInstance: {} + f:source: + f:component: {} + f:host: {} + f:type: {} + manager: kubelet + operation: Update + time: "2024-03-05T15:57:32Z" + name: monitoring-0.17b9e97c0507a433 namespace: kuttl-test-internal-boxer -reason: EnsuredLoadBalancer -reportingComponent: "" -reportingInstance: "" +reason: Pulling +reportingComponent: kubelet +reportingInstance: gke-jen-pg-687-895efcafe-default-pool-180f5d1b-q1sb source: - component: service-controller + component: kubelet + host: gke-jen-pg-687-895efcafe-default-pool-180f5d1b-q1sb type: Normal case.go:366: resource Event:kuttl-test-internal-boxer/: .reason: value mismatch, expected: EnsuredLoadBalancer != actual: Pulling case.go:366: --- Event:kuttl-test-internal-boxer/ +++ Event:kuttl-test-internal-boxer/monitoring-0.17b9e97c0ae35152 @@ -1,17 +1,46 @@ apiVersion: v1 count: 1 +eventTime: null +firstTimestamp: "2024-03-05T15:57:32Z" involvedObject: apiVersion: v1 - kind: Service - name: monitoring-service + fieldPath: spec.containers{monitoring} + kind: Pod + name: monitoring-0 + namespace: kuttl-test-internal-boxer + resourceVersion: "3877" + uid: dacbfee3-eac4-4ca3-ac54-8b80e9c8a721 kind: Event -message: Ensured load balancer +lastTimestamp: "2024-03-05T15:57:32Z" +message: Successfully pulled image "perconalab/pmm-server:dev-latest" in 98.243152ms + (98.256683ms including waiting) metadata: + managedFields: + - apiVersion: v1 + fieldsType: FieldsV1 + fieldsV1: + f:count: {} + f:firstTimestamp: {} + f:involvedObject: {} + f:lastTimestamp: {} + f:message: {} + f:reason: {} + f:reportingComponent: {} + f:reportingInstance: {} + f:source: + f:component: {} + f:host: {} + f:type: {} + manager: kubelet + operation: Update + time: "2024-03-05T15:57:32Z" + name: monitoring-0.17b9e97c0ae35152 namespace: kuttl-test-internal-boxer -reason: EnsuredLoadBalancer -reportingComponent: "" -reportingInstance: "" +reason: Pulled +reportingComponent: kubelet +reportingInstance: gke-jen-pg-687-895efcafe-default-pool-180f5d1b-q1sb source: - component: service-controller + component: kubelet + host: gke-jen-pg-687-895efcafe-default-pool-180f5d1b-q1sb type: Normal case.go:366: resource Event:kuttl-test-internal-boxer/: .reason: value mismatch, expected: EnsuredLoadBalancer != actual: Pulled case.go:366: --- Event:kuttl-test-internal-boxer/ +++ Event:kuttl-test-internal-boxer/monitoring-0.17b9e97c0be1724b @@ -1,17 +1,45 @@ apiVersion: v1 count: 1 +eventTime: null +firstTimestamp: "2024-03-05T15:57:32Z" involvedObject: apiVersion: v1 - kind: Service - name: monitoring-service + fieldPath: spec.containers{monitoring} + kind: Pod + name: monitoring-0 + namespace: kuttl-test-internal-boxer + resourceVersion: "3877" + uid: dacbfee3-eac4-4ca3-ac54-8b80e9c8a721 kind: Event -message: Ensured load balancer +lastTimestamp: "2024-03-05T15:57:32Z" +message: Created container monitoring metadata: + managedFields: + - apiVersion: v1 + fieldsType: FieldsV1 + fieldsV1: + f:count: {} + f:firstTimestamp: {} + f:involvedObject: {} + f:lastTimestamp: {} + f:message: {} + f:reason: {} + f:reportingComponent: {} + f:reportingInstance: {} + f:source: + f:component: {} + f:host: {} + f:type: {} + manager: kubelet + operation: Update + time: "2024-03-05T15:57:32Z" + name: monitoring-0.17b9e97c0be1724b namespace: kuttl-test-internal-boxer -reason: EnsuredLoadBalancer -reportingComponent: "" -reportingInstance: "" +reason: Created +reportingComponent: kubelet +reportingInstance: gke-jen-pg-687-895efcafe-default-pool-180f5d1b-q1sb source: - component: service-controller + component: kubelet + host: gke-jen-pg-687-895efcafe-default-pool-180f5d1b-q1sb type: Normal case.go:366: resource Event:kuttl-test-internal-boxer/: .reportingComponent: value mismatch, expected: != actual: kubelet case.go:366: --- Event:kuttl-test-internal-boxer/ +++ Event:kuttl-test-internal-boxer/monitoring-0.17b9e97c0f26964b @@ -1,17 +1,45 @@ apiVersion: v1 count: 1 +eventTime: null +firstTimestamp: "2024-03-05T15:57:32Z" involvedObject: apiVersion: v1 - kind: Service - name: monitoring-service + fieldPath: spec.containers{monitoring} + kind: Pod + name: monitoring-0 + namespace: kuttl-test-internal-boxer + resourceVersion: "3877" + uid: dacbfee3-eac4-4ca3-ac54-8b80e9c8a721 kind: Event -message: Ensured load balancer +lastTimestamp: "2024-03-05T15:57:32Z" +message: Started container monitoring metadata: + managedFields: + - apiVersion: v1 + fieldsType: FieldsV1 + fieldsV1: + f:count: {} + f:firstTimestamp: {} + f:involvedObject: {} + f:lastTimestamp: {} + f:message: {} + f:reason: {} + f:reportingComponent: {} + f:reportingInstance: {} + f:source: + f:component: {} + f:host: {} + f:type: {} + manager: kubelet + operation: Update + time: "2024-03-05T15:57:32Z" + name: monitoring-0.17b9e97c0f26964b namespace: kuttl-test-internal-boxer -reason: EnsuredLoadBalancer -reportingComponent: "" -reportingInstance: "" +reason: Started +reportingComponent: kubelet +reportingInstance: gke-jen-pg-687-895efcafe-default-pool-180f5d1b-q1sb source: - component: service-controller + component: kubelet + host: gke-jen-pg-687-895efcafe-default-pool-180f5d1b-q1sb type: Normal case.go:366: resource Event:kuttl-test-internal-boxer/: .reason: value mismatch, expected: EnsuredLoadBalancer != actual: Started case.go:366: --- Event:kuttl-test-internal-boxer/ +++ Event:kuttl-test-internal-boxer/monitoring-service.17b9e978eade694c @@ -1,15 +1,39 @@ apiVersion: v1 count: 1 +eventTime: null +firstTimestamp: "2024-03-05T15:57:19Z" involvedObject: apiVersion: v1 kind: Service name: monitoring-service + namespace: kuttl-test-internal-boxer + resourceVersion: "3829" + uid: 0a917e01-ba24-4cc7-ac74-446ed8826858 kind: Event -message: Ensured load balancer +lastTimestamp: "2024-03-05T15:57:19Z" +message: Ensuring load balancer metadata: + managedFields: + - apiVersion: v1 + fieldsType: FieldsV1 + fieldsV1: + f:count: {} + f:firstTimestamp: {} + f:involvedObject: {} + f:lastTimestamp: {} + f:message: {} + f:reason: {} + f:reportingComponent: {} + f:source: + f:component: {} + f:type: {} + manager: kube-controller-manager + operation: Update + time: "2024-03-05T15:57:19Z" + name: monitoring-service.17b9e978eade694c namespace: kuttl-test-internal-boxer -reason: EnsuredLoadBalancer -reportingComponent: "" +reason: EnsuringLoadBalancer +reportingComponent: service-controller reportingInstance: "" source: component: service-controller case.go:366: resource Event:kuttl-test-internal-boxer/: .message: value mismatch, expected: Ensured load balancer != actual: Ensuring load balancer case.go:366: --- Event:kuttl-test-internal-boxer/ +++ Event:kuttl-test-internal-boxer/monitoring-service.17b9e981ee9187bf @@ -1,15 +1,39 @@ apiVersion: v1 count: 1 +eventTime: null +firstTimestamp: "2024-03-05T15:57:57Z" involvedObject: apiVersion: v1 kind: Service name: monitoring-service + namespace: kuttl-test-internal-boxer + resourceVersion: "3829" + uid: 0a917e01-ba24-4cc7-ac74-446ed8826858 kind: Event +lastTimestamp: "2024-03-05T15:57:57Z" message: Ensured load balancer metadata: + managedFields: + - apiVersion: v1 + fieldsType: FieldsV1 + fieldsV1: + f:count: {} + f:firstTimestamp: {} + f:involvedObject: {} + f:lastTimestamp: {} + f:message: {} + f:reason: {} + f:reportingComponent: {} + f:source: + f:component: {} + f:type: {} + manager: kube-controller-manager + operation: Update + time: "2024-03-05T15:57:57Z" + name: monitoring-service.17b9e981ee9187bf namespace: kuttl-test-internal-boxer reason: EnsuredLoadBalancer -reportingComponent: "" +reportingComponent: service-controller reportingInstance: "" source: component: service-controller case.go:366: resource Event:kuttl-test-internal-boxer/: .reportingComponent: value mismatch, expected: != actual: service-controller case.go:366: --- Event:kuttl-test-internal-boxer/ +++ Event:kuttl-test-internal-boxer/monitoring.17b9e978f2c06c00 @@ -1,17 +1,42 @@ apiVersion: v1 count: 1 +eventTime: null +firstTimestamp: "2024-03-05T15:57:19Z" involvedObject: - apiVersion: v1 - kind: Service - name: monitoring-service + apiVersion: apps/v1 + kind: StatefulSet + name: monitoring + namespace: kuttl-test-internal-boxer + resourceVersion: "3833" + uid: 3bb08919-08fc-443f-812e-b3bc8cb23d83 kind: Event -message: Ensured load balancer +lastTimestamp: "2024-03-05T15:57:19Z" +message: create Claim pmmdata-monitoring-0 Pod monitoring-0 in StatefulSet monitoring + success metadata: + managedFields: + - apiVersion: v1 + fieldsType: FieldsV1 + fieldsV1: + f:count: {} + f:firstTimestamp: {} + f:involvedObject: {} + f:lastTimestamp: {} + f:message: {} + f:reason: {} + f:reportingComponent: {} + f:source: + f:component: {} + f:type: {} + manager: kube-controller-manager + operation: Update + time: "2024-03-05T15:57:19Z" + name: monitoring.17b9e978f2c06c00 namespace: kuttl-test-internal-boxer -reason: EnsuredLoadBalancer -reportingComponent: "" +reason: SuccessfulCreate +reportingComponent: statefulset-controller reportingInstance: "" source: - component: service-controller + component: statefulset-controller type: Normal case.go:366: resource Event:kuttl-test-internal-boxer/: .reason: value mismatch, expected: EnsuredLoadBalancer != actual: SuccessfulCreate case.go:366: --- Event:kuttl-test-internal-boxer/ +++ Event:kuttl-test-internal-boxer/monitoring.17b9e978f3fef279 @@ -1,17 +1,41 @@ apiVersion: v1 count: 1 +eventTime: null +firstTimestamp: "2024-03-05T15:57:19Z" involvedObject: - apiVersion: v1 - kind: Service - name: monitoring-service + apiVersion: apps/v1 + kind: StatefulSet + name: monitoring + namespace: kuttl-test-internal-boxer + resourceVersion: "3833" + uid: 3bb08919-08fc-443f-812e-b3bc8cb23d83 kind: Event -message: Ensured load balancer +lastTimestamp: "2024-03-05T15:57:19Z" +message: create Pod monitoring-0 in StatefulSet monitoring successful metadata: + managedFields: + - apiVersion: v1 + fieldsType: FieldsV1 + fieldsV1: + f:count: {} + f:firstTimestamp: {} + f:involvedObject: {} + f:lastTimestamp: {} + f:message: {} + f:reason: {} + f:reportingComponent: {} + f:source: + f:component: {} + f:type: {} + manager: kube-controller-manager + operation: Update + time: "2024-03-05T15:57:19Z" + name: monitoring.17b9e978f3fef279 namespace: kuttl-test-internal-boxer -reason: EnsuredLoadBalancer -reportingComponent: "" +reason: SuccessfulCreate +reportingComponent: statefulset-controller reportingInstance: "" source: - component: service-controller + component: statefulset-controller type: Normal case.go:366: resource Event:kuttl-test-internal-boxer/: .reason: value mismatch, expected: EnsuredLoadBalancer != actual: SuccessfulCreate case.go:366: --- Event:kuttl-test-internal-boxer/ +++ Event:kuttl-test-internal-boxer/pg-client-689f5689bd-5sfr8.17b9e9778a52f0e8 @@ -1,17 +1,42 @@ apiVersion: v1 count: 1 +eventTime: null +firstTimestamp: "2024-03-05T15:57:13Z" involvedObject: apiVersion: v1 - kind: Service - name: monitoring-service + kind: Pod + name: pg-client-689f5689bd-5sfr8 + namespace: kuttl-test-internal-boxer + resourceVersion: "3761" + uid: 6abf2d4a-8d2a-4733-8416-5c514b2fe750 kind: Event -message: Ensured load balancer +lastTimestamp: "2024-03-05T15:57:13Z" +message: Successfully assigned kuttl-test-internal-boxer/pg-client-689f5689bd-5sfr8 + to gke-jen-pg-687-895efcafe-default-pool-180f5d1b-q1sb metadata: + managedFields: + - apiVersion: v1 + fieldsType: FieldsV1 + fieldsV1: + f:count: {} + f:firstTimestamp: {} + f:involvedObject: {} + f:lastTimestamp: {} + f:message: {} + f:reason: {} + f:reportingComponent: {} + f:source: + f:component: {} + f:type: {} + manager: kube-scheduler + operation: Update + time: "2024-03-05T15:57:13Z" + name: pg-client-689f5689bd-5sfr8.17b9e9778a52f0e8 namespace: kuttl-test-internal-boxer -reason: EnsuredLoadBalancer -reportingComponent: "" +reason: Scheduled +reportingComponent: default-scheduler reportingInstance: "" source: - component: service-controller + component: default-scheduler type: Normal case.go:366: resource Event:kuttl-test-internal-boxer/: .reportingComponent: value mismatch, expected: != actual: default-scheduler case.go:366: --- Event:kuttl-test-internal-boxer/ +++ Event:kuttl-test-internal-boxer/pg-client-689f5689bd-5sfr8.17b9e977cb2c9c0e @@ -1,17 +1,45 @@ apiVersion: v1 count: 1 +eventTime: null +firstTimestamp: "2024-03-05T15:57:14Z" involvedObject: apiVersion: v1 - kind: Service - name: monitoring-service + fieldPath: spec.containers{pg-client} + kind: Pod + name: pg-client-689f5689bd-5sfr8 + namespace: kuttl-test-internal-boxer + resourceVersion: "3763" + uid: 6abf2d4a-8d2a-4733-8416-5c514b2fe750 kind: Event -message: Ensured load balancer +lastTimestamp: "2024-03-05T15:57:14Z" +message: Pulling image "perconalab/percona-distribution-postgresql:15" metadata: + managedFields: + - apiVersion: v1 + fieldsType: FieldsV1 + fieldsV1: + f:count: {} + f:firstTimestamp: {} + f:involvedObject: {} + f:lastTimestamp: {} + f:message: {} + f:reason: {} + f:reportingComponent: {} + f:reportingInstance: {} + f:source: + f:component: {} + f:host: {} + f:type: {} + manager: kubelet + operation: Update + time: "2024-03-05T15:57:14Z" + name: pg-client-689f5689bd-5sfr8.17b9e977cb2c9c0e namespace: kuttl-test-internal-boxer -reason: EnsuredLoadBalancer -reportingComponent: "" -reportingInstance: "" +reason: Pulling +reportingComponent: kubelet +reportingInstance: gke-jen-pg-687-895efcafe-default-pool-180f5d1b-q1sb source: - component: service-controller + component: kubelet + host: gke-jen-pg-687-895efcafe-default-pool-180f5d1b-q1sb type: Normal case.go:366: resource Event:kuttl-test-internal-boxer/: .reportingInstance: value mismatch, expected: != actual: gke-jen-pg-687-895efcafe-default-pool-180f5d1b-q1sb case.go:366: --- Event:kuttl-test-internal-boxer/ +++ Event:kuttl-test-internal-boxer/pg-client-689f5689bd-5sfr8.17b9e977d132fb40 @@ -1,17 +1,46 @@ apiVersion: v1 count: 1 +eventTime: null +firstTimestamp: "2024-03-05T15:57:14Z" involvedObject: apiVersion: v1 - kind: Service - name: monitoring-service + fieldPath: spec.containers{pg-client} + kind: Pod + name: pg-client-689f5689bd-5sfr8 + namespace: kuttl-test-internal-boxer + resourceVersion: "3763" + uid: 6abf2d4a-8d2a-4733-8416-5c514b2fe750 kind: Event -message: Ensured load balancer +lastTimestamp: "2024-03-05T15:57:14Z" +message: Successfully pulled image "perconalab/percona-distribution-postgresql:15" + in 101.050083ms (101.05782ms including waiting) metadata: + managedFields: + - apiVersion: v1 + fieldsType: FieldsV1 + fieldsV1: + f:count: {} + f:firstTimestamp: {} + f:involvedObject: {} + f:lastTimestamp: {} + f:message: {} + f:reason: {} + f:reportingComponent: {} + f:reportingInstance: {} + f:source: + f:component: {} + f:host: {} + f:type: {} + manager: kubelet + operation: Update + time: "2024-03-05T15:57:14Z" + name: pg-client-689f5689bd-5sfr8.17b9e977d132fb40 namespace: kuttl-test-internal-boxer -reason: EnsuredLoadBalancer -reportingComponent: "" -reportingInstance: "" +reason: Pulled +reportingComponent: kubelet +reportingInstance: gke-jen-pg-687-895efcafe-default-pool-180f5d1b-q1sb source: - component: service-controller + component: kubelet + host: gke-jen-pg-687-895efcafe-default-pool-180f5d1b-q1sb type: Normal case.go:366: resource Event:kuttl-test-internal-boxer/: .source.component: value mismatch, expected: service-controller != actual: kubelet case.go:366: --- Event:kuttl-test-internal-boxer/ +++ Event:kuttl-test-internal-boxer/pg-client-689f5689bd-5sfr8.17b9e977d26c817e @@ -1,17 +1,45 @@ apiVersion: v1 count: 1 +eventTime: null +firstTimestamp: "2024-03-05T15:57:14Z" involvedObject: apiVersion: v1 - kind: Service - name: monitoring-service + fieldPath: spec.containers{pg-client} + kind: Pod + name: pg-client-689f5689bd-5sfr8 + namespace: kuttl-test-internal-boxer + resourceVersion: "3763" + uid: 6abf2d4a-8d2a-4733-8416-5c514b2fe750 kind: Event -message: Ensured load balancer +lastTimestamp: "2024-03-05T15:57:14Z" +message: Created container pg-client metadata: + managedFields: + - apiVersion: v1 + fieldsType: FieldsV1 + fieldsV1: + f:count: {} + f:firstTimestamp: {} + f:involvedObject: {} + f:lastTimestamp: {} + f:message: {} + f:reason: {} + f:reportingComponent: {} + f:reportingInstance: {} + f:source: + f:component: {} + f:host: {} + f:type: {} + manager: kubelet + operation: Update + time: "2024-03-05T15:57:14Z" + name: pg-client-689f5689bd-5sfr8.17b9e977d26c817e namespace: kuttl-test-internal-boxer -reason: EnsuredLoadBalancer -reportingComponent: "" -reportingInstance: "" +reason: Created +reportingComponent: kubelet +reportingInstance: gke-jen-pg-687-895efcafe-default-pool-180f5d1b-q1sb source: - component: service-controller + component: kubelet + host: gke-jen-pg-687-895efcafe-default-pool-180f5d1b-q1sb type: Normal case.go:366: resource Event:kuttl-test-internal-boxer/: .involvedObject.kind: value mismatch, expected: Service != actual: Pod case.go:366: --- Event:kuttl-test-internal-boxer/ +++ Event:kuttl-test-internal-boxer/pg-client-689f5689bd-5sfr8.17b9e977d5f418b5 @@ -1,17 +1,45 @@ apiVersion: v1 count: 1 +eventTime: null +firstTimestamp: "2024-03-05T15:57:14Z" involvedObject: apiVersion: v1 - kind: Service - name: monitoring-service + fieldPath: spec.containers{pg-client} + kind: Pod + name: pg-client-689f5689bd-5sfr8 + namespace: kuttl-test-internal-boxer + resourceVersion: "3763" + uid: 6abf2d4a-8d2a-4733-8416-5c514b2fe750 kind: Event -message: Ensured load balancer +lastTimestamp: "2024-03-05T15:57:14Z" +message: Started container pg-client metadata: + managedFields: + - apiVersion: v1 + fieldsType: FieldsV1 + fieldsV1: + f:count: {} + f:firstTimestamp: {} + f:involvedObject: {} + f:lastTimestamp: {} + f:message: {} + f:reason: {} + f:reportingComponent: {} + f:reportingInstance: {} + f:source: + f:component: {} + f:host: {} + f:type: {} + manager: kubelet + operation: Update + time: "2024-03-05T15:57:14Z" + name: pg-client-689f5689bd-5sfr8.17b9e977d5f418b5 namespace: kuttl-test-internal-boxer -reason: EnsuredLoadBalancer -reportingComponent: "" -reportingInstance: "" +reason: Started +reportingComponent: kubelet +reportingInstance: gke-jen-pg-687-895efcafe-default-pool-180f5d1b-q1sb source: - component: service-controller + component: kubelet + host: gke-jen-pg-687-895efcafe-default-pool-180f5d1b-q1sb type: Normal case.go:366: resource Event:kuttl-test-internal-boxer/: .reason: value mismatch, expected: EnsuredLoadBalancer != actual: Started case.go:366: --- Event:kuttl-test-internal-boxer/ +++ Event:kuttl-test-internal-boxer/pg-client-689f5689bd.17b9e977896191b1 @@ -1,17 +1,41 @@ apiVersion: v1 count: 1 +eventTime: null +firstTimestamp: "2024-03-05T15:57:13Z" involvedObject: - apiVersion: v1 - kind: Service - name: monitoring-service + apiVersion: apps/v1 + kind: ReplicaSet + name: pg-client-689f5689bd + namespace: kuttl-test-internal-boxer + resourceVersion: "3760" + uid: 701e5b99-a189-4ea9-a419-2ae0290f5749 kind: Event -message: Ensured load balancer +lastTimestamp: "2024-03-05T15:57:13Z" +message: 'Created pod: pg-client-689f5689bd-5sfr8' metadata: + managedFields: + - apiVersion: v1 + fieldsType: FieldsV1 + fieldsV1: + f:count: {} + f:firstTimestamp: {} + f:involvedObject: {} + f:lastTimestamp: {} + f:message: {} + f:reason: {} + f:reportingComponent: {} + f:source: + f:component: {} + f:type: {} + manager: kube-controller-manager + operation: Update + time: "2024-03-05T15:57:13Z" + name: pg-client-689f5689bd.17b9e977896191b1 namespace: kuttl-test-internal-boxer -reason: EnsuredLoadBalancer -reportingComponent: "" +reason: SuccessfulCreate +reportingComponent: replicaset-controller reportingInstance: "" source: - component: service-controller + component: replicaset-controller type: Normal case.go:366: resource Event:kuttl-test-internal-boxer/: .involvedObject.apiVersion: value mismatch, expected: v1 != actual: apps/v1 case.go:366: --- Event:kuttl-test-internal-boxer/ +++ Event:kuttl-test-internal-boxer/pg-client.17b9e97787c415aa @@ -1,17 +1,41 @@ apiVersion: v1 count: 1 +eventTime: null +firstTimestamp: "2024-03-05T15:57:13Z" involvedObject: - apiVersion: v1 - kind: Service - name: monitoring-service + apiVersion: apps/v1 + kind: Deployment + name: pg-client + namespace: kuttl-test-internal-boxer + resourceVersion: "3759" + uid: 199500fb-89d3-405e-b83c-7f7d6d5aa6e6 kind: Event -message: Ensured load balancer +lastTimestamp: "2024-03-05T15:57:13Z" +message: Scaled up replica set pg-client-689f5689bd to 1 metadata: + managedFields: + - apiVersion: v1 + fieldsType: FieldsV1 + fieldsV1: + f:count: {} + f:firstTimestamp: {} + f:involvedObject: {} + f:lastTimestamp: {} + f:message: {} + f:reason: {} + f:reportingComponent: {} + f:source: + f:component: {} + f:type: {} + manager: kube-controller-manager + operation: Update + time: "2024-03-05T15:57:13Z" + name: pg-client.17b9e97787c415aa namespace: kuttl-test-internal-boxer -reason: EnsuredLoadBalancer -reportingComponent: "" +reason: ScalingReplicaSet +reportingComponent: deployment-controller reportingInstance: "" source: - component: service-controller + component: deployment-controller type: Normal case.go:366: resource Event:kuttl-test-internal-boxer/: .reason: value mismatch, expected: EnsuredLoadBalancer != actual: ScalingReplicaSet case.go:366: --- Event:kuttl-test-internal-boxer/ +++ Event:kuttl-test-internal-boxer/pmmdata-monitoring-0.17b9e978f2cb6608 @@ -1,17 +1,41 @@ apiVersion: v1 count: 1 +eventTime: null +firstTimestamp: "2024-03-05T15:57:19Z" involvedObject: apiVersion: v1 - kind: Service - name: monitoring-service + kind: PersistentVolumeClaim + name: pmmdata-monitoring-0 + namespace: kuttl-test-internal-boxer + resourceVersion: "3835" + uid: bd15bf1b-fe46-4802-942f-532ea3704dd3 kind: Event -message: Ensured load balancer +lastTimestamp: "2024-03-05T15:57:19Z" +message: waiting for first consumer to be created before binding metadata: + managedFields: + - apiVersion: v1 + fieldsType: FieldsV1 + fieldsV1: + f:count: {} + f:firstTimestamp: {} + f:involvedObject: {} + f:lastTimestamp: {} + f:message: {} + f:reason: {} + f:reportingComponent: {} + f:source: + f:component: {} + f:type: {} + manager: kube-controller-manager + operation: Update + time: "2024-03-05T15:57:19Z" + name: pmmdata-monitoring-0.17b9e978f2cb6608 namespace: kuttl-test-internal-boxer -reason: EnsuredLoadBalancer -reportingComponent: "" +reason: WaitForFirstConsumer +reportingComponent: persistentvolume-controller reportingInstance: "" source: - component: service-controller + component: persistentvolume-controller type: Normal case.go:366: resource Event:kuttl-test-internal-boxer/: .involvedObject.kind: value mismatch, expected: Service != actual: PersistentVolumeClaim case.go:366: --- Event:kuttl-test-internal-boxer/ +++ Event:kuttl-test-internal-boxer/pmmdata-monitoring-0.17b9e978f61285fd @@ -1,17 +1,42 @@ apiVersion: v1 -count: 1 +count: 3 +eventTime: null +firstTimestamp: "2024-03-05T15:57:19Z" involvedObject: apiVersion: v1 - kind: Service - name: monitoring-service + kind: PersistentVolumeClaim + name: pmmdata-monitoring-0 + namespace: kuttl-test-internal-boxer + resourceVersion: "3839" + uid: bd15bf1b-fe46-4802-942f-532ea3704dd3 kind: Event -message: Ensured load balancer +lastTimestamp: "2024-03-05T15:57:21Z" +message: waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" + or manually created by system administrator metadata: + managedFields: + - apiVersion: v1 + fieldsType: FieldsV1 + fieldsV1: + f:count: {} + f:firstTimestamp: {} + f:involvedObject: {} + f:lastTimestamp: {} + f:message: {} + f:reason: {} + f:reportingComponent: {} + f:source: + f:component: {} + f:type: {} + manager: kube-controller-manager + operation: Update + time: "2024-03-05T15:57:21Z" + name: pmmdata-monitoring-0.17b9e978f61285fd namespace: kuttl-test-internal-boxer -reason: EnsuredLoadBalancer -reportingComponent: "" +reason: ExternalProvisioning +reportingComponent: persistentvolume-controller reportingInstance: "" source: - component: service-controller + component: persistentvolume-controller type: Normal case.go:366: resource Event:kuttl-test-internal-boxer/: .reason: value mismatch, expected: EnsuredLoadBalancer != actual: ExternalProvisioning case.go:366: --- Event:kuttl-test-internal-boxer/ +++ Event:kuttl-test-internal-boxer/pmmdata-monitoring-0.17b9e978f841030f @@ -1,17 +1,41 @@ apiVersion: v1 count: 1 +eventTime: null +firstTimestamp: "2024-03-05T15:57:19Z" involvedObject: apiVersion: v1 - kind: Service - name: monitoring-service + kind: PersistentVolumeClaim + name: pmmdata-monitoring-0 + namespace: kuttl-test-internal-boxer + resourceVersion: "3839" + uid: bd15bf1b-fe46-4802-942f-532ea3704dd3 kind: Event -message: Ensured load balancer +lastTimestamp: "2024-03-05T15:57:19Z" +message: External provisioner is provisioning volume for claim "kuttl-test-internal-boxer/pmmdata-monitoring-0" metadata: + managedFields: + - apiVersion: v1 + fieldsType: FieldsV1 + fieldsV1: + f:count: {} + f:firstTimestamp: {} + f:involvedObject: {} + f:lastTimestamp: {} + f:message: {} + f:reason: {} + f:reportingComponent: {} + f:source: + f:component: {} + f:type: {} + manager: csi-provisioner + operation: Update + time: "2024-03-05T15:57:19Z" + name: pmmdata-monitoring-0.17b9e978f841030f namespace: kuttl-test-internal-boxer -reason: EnsuredLoadBalancer -reportingComponent: "" +reason: Provisioning +reportingComponent: pd.csi.storage.gke.io_gke-2343d8e359cd44729a5d-a5ba-49fc-vm_8d4f97ec-3d2c-4b00-863a-28fddc97f027 reportingInstance: "" source: - component: service-controller + component: pd.csi.storage.gke.io_gke-2343d8e359cd44729a5d-a5ba-49fc-vm_8d4f97ec-3d2c-4b00-863a-28fddc97f027 type: Normal case.go:366: resource Event:kuttl-test-internal-boxer/: .involvedObject.kind: value mismatch, expected: Service != actual: PersistentVolumeClaim case.go:366: --- Event:kuttl-test-internal-boxer/ +++ Event:kuttl-test-internal-boxer/pmmdata-monitoring-0.17b9e979d6db55f5 @@ -1,17 +1,41 @@ apiVersion: v1 count: 1 +eventTime: null +firstTimestamp: "2024-03-05T15:57:23Z" involvedObject: apiVersion: v1 - kind: Service - name: monitoring-service + kind: PersistentVolumeClaim + name: pmmdata-monitoring-0 + namespace: kuttl-test-internal-boxer + resourceVersion: "3839" + uid: bd15bf1b-fe46-4802-942f-532ea3704dd3 kind: Event -message: Ensured load balancer +lastTimestamp: "2024-03-05T15:57:23Z" +message: Successfully provisioned volume pvc-bd15bf1b-fe46-4802-942f-532ea3704dd3 metadata: + managedFields: + - apiVersion: v1 + fieldsType: FieldsV1 + fieldsV1: + f:count: {} + f:firstTimestamp: {} + f:involvedObject: {} + f:lastTimestamp: {} + f:message: {} + f:reason: {} + f:reportingComponent: {} + f:source: + f:component: {} + f:type: {} + manager: csi-provisioner + operation: Update + time: "2024-03-05T15:57:23Z" + name: pmmdata-monitoring-0.17b9e979d6db55f5 namespace: kuttl-test-internal-boxer -reason: EnsuredLoadBalancer -reportingComponent: "" +reason: ProvisioningSucceeded +reportingComponent: pd.csi.storage.gke.io_gke-2343d8e359cd44729a5d-a5ba-49fc-vm_8d4f97ec-3d2c-4b00-863a-28fddc97f027 reportingInstance: "" source: - component: service-controller + component: pd.csi.storage.gke.io_gke-2343d8e359cd44729a5d-a5ba-49fc-vm_8d4f97ec-3d2c-4b00-863a-28fddc97f027 type: Normal case.go:366: resource Event:kuttl-test-internal-boxer/: .reportingComponent: value mismatch, expected: != actual: pd.csi.storage.gke.io_gke-2343d8e359cd44729a5d-a5ba-49fc-vm_8d4f97ec-3d2c-4b00-863a-28fddc97f027 logger.go:42: 15:59:21 | monitoring | monitoring events from ns kuttl-test-internal-boxer: logger.go:42: 15:59:21 | monitoring | 2024-03-05 15:57:13 +0000 UTC Normal Pod pg-client-689f5689bd-5sfr8 Scheduled Successfully assigned kuttl-test-internal-boxer/pg-client-689f5689bd-5sfr8 to gke-jen-pg-687-895efcafe-default-pool-180f5d1b-q1sb default-scheduler logger.go:42: 15:59:21 | monitoring | 2024-03-05 15:57:13 +0000 UTC Normal ReplicaSet.apps pg-client-689f5689bd SuccessfulCreate Created pod: pg-client-689f5689bd-5sfr8 replicaset-controller logger.go:42: 15:59:21 | monitoring | 2024-03-05 15:57:13 +0000 UTC Normal Deployment.apps pg-client ScalingReplicaSet Scaled up replica set pg-client-689f5689bd to 1 deployment-controller logger.go:42: 15:59:21 | monitoring | 2024-03-05 15:57:14 +0000 UTC Normal Pod pg-client-689f5689bd-5sfr8.spec.containers{pg-client} Pulling Pulling image "perconalab/percona-distribution-postgresql:15" kubelet logger.go:42: 15:59:21 | monitoring | 2024-03-05 15:57:14 +0000 UTC Normal Pod pg-client-689f5689bd-5sfr8.spec.containers{pg-client} Pulled Successfully pulled image "perconalab/percona-distribution-postgresql:15" in 101.050083ms (101.05782ms including waiting) kubelet logger.go:42: 15:59:21 | monitoring | 2024-03-05 15:57:14 +0000 UTC Normal Pod pg-client-689f5689bd-5sfr8.spec.containers{pg-client} Created Created container pg-client kubelet logger.go:42: 15:59:21 | monitoring | 2024-03-05 15:57:14 +0000 UTC Normal Pod pg-client-689f5689bd-5sfr8.spec.containers{pg-client} Started Started container pg-client kubelet logger.go:42: 15:59:21 | monitoring | 2024-03-05 15:57:19 +0000 UTC Normal Service monitoring-service EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 15:59:21 | monitoring | 2024-03-05 15:57:19 +0000 UTC Normal StatefulSet.apps monitoring SuccessfulCreate create Claim pmmdata-monitoring-0 Pod monitoring-0 in StatefulSet monitoring success statefulset-controller logger.go:42: 15:59:21 | monitoring | 2024-03-05 15:57:19 +0000 UTC Normal StatefulSet.apps monitoring SuccessfulCreate create Pod monitoring-0 in StatefulSet monitoring successful statefulset-controller logger.go:42: 15:59:21 | monitoring | 2024-03-05 15:57:19 +0000 UTC Normal PersistentVolumeClaim pmmdata-monitoring-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 15:59:21 | monitoring | 2024-03-05 15:57:19 +0000 UTC Normal PersistentVolumeClaim pmmdata-monitoring-0 ExternalProvisioning waiting for a volume to be created, either by external provisioner "pd.csi.storage.gke.io" or manually created by system administrator persistentvolume-controller logger.go:42: 15:59:21 | monitoring | 2024-03-05 15:57:19 +0000 UTC Normal PersistentVolumeClaim pmmdata-monitoring-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-internal-boxer/pmmdata-monitoring-0" pd.csi.storage.gke.io_gke-2343d8e359cd44729a5d-a5ba-49fc-vm_8d4f97ec-3d2c-4b00-863a-28fddc97f027 logger.go:42: 15:59:21 | monitoring | 2024-03-05 15:57:23 +0000 UTC Normal Pod monitoring-0 Scheduled Successfully assigned kuttl-test-internal-boxer/monitoring-0 to gke-jen-pg-687-895efcafe-default-pool-180f5d1b-q1sb default-scheduler logger.go:42: 15:59:21 | monitoring | 2024-03-05 15:57:23 +0000 UTC Normal PersistentVolumeClaim pmmdata-monitoring-0 ProvisioningSucceeded Successfully provisioned volume pvc-bd15bf1b-fe46-4802-942f-532ea3704dd3 pd.csi.storage.gke.io_gke-2343d8e359cd44729a5d-a5ba-49fc-vm_8d4f97ec-3d2c-4b00-863a-28fddc97f027 logger.go:42: 15:59:21 | monitoring | 2024-03-05 15:57:30 +0000 UTC Normal Pod monitoring-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-bd15bf1b-fe46-4802-942f-532ea3704dd3" attachdetach-controller logger.go:42: 15:59:21 | monitoring | 2024-03-05 15:57:32 +0000 UTC Normal Pod monitoring-0.spec.containers{monitoring} Pulling Pulling image "perconalab/pmm-server:dev-latest" kubelet logger.go:42: 15:59:21 | monitoring | 2024-03-05 15:57:32 +0000 UTC Normal Pod monitoring-0.spec.containers{monitoring} Pulled Successfully pulled image "perconalab/pmm-server:dev-latest" in 98.243152ms (98.256683ms including waiting) kubelet logger.go:42: 15:59:21 | monitoring | 2024-03-05 15:57:32 +0000 UTC Normal Pod monitoring-0.spec.containers{monitoring} Created Created container monitoring kubelet logger.go:42: 15:59:21 | monitoring | 2024-03-05 15:57:32 +0000 UTC Normal Pod monitoring-0.spec.containers{monitoring} Started Started container monitoring kubelet logger.go:42: 15:59:21 | monitoring | 2024-03-05 15:57:57 +0000 UTC Normal Service monitoring-service EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 15:59:21 | monitoring | Deleting namespace: kuttl-test-internal-boxer === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (188.15s) --- FAIL: kuttl/harness (0.00s) --- FAIL: kuttl/harness/monitoring (185.37s) FAIL