=== RUN kuttl harness.go:459: starting setup harness.go:254: running tests using configured kubeconfig. harness.go:277: Successful connection to cluster at: https://34.60.89.24 harness.go:362: running tests harness.go:74: going to run test suite with timeout of 180 seconds for each step harness.go:374: testsuite: e2e-tests/tests has 26 tests === RUN kuttl/harness === RUN kuttl/harness/monitoring-pmm3 === PAUSE kuttl/harness/monitoring-pmm3 === CONT kuttl/harness/monitoring-pmm3 logger.go:42: 16:35:56 | monitoring-pmm3 | Creating namespace: kuttl-test-worthy-eagle logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | starting test step 0-deploy-operator logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | running command: [sh -c set -o errexit set -o xtrace source ../../functions init_temp_dir # do this only in the first TestStep deploy_operator deploy_client] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | + source ../../functions logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ realpath ../../.. logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | ++ CERT_MANAGER_VER=1.18.2 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | ++++ pwd logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests/tests/monitoring-pmm3 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | ++ test_name=monitoring-pmm3 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests/vars.sh logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/deploy logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/deploy logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests/conf logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests/conf logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export TEMP_DIR=/tmp/kuttl/pg/monitoring-pmm3 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ TEMP_DIR=/tmp/kuttl/pg/monitoring-pmm3 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export GIT_BRANCH=PR-1273 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ GIT_BRANCH=PR-1273 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export VERSION=PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ VERSION=PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export SKIP_TEST_WARNINGS=true logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ SKIP_TEST_WARNINGS=true logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ command -v oc logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export PG_VER=17 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ PG_VER=17 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export BUCKET=pg-operator-testing logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ BUCKET=pg-operator-testing logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export PGOV1_TAG=1.4.0 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ PGOV1_TAG=1.4.0 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export PGOV1_VER=14 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ PGOV1_VER=14 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export MINIO_VER=5.4.0 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ MINIO_VER=5.4.0 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export REGISTRY_NAME=docker.io logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ REGISTRY_NAME=docker.io logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export REGISTRY_NAME_FULL=docker.io/ logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ REGISTRY_NAME_FULL=docker.io/ logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | ++++ printenv logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | ++++ grep -E '^IMAGE' logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | ++++ awk -F= '{print $1}' logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | ++++ eval 'echo $IMAGE_BACKREST' logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++++ echo perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ var_value=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest17 == docker.io/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest17 == percona/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest17 == perconalab/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ echo IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ echo perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | ++++ eval 'echo $IMAGE_PMM3_CLIENT' logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++++ echo perconalab/pmm-client:3-dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ var_value=perconalab/pmm-client:3-dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/pmm-client:3-dev-latest == docker.io/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/pmm-client:3-dev-latest == percona/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/pmm-client:3-dev-latest == perconalab/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ new_value=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ echo IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ echo perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | ++++ eval 'echo $IMAGE' logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++++ echo perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ var_value=perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:PR-1273-332364a9b == docker.io/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:PR-1273-332364a9b == percona/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:PR-1273-332364a9b == perconalab/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ new_value=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ echo IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | ++++ eval 'echo $IMAGE_UPGRADE' logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++++ echo perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ var_value=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == docker.io/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == percona/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == perconalab/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ echo IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | ++++ eval 'echo $IMAGE_BASE' logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++++ echo perconalab/percona-postgresql-operator logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ var_value=perconalab/percona-postgresql-operator logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator == docker.io/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator == percona/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator == perconalab/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ new_value=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ echo IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | ++++ eval 'echo $IMAGE_PMM3_SERVER' logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++++ echo perconalab/pmm-server:3-dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ var_value=perconalab/pmm-server:3-dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/pmm-server:3-dev-latest == docker.io/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/pmm-server:3-dev-latest == percona/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/pmm-server:3-dev-latest == perconalab/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ new_value=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ echo IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | ++++ eval 'echo $IMAGE_PGBOUNCER' logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++++ echo perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ var_value=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer17 == docker.io/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer17 == percona/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer17 == perconalab/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ echo IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | ++++ eval 'echo $IMAGE_POSTGRESQL' logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++++ echo perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ var_value=perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:main-ppg17-postgres == docker.io/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:main-ppg17-postgres == percona/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:main-ppg17-postgres == perconalab/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ echo IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | ++++ eval 'echo $IMAGE_PMM_SERVER' logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++++ echo perconalab/pmm-server:dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ var_value=perconalab/pmm-server:dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/pmm-server:dev-latest == docker.io/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/pmm-server:dev-latest == percona/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/pmm-server:dev-latest == perconalab/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ new_value=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ echo IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | ++++ eval 'echo $IMAGE_PMM_CLIENT' logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++++ echo perconalab/pmm-client:dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ var_value=perconalab/pmm-client:dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/pmm-client:dev-latest == docker.io/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/pmm-client:dev-latest == percona/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ [[ perconalab/pmm-client:dev-latest == perconalab/* ]] logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ new_value=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ export IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ echo IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | ++++ which gdate logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | ++++ which date logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ date=/usr/sbin/date logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | ++++ which gsed logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | ++++ which sed logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | +++ sed=/usr/sbin/sed logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | ++ oc get projects logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | + init_temp_dir logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | + rm -rf /tmp/kuttl/pg/monitoring-pmm3 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | + mkdir -p /tmp/kuttl/pg/monitoring-pmm3 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | + deploy_operator logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | + local cw_prefix= logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | + destroy_operator logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | + kubectl -n pg-operator delete deployment percona-postgresql-operator --force --grace-period=0 logger.go:42: 16:35:56 | monitoring-pmm3/0-deploy-operator | Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. logger.go:42: 16:35:57 | monitoring-pmm3/0-deploy-operator | deployment.apps "percona-postgresql-operator" force deleted from pg-operator namespace logger.go:42: 16:35:57 | monitoring-pmm3/0-deploy-operator | + [[ -n pg-operator ]] logger.go:42: 16:35:57 | monitoring-pmm3/0-deploy-operator | + kubectl delete namespace pg-operator --force --grace-period=0 logger.go:42: 16:35:57 | monitoring-pmm3/0-deploy-operator | Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. logger.go:42: 16:35:57 | monitoring-pmm3/0-deploy-operator | namespace "pg-operator" force deleted logger.go:42: 16:36:04 | monitoring-pmm3/0-deploy-operator | + [[ -n pg-operator ]] logger.go:42: 16:36:04 | monitoring-pmm3/0-deploy-operator | + create_namespace pg-operator logger.go:42: 16:36:04 | monitoring-pmm3/0-deploy-operator | + local namespace=pg-operator logger.go:42: 16:36:04 | monitoring-pmm3/0-deploy-operator | + [[ -n '' ]] logger.go:42: 16:36:04 | monitoring-pmm3/0-deploy-operator | + kubectl delete namespace pg-operator --ignore-not-found logger.go:42: 16:36:05 | monitoring-pmm3/0-deploy-operator | + kubectl wait --for=delete namespace pg-operator logger.go:42: 16:36:05 | monitoring-pmm3/0-deploy-operator | + kubectl create namespace pg-operator logger.go:42: 16:36:06 | monitoring-pmm3/0-deploy-operator | namespace/pg-operator created logger.go:42: 16:36:06 | monitoring-pmm3/0-deploy-operator | + cw_prefix=cw- logger.go:42: 16:36:06 | monitoring-pmm3/0-deploy-operator | + kubectl -n pg-operator apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pg-operator_PR-1273/deploy/crd.yaml logger.go:42: 16:36:07 | monitoring-pmm3/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/crunchybridgeclusters.postgres-operator.crunchydata.com serverside-applied logger.go:42: 16:36:07 | monitoring-pmm3/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconapgbackups.pgv2.percona.com serverside-applied logger.go:42: 16:36:09 | monitoring-pmm3/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconapgclusters.pgv2.percona.com serverside-applied logger.go:42: 16:36:09 | monitoring-pmm3/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconapgrestores.pgv2.percona.com serverside-applied logger.go:42: 16:36:10 | monitoring-pmm3/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconapgupgrades.pgv2.percona.com serverside-applied logger.go:42: 16:36:10 | monitoring-pmm3/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/pgadmins.postgres-operator.crunchydata.com serverside-applied logger.go:42: 16:36:11 | monitoring-pmm3/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/pgupgrades.postgres-operator.crunchydata.com serverside-applied logger.go:42: 16:36:14 | monitoring-pmm3/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/postgresclusters.postgres-operator.crunchydata.com serverside-applied logger.go:42: 16:36:14 | monitoring-pmm3/0-deploy-operator | + kubectl -n pg-operator apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pg-operator_PR-1273/deploy/cw-rbac.yaml logger.go:42: 16:36:15 | monitoring-pmm3/0-deploy-operator | serviceaccount/percona-postgresql-operator serverside-applied logger.go:42: 16:36:15 | monitoring-pmm3/0-deploy-operator | clusterrole.rbac.authorization.k8s.io/percona-postgresql-operator serverside-applied logger.go:42: 16:36:15 | monitoring-pmm3/0-deploy-operator | clusterrolebinding.rbac.authorization.k8s.io/percona-postgresql-operator serverside-applied logger.go:42: 16:36:15 | monitoring-pmm3/0-deploy-operator | + local disable_telemetry=true logger.go:42: 16:36:15 | monitoring-pmm3/0-deploy-operator | + '[' monitoring-pmm3 == telemetry-transfer ']' logger.go:42: 16:36:15 | monitoring-pmm3/0-deploy-operator | + yq eval '.spec.template.spec.containers[0].image = "docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b"' /mnt/jenkins/workspace/cloud-pg-operator_PR-1273/deploy/cw-operator.yaml logger.go:42: 16:36:15 | monitoring-pmm3/0-deploy-operator | + yq eval '(.spec.template.spec.containers[] | select(.name=="operator") | .env[] | select(.name=="DISABLE_TELEMETRY") | .value) = "true"' - logger.go:42: 16:36:15 | monitoring-pmm3/0-deploy-operator | + kubectl -n pg-operator apply -f - logger.go:42: 16:36:16 | monitoring-pmm3/0-deploy-operator | deployment.apps/percona-postgresql-operator created logger.go:42: 16:36:16 | monitoring-pmm3/0-deploy-operator | + deploy_client logger.go:42: 16:36:16 | monitoring-pmm3/0-deploy-operator | + kubectl -n kuttl-test-worthy-eagle apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests/conf/client.yaml logger.go:42: 16:36:17 | monitoring-pmm3/0-deploy-operator | deployment.apps/pg-client created logger.go:42: 16:36:18 | monitoring-pmm3/0-deploy-operator | running command: [sh -c kubectl assert exist-enhanced deployment percona-postgresql-operator -n ${OPERATOR_NS:-$NAMESPACE} --field-selector status.readyReplicas=1] logger.go:42: 16:36:18 | monitoring-pmm3/0-deploy-operator | ASSERT deployment percona-postgresql-operator matching field criteria 'status.readyReplicas=1' should exist. logger.go:42: 16:36:19 | monitoring-pmm3/0-deploy-operator | INFO Found 1 resource(s). logger.go:42: 16:36:19 | monitoring-pmm3/0-deploy-operator | NAME NAMESPACE COL0 logger.go:42: 16:36:19 | monitoring-pmm3/0-deploy-operator | percona-postgresql-operator pg-operator 1 logger.go:42: 16:36:19 | monitoring-pmm3/0-deploy-operator | ASSERT PASS logger.go:42: 16:36:19 | monitoring-pmm3/0-deploy-operator | test step completed 0-deploy-operator logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | starting test step 1-deploy-pmm-server logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | running command: [sh -c set -o errexit set -o xtrace source ../../functions deploy_pmm3_server] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | + source ../../functions logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ realpath ../../.. logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | ++ CERT_MANAGER_VER=1.18.2 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | ++++ pwd logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests/tests/monitoring-pmm3 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | ++ test_name=monitoring-pmm3 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests/vars.sh logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/deploy logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/deploy logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests/conf logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests/conf logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export TEMP_DIR=/tmp/kuttl/pg/monitoring-pmm3 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ TEMP_DIR=/tmp/kuttl/pg/monitoring-pmm3 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export GIT_BRANCH=PR-1273 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ GIT_BRANCH=PR-1273 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export VERSION=PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ VERSION=PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export SKIP_TEST_WARNINGS=true logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ SKIP_TEST_WARNINGS=true logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ command -v oc logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export PG_VER=17 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ PG_VER=17 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export BUCKET=pg-operator-testing logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ BUCKET=pg-operator-testing logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export PGOV1_TAG=1.4.0 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ PGOV1_TAG=1.4.0 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export PGOV1_VER=14 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ PGOV1_VER=14 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export MINIO_VER=5.4.0 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ MINIO_VER=5.4.0 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export REGISTRY_NAME=docker.io logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ REGISTRY_NAME=docker.io logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export REGISTRY_NAME_FULL=docker.io/ logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ REGISTRY_NAME_FULL=docker.io/ logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | ++++ printenv logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | ++++ grep -E '^IMAGE' logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | ++++ awk -F= '{print $1}' logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | ++++ eval 'echo $IMAGE_BACKREST' logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++++ echo perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ var_value=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest17 == docker.io/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest17 == percona/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest17 == perconalab/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ echo IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ echo perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | ++++ eval 'echo $IMAGE_PMM3_CLIENT' logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++++ echo perconalab/pmm-client:3-dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ var_value=perconalab/pmm-client:3-dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/pmm-client:3-dev-latest == docker.io/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/pmm-client:3-dev-latest == percona/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/pmm-client:3-dev-latest == perconalab/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ new_value=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ echo IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ echo perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | ++++ eval 'echo $IMAGE' logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++++ echo perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ var_value=perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/percona-postgresql-operator:PR-1273-332364a9b == docker.io/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/percona-postgresql-operator:PR-1273-332364a9b == percona/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/percona-postgresql-operator:PR-1273-332364a9b == perconalab/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ new_value=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ echo IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | ++++ eval 'echo $IMAGE_UPGRADE' logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++++ echo perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ var_value=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == docker.io/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == percona/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == perconalab/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ echo IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | ++++ eval 'echo $IMAGE_BASE' logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++++ echo perconalab/percona-postgresql-operator logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ var_value=perconalab/percona-postgresql-operator logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/percona-postgresql-operator == docker.io/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/percona-postgresql-operator == percona/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/percona-postgresql-operator == perconalab/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ new_value=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ echo IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | ++++ eval 'echo $IMAGE_PMM3_SERVER' logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++++ echo perconalab/pmm-server:3-dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ var_value=perconalab/pmm-server:3-dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/pmm-server:3-dev-latest == docker.io/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/pmm-server:3-dev-latest == percona/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/pmm-server:3-dev-latest == perconalab/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ new_value=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ echo IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | ++++ eval 'echo $IMAGE_PGBOUNCER' logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++++ echo perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ var_value=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer17 == docker.io/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer17 == percona/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer17 == perconalab/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ echo IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | ++++ eval 'echo $IMAGE_POSTGRESQL' logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++++ echo perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ var_value=perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/percona-postgresql-operator:main-ppg17-postgres == docker.io/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/percona-postgresql-operator:main-ppg17-postgres == percona/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/percona-postgresql-operator:main-ppg17-postgres == perconalab/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ echo IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | ++++ eval 'echo $IMAGE_PMM_SERVER' logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++++ echo perconalab/pmm-server:dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ var_value=perconalab/pmm-server:dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/pmm-server:dev-latest == docker.io/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/pmm-server:dev-latest == percona/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/pmm-server:dev-latest == perconalab/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ new_value=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ echo IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | ++++ eval 'echo $IMAGE_PMM_CLIENT' logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++++ echo perconalab/pmm-client:dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ var_value=perconalab/pmm-client:dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/pmm-client:dev-latest == docker.io/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/pmm-client:dev-latest == percona/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ [[ perconalab/pmm-client:dev-latest == perconalab/* ]] logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ new_value=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ export IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ echo IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | ++++ which gdate logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | ++++ which date logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ date=/usr/sbin/date logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | ++++ which gsed logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | ++++ which sed logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | +++ sed=/usr/sbin/sed logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | ++ oc get projects logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | + deploy_pmm3_server logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | + helm uninstall -n kuttl-test-worthy-eagle monitoring logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | Error: uninstall: Release not loaded: monitoring: release: not found logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | + : logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | + helm repo remove percona logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | "percona" has been removed from your repositories logger.go:42: 16:36:19 | monitoring-pmm3/1-deploy-pmm-server | + kubectl delete clusterrole monitoring --ignore-not-found logger.go:42: 16:36:20 | monitoring-pmm3/1-deploy-pmm-server | + kubectl delete clusterrolebinding monitoring --ignore-not-found logger.go:42: 16:36:20 | monitoring-pmm3/1-deploy-pmm-server | + helm repo add percona https://percona.github.io/percona-helm-charts/ logger.go:42: 16:36:21 | monitoring-pmm3/1-deploy-pmm-server | "percona" has been added to your repositories logger.go:42: 16:36:21 | monitoring-pmm3/1-deploy-pmm-server | + helm repo update logger.go:42: 16:36:21 | monitoring-pmm3/1-deploy-pmm-server | Hang tight while we grab the latest from your chart repositories... logger.go:42: 16:36:21 | monitoring-pmm3/1-deploy-pmm-server | ...Successfully got an update from the "percona" chart repository logger.go:42: 16:36:21 | monitoring-pmm3/1-deploy-pmm-server | Update Complete. ⎈Happy Helming!⎈ logger.go:42: 16:36:21 | monitoring-pmm3/1-deploy-pmm-server | + [[ -n '' ]] logger.go:42: 16:36:21 | monitoring-pmm3/1-deploy-pmm-server | + retry 10 60 helm install monitoring percona/pmm -n kuttl-test-worthy-eagle --set fullnameOverride=monitoring --set image.tag=3-dev-latest --set image.repository=docker.io/perconalab/pmm-server --set service.type=LoadBalancer --force logger.go:42: 16:36:21 | monitoring-pmm3/1-deploy-pmm-server | + local max=10 logger.go:42: 16:36:21 | monitoring-pmm3/1-deploy-pmm-server | + local delay=60 logger.go:42: 16:36:21 | monitoring-pmm3/1-deploy-pmm-server | + shift 2 logger.go:42: 16:36:21 | monitoring-pmm3/1-deploy-pmm-server | + local n=1 logger.go:42: 16:36:21 | monitoring-pmm3/1-deploy-pmm-server | + helm install monitoring percona/pmm -n kuttl-test-worthy-eagle --set fullnameOverride=monitoring --set image.tag=3-dev-latest --set image.repository=docker.io/perconalab/pmm-server --set service.type=LoadBalancer --force logger.go:42: 16:36:25 | monitoring-pmm3/1-deploy-pmm-server | NAME: monitoring logger.go:42: 16:36:25 | monitoring-pmm3/1-deploy-pmm-server | LAST DEPLOYED: Thu Sep 25 16:36:22 2025 logger.go:42: 16:36:25 | monitoring-pmm3/1-deploy-pmm-server | NAMESPACE: kuttl-test-worthy-eagle logger.go:42: 16:36:25 | monitoring-pmm3/1-deploy-pmm-server | STATUS: deployed logger.go:42: 16:36:25 | monitoring-pmm3/1-deploy-pmm-server | REVISION: 1 logger.go:42: 16:36:25 | monitoring-pmm3/1-deploy-pmm-server | TEST SUITE: None logger.go:42: 16:36:25 | monitoring-pmm3/1-deploy-pmm-server | NOTES: logger.go:42: 16:36:25 | monitoring-pmm3/1-deploy-pmm-server | Percona Monitoring and Management (PMM) logger.go:42: 16:36:25 | monitoring-pmm3/1-deploy-pmm-server | logger.go:42: 16:36:25 | monitoring-pmm3/1-deploy-pmm-server | An open source database monitoring, observability and management tool logger.go:42: 16:36:25 | monitoring-pmm3/1-deploy-pmm-server | Check more info here: https://docs.percona.com/percona-monitoring-and-management/index.html logger.go:42: 16:36:25 | monitoring-pmm3/1-deploy-pmm-server | logger.go:42: 16:36:25 | monitoring-pmm3/1-deploy-pmm-server | Get the application URL: logger.go:42: 16:36:25 | monitoring-pmm3/1-deploy-pmm-server | NOTE: It may take a few minutes for the LoadBalancer IP to be available. logger.go:42: 16:36:25 | monitoring-pmm3/1-deploy-pmm-server | You can watch the status of by running 'kubectl get --namespace kuttl-test-worthy-eagle svc -w monitoring-service' logger.go:42: 16:36:25 | monitoring-pmm3/1-deploy-pmm-server | export SERVICE_IP=$(kubectl get svc --namespace kuttl-test-worthy-eagle monitoring-service -o jsonpath="{.status.loadBalancer.ingress[0].ip}") logger.go:42: 16:36:25 | monitoring-pmm3/1-deploy-pmm-server | echo https://$SERVICE_IP: logger.go:42: 16:36:25 | monitoring-pmm3/1-deploy-pmm-server | logger.go:42: 16:36:25 | monitoring-pmm3/1-deploy-pmm-server | Get password for the "admin" user: logger.go:42: 16:36:25 | monitoring-pmm3/1-deploy-pmm-server | export ADMIN_PASS=$(kubectl get secret pmm-secret --namespace kuttl-test-worthy-eagle -o jsonpath='{.data.PMM_ADMIN_PASSWORD}' | base64 --decode) logger.go:42: 16:36:25 | monitoring-pmm3/1-deploy-pmm-server | echo $ADMIN_PASS logger.go:42: 16:37:53 | monitoring-pmm3/1-deploy-pmm-server | test step completed 1-deploy-pmm-server logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | starting test step 2-create-pmm-secret logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | running command: [sh -c set -o errexit set -o xtrace source ../../functions kubectl create -n "${NAMESPACE}" secret generic monitoring-pmm3-pmm-secret --from-literal=PMM_SERVER_TOKEN="" || true] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | + source ../../functions logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ realpath ../../.. logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | ++ CERT_MANAGER_VER=1.18.2 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | ++++ pwd logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests/tests/monitoring-pmm3 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | ++ test_name=monitoring-pmm3 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests/vars.sh logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/deploy logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/deploy logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests/conf logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests/conf logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export TEMP_DIR=/tmp/kuttl/pg/monitoring-pmm3 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ TEMP_DIR=/tmp/kuttl/pg/monitoring-pmm3 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export GIT_BRANCH=PR-1273 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ GIT_BRANCH=PR-1273 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export VERSION=PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ VERSION=PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export SKIP_TEST_WARNINGS=true logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ SKIP_TEST_WARNINGS=true logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ command -v oc logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export PG_VER=17 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ PG_VER=17 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export BUCKET=pg-operator-testing logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ BUCKET=pg-operator-testing logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export PGOV1_TAG=1.4.0 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ PGOV1_TAG=1.4.0 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export PGOV1_VER=14 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ PGOV1_VER=14 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export MINIO_VER=5.4.0 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ MINIO_VER=5.4.0 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export REGISTRY_NAME=docker.io logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ REGISTRY_NAME=docker.io logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export REGISTRY_NAME_FULL=docker.io/ logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ REGISTRY_NAME_FULL=docker.io/ logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | ++++ printenv logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | ++++ grep -E '^IMAGE' logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | ++++ awk -F= '{print $1}' logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | ++++ eval 'echo $IMAGE_BACKREST' logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++++ echo perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ var_value=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest17 == docker.io/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest17 == percona/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest17 == perconalab/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ echo IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ echo perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | ++++ eval 'echo $IMAGE_PMM3_CLIENT' logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++++ echo perconalab/pmm-client:3-dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ var_value=perconalab/pmm-client:3-dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/pmm-client:3-dev-latest == docker.io/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/pmm-client:3-dev-latest == percona/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/pmm-client:3-dev-latest == perconalab/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ new_value=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ echo IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ echo perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | ++++ eval 'echo $IMAGE' logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++++ echo perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ var_value=perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/percona-postgresql-operator:PR-1273-332364a9b == docker.io/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/percona-postgresql-operator:PR-1273-332364a9b == percona/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/percona-postgresql-operator:PR-1273-332364a9b == perconalab/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ new_value=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ echo IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | ++++ eval 'echo $IMAGE_UPGRADE' logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++++ echo perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ var_value=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == docker.io/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == percona/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == perconalab/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ echo IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | ++++ eval 'echo $IMAGE_BASE' logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++++ echo perconalab/percona-postgresql-operator logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ var_value=perconalab/percona-postgresql-operator logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/percona-postgresql-operator == docker.io/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/percona-postgresql-operator == percona/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/percona-postgresql-operator == perconalab/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ new_value=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ echo IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | ++++ eval 'echo $IMAGE_PMM3_SERVER' logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++++ echo perconalab/pmm-server:3-dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ var_value=perconalab/pmm-server:3-dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/pmm-server:3-dev-latest == docker.io/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/pmm-server:3-dev-latest == percona/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/pmm-server:3-dev-latest == perconalab/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ new_value=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ echo IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | ++++ eval 'echo $IMAGE_PGBOUNCER' logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++++ echo perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ var_value=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer17 == docker.io/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer17 == percona/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer17 == perconalab/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ echo IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | ++++ eval 'echo $IMAGE_POSTGRESQL' logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++++ echo perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ var_value=perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/percona-postgresql-operator:main-ppg17-postgres == docker.io/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/percona-postgresql-operator:main-ppg17-postgres == percona/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/percona-postgresql-operator:main-ppg17-postgres == perconalab/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ echo IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | ++++ eval 'echo $IMAGE_PMM_SERVER' logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++++ echo perconalab/pmm-server:dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ var_value=perconalab/pmm-server:dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/pmm-server:dev-latest == docker.io/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/pmm-server:dev-latest == percona/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/pmm-server:dev-latest == perconalab/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ new_value=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ echo IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | ++++ eval 'echo $IMAGE_PMM_CLIENT' logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++++ echo perconalab/pmm-client:dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ var_value=perconalab/pmm-client:dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/pmm-client:dev-latest == docker.io/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/pmm-client:dev-latest == percona/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ [[ perconalab/pmm-client:dev-latest == perconalab/* ]] logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ new_value=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ export IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ echo IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | ++++ which gdate logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | ++++ which date logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ date=/usr/sbin/date logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | ++++ which gsed logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | ++++ which sed logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | +++ sed=/usr/sbin/sed logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | ++ oc get projects logger.go:42: 16:37:53 | monitoring-pmm3/2-create-pmm-secret | + kubectl create -n kuttl-test-worthy-eagle secret generic monitoring-pmm3-pmm-secret --from-literal=PMM_SERVER_TOKEN= logger.go:42: 16:37:54 | monitoring-pmm3/2-create-pmm-secret | secret/monitoring-pmm3-pmm-secret created logger.go:42: 16:37:54 | monitoring-pmm3/2-create-pmm-secret | test step completed 2-create-pmm-secret logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | starting test step 3-create-cluster logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | running command: [sh -c set -o errexit set -o xtrace source ../../functions get_cr \ | yq eval '.spec.pmm.enabled=true' - \ | yq eval ".spec.pmm.image=\"${IMAGE_PMM3_CLIENT}\"" - \ | kubectl -n "${NAMESPACE}" apply -f -] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | + source ../../functions logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ realpath ../../.. logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | ++ CERT_MANAGER_VER=1.18.2 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | ++++ pwd logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests/tests/monitoring-pmm3 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | ++ test_name=monitoring-pmm3 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests/vars.sh logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/deploy logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/deploy logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests/conf logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests/conf logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export TEMP_DIR=/tmp/kuttl/pg/monitoring-pmm3 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ TEMP_DIR=/tmp/kuttl/pg/monitoring-pmm3 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export GIT_BRANCH=PR-1273 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ GIT_BRANCH=PR-1273 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export VERSION=PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ VERSION=PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export SKIP_TEST_WARNINGS=true logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ SKIP_TEST_WARNINGS=true logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ command -v oc logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export PG_VER=17 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ PG_VER=17 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export BUCKET=pg-operator-testing logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ BUCKET=pg-operator-testing logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export PGOV1_TAG=1.4.0 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ PGOV1_TAG=1.4.0 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export PGOV1_VER=14 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ PGOV1_VER=14 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export MINIO_VER=5.4.0 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ MINIO_VER=5.4.0 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export REGISTRY_NAME=docker.io logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ REGISTRY_NAME=docker.io logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export REGISTRY_NAME_FULL=docker.io/ logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ REGISTRY_NAME_FULL=docker.io/ logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | ++++ printenv logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | ++++ grep -E '^IMAGE' logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | ++++ awk -F= '{print $1}' logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | ++++ eval 'echo $IMAGE_BACKREST' logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++++ echo perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ var_value=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest17 == docker.io/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest17 == percona/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest17 == perconalab/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ echo IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ echo perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | ++++ eval 'echo $IMAGE_PMM3_CLIENT' logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++++ echo perconalab/pmm-client:3-dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ var_value=perconalab/pmm-client:3-dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/pmm-client:3-dev-latest == docker.io/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/pmm-client:3-dev-latest == percona/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/pmm-client:3-dev-latest == perconalab/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ new_value=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ echo IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ echo perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | ++++ eval 'echo $IMAGE' logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++++ echo perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ var_value=perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/percona-postgresql-operator:PR-1273-332364a9b == docker.io/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/percona-postgresql-operator:PR-1273-332364a9b == percona/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/percona-postgresql-operator:PR-1273-332364a9b == perconalab/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ new_value=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ echo IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | ++++ eval 'echo $IMAGE_UPGRADE' logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++++ echo perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ var_value=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == docker.io/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == percona/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == perconalab/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ echo IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | ++++ eval 'echo $IMAGE_BASE' logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++++ echo perconalab/percona-postgresql-operator logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ var_value=perconalab/percona-postgresql-operator logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/percona-postgresql-operator == docker.io/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/percona-postgresql-operator == percona/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/percona-postgresql-operator == perconalab/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ new_value=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ echo IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | ++++ eval 'echo $IMAGE_PMM3_SERVER' logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++++ echo perconalab/pmm-server:3-dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ var_value=perconalab/pmm-server:3-dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/pmm-server:3-dev-latest == docker.io/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/pmm-server:3-dev-latest == percona/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/pmm-server:3-dev-latest == perconalab/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ new_value=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ echo IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | ++++ eval 'echo $IMAGE_PGBOUNCER' logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++++ echo perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ var_value=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer17 == docker.io/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer17 == percona/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer17 == perconalab/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ echo IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | ++++ eval 'echo $IMAGE_POSTGRESQL' logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++++ echo perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ var_value=perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/percona-postgresql-operator:main-ppg17-postgres == docker.io/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/percona-postgresql-operator:main-ppg17-postgres == percona/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/percona-postgresql-operator:main-ppg17-postgres == perconalab/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ echo IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | ++++ eval 'echo $IMAGE_PMM_SERVER' logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++++ echo perconalab/pmm-server:dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ var_value=perconalab/pmm-server:dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/pmm-server:dev-latest == docker.io/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/pmm-server:dev-latest == percona/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/pmm-server:dev-latest == perconalab/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ new_value=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ echo IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | ++++ eval 'echo $IMAGE_PMM_CLIENT' logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++++ echo perconalab/pmm-client:dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ var_value=perconalab/pmm-client:dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/pmm-client:dev-latest == docker.io/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/pmm-client:dev-latest == percona/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ [[ perconalab/pmm-client:dev-latest == perconalab/* ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ new_value=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ export IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ echo IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | ++++ which gdate logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | ++++ which date logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ date=/usr/sbin/date logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | ++++ which gsed logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | ++++ which sed logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | +++ sed=/usr/sbin/sed logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | ++ oc get projects logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | + get_cr logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | + local cr_name= logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | + '[' -z ']' logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | + cr_name=monitoring-pmm3 logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | + local repo_path= logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | + local source_path= logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | + yq eval .spec.pmm.enabled=true - logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | + yq eval ' logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | .metadata.name = "monitoring-pmm3" | logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | .metadata.labels = {"e2e":"monitoring-pmm3"} | logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | .spec.postgresVersion = 17 | logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | .spec.users += [{"name":"postgres","password":{"type":"AlphaNumeric"}}] | logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | .spec.users += [{"name":"monitoring-pmm3","password":{"type":"AlphaNumeric"}}] | logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | .spec.image = "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" | logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | .spec.initContainer.image = "docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b" | logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | .spec.backups.pgbackrest.image = "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17" | logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | .spec.proxy.pgBouncer.image = "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17" | logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | .spec.pmm.image = "docker.io/perconalab/pmm-client:dev-latest" | logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | .spec.pmm.secret = "monitoring-pmm3-pmm-secret" | logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | .spec.pmm.customClusterName = "monitoring-pmm3-pmm-custom-name" | logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | .spec.pmm.postgresParams = "--environment=dev-postgres" logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | ' /mnt/jenkins/workspace/cloud-pg-operator_PR-1273/deploy/cr.yaml logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | + yq eval '.spec.pmm.image="docker.io/perconalab/pmm-client:3-dev-latest"' - logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | + kubectl -n kuttl-test-worthy-eagle apply -f - logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | + [[ -n '' ]] logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | + case $test_name in logger.go:42: 16:37:54 | monitoring-pmm3/3-create-cluster | + cat /tmp/kuttl/pg/monitoring-pmm3/cr.yaml logger.go:42: 16:37:56 | monitoring-pmm3/3-create-cluster | perconapgcluster.pgv2.percona.com/monitoring-pmm3 created logger.go:42: 16:39:17 | monitoring-pmm3/3-create-cluster | test step completed 3-create-cluster logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | starting test step 4-update-pmm-server-token logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | running command: [sh -c set -o errexit set -o xtrace source ../../functions sts=$(kubectl -n "${NAMESPACE}" get sts --selector=postgres-operator.crunchydata.com/instance-set=instance1 -o jsonpath='{.items[*].metadata.name}') for st in $sts; do wait_for_generation "sts/$st" 2 done token=$(generate_pmm3_server_token) [[ -n ${token} && ${token} != null ]] \ && kubectl -n ${NAMESPACE} patch secret monitoring-pmm3-pmm-secret --type merge --patch '{"stringData": {"PMM_SERVER_TOKEN": "'${token}'"}}' \ || true sleep 10 sts=$(kubectl -n "${NAMESPACE}" get sts --selector=postgres-operator.crunchydata.com/instance-set=instance1 -o jsonpath='{.items[*].metadata.name}') for st in $sts; do wait_for_generation "sts/$st" 3 done sleep 25] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | + source ../../functions logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ realpath ../../.. logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | ++ CERT_MANAGER_VER=1.18.2 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | ++++ pwd logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests/tests/monitoring-pmm3 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | ++ test_name=monitoring-pmm3 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests/vars.sh logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/deploy logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/deploy logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests/conf logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests/conf logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export TEMP_DIR=/tmp/kuttl/pg/monitoring-pmm3 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ TEMP_DIR=/tmp/kuttl/pg/monitoring-pmm3 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export GIT_BRANCH=PR-1273 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ GIT_BRANCH=PR-1273 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export VERSION=PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ VERSION=PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export SKIP_TEST_WARNINGS=true logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ SKIP_TEST_WARNINGS=true logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ command -v oc logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export PG_VER=17 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ PG_VER=17 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export BUCKET=pg-operator-testing logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ BUCKET=pg-operator-testing logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export PGOV1_TAG=1.4.0 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ PGOV1_TAG=1.4.0 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export PGOV1_VER=14 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ PGOV1_VER=14 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export MINIO_VER=5.4.0 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ MINIO_VER=5.4.0 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export REGISTRY_NAME=docker.io logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ REGISTRY_NAME=docker.io logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export REGISTRY_NAME_FULL=docker.io/ logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ REGISTRY_NAME_FULL=docker.io/ logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | ++++ printenv logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | ++++ grep -E '^IMAGE' logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | ++++ awk -F= '{print $1}' logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | ++++ eval 'echo $IMAGE_BACKREST' logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++++ echo perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ var_value=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest17 == docker.io/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest17 == percona/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest17 == perconalab/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ echo IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ echo perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | ++++ eval 'echo $IMAGE_PMM3_CLIENT' logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++++ echo perconalab/pmm-client:3-dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ var_value=perconalab/pmm-client:3-dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/pmm-client:3-dev-latest == docker.io/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/pmm-client:3-dev-latest == percona/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/pmm-client:3-dev-latest == perconalab/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ new_value=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ echo IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ echo perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | ++++ eval 'echo $IMAGE' logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++++ echo perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ var_value=perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/percona-postgresql-operator:PR-1273-332364a9b == docker.io/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/percona-postgresql-operator:PR-1273-332364a9b == percona/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/percona-postgresql-operator:PR-1273-332364a9b == perconalab/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ new_value=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ echo IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | ++++ eval 'echo $IMAGE_UPGRADE' logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++++ echo perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ var_value=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == docker.io/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == percona/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == perconalab/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ echo IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | ++++ eval 'echo $IMAGE_BASE' logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++++ echo perconalab/percona-postgresql-operator logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ var_value=perconalab/percona-postgresql-operator logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/percona-postgresql-operator == docker.io/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/percona-postgresql-operator == percona/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/percona-postgresql-operator == perconalab/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ new_value=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ echo IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | ++++ eval 'echo $IMAGE_PMM3_SERVER' logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++++ echo perconalab/pmm-server:3-dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ var_value=perconalab/pmm-server:3-dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/pmm-server:3-dev-latest == docker.io/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/pmm-server:3-dev-latest == percona/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/pmm-server:3-dev-latest == perconalab/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ new_value=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ echo IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | ++++ eval 'echo $IMAGE_PGBOUNCER' logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++++ echo perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ var_value=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer17 == docker.io/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer17 == percona/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer17 == perconalab/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ echo IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | ++++ eval 'echo $IMAGE_POSTGRESQL' logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++++ echo perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ var_value=perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/percona-postgresql-operator:main-ppg17-postgres == docker.io/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/percona-postgresql-operator:main-ppg17-postgres == percona/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/percona-postgresql-operator:main-ppg17-postgres == perconalab/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ echo IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | ++++ eval 'echo $IMAGE_PMM_SERVER' logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++++ echo perconalab/pmm-server:dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ var_value=perconalab/pmm-server:dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/pmm-server:dev-latest == docker.io/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/pmm-server:dev-latest == percona/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/pmm-server:dev-latest == perconalab/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ new_value=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ echo IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | ++++ eval 'echo $IMAGE_PMM_CLIENT' logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++++ echo perconalab/pmm-client:dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ var_value=perconalab/pmm-client:dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/pmm-client:dev-latest == docker.io/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/pmm-client:dev-latest == percona/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ [[ perconalab/pmm-client:dev-latest == perconalab/* ]] logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ new_value=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ export IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ echo IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | ++++ which gdate logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | ++++ which date logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ date=/usr/sbin/date logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | ++++ which gsed logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | ++++ which sed logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | +++ sed=/usr/sbin/sed logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | ++ oc get projects logger.go:42: 16:39:17 | monitoring-pmm3/4-update-pmm-server-token | ++ kubectl -n kuttl-test-worthy-eagle get sts --selector=postgres-operator.crunchydata.com/instance-set=instance1 -o 'jsonpath={.items[*].metadata.name}' logger.go:42: 16:39:18 | monitoring-pmm3/4-update-pmm-server-token | + sts='monitoring-pmm3-instance1-g7tq monitoring-pmm3-instance1-pt7m monitoring-pmm3-instance1-rrwx' logger.go:42: 16:39:18 | monitoring-pmm3/4-update-pmm-server-token | + for st in $sts logger.go:42: 16:39:18 | monitoring-pmm3/4-update-pmm-server-token | + wait_for_generation sts/monitoring-pmm3-instance1-g7tq 2 logger.go:42: 16:39:18 | monitoring-pmm3/4-update-pmm-server-token | + local resource=sts/monitoring-pmm3-instance1-g7tq logger.go:42: 16:39:18 | monitoring-pmm3/4-update-pmm-server-token | + local target_generation=2 logger.go:42: 16:39:18 | monitoring-pmm3/4-update-pmm-server-token | + echo 'Waiting for sts/monitoring-pmm3-instance1-g7tq to reach generation 2...' logger.go:42: 16:39:18 | monitoring-pmm3/4-update-pmm-server-token | Waiting for sts/monitoring-pmm3-instance1-g7tq to reach generation 2... logger.go:42: 16:39:18 | monitoring-pmm3/4-update-pmm-server-token | + true logger.go:42: 16:39:18 | monitoring-pmm3/4-update-pmm-server-token | ++ kubectl -n kuttl-test-worthy-eagle get sts/monitoring-pmm3-instance1-g7tq -o 'jsonpath={.metadata.generation}' logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | + current_generation=2 logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | + '[' 2 -eq 2 ']' logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | + echo 'Resource sts/monitoring-pmm3-instance1-g7tq has reached generation 2.' logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | Resource sts/monitoring-pmm3-instance1-g7tq has reached generation 2. logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | + break logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | + for st in $sts logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | + wait_for_generation sts/monitoring-pmm3-instance1-pt7m 2 logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | + local resource=sts/monitoring-pmm3-instance1-pt7m logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | + local target_generation=2 logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | + echo 'Waiting for sts/monitoring-pmm3-instance1-pt7m to reach generation 2...' logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | Waiting for sts/monitoring-pmm3-instance1-pt7m to reach generation 2... logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | + true logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | ++ kubectl -n kuttl-test-worthy-eagle get sts/monitoring-pmm3-instance1-pt7m -o 'jsonpath={.metadata.generation}' logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | + current_generation=2 logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | + '[' 2 -eq 2 ']' logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | + echo 'Resource sts/monitoring-pmm3-instance1-pt7m has reached generation 2.' logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | Resource sts/monitoring-pmm3-instance1-pt7m has reached generation 2. logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | + break logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | + for st in $sts logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | + wait_for_generation sts/monitoring-pmm3-instance1-rrwx 2 logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | + local resource=sts/monitoring-pmm3-instance1-rrwx logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | + local target_generation=2 logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | + echo 'Waiting for sts/monitoring-pmm3-instance1-rrwx to reach generation 2...' logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | Waiting for sts/monitoring-pmm3-instance1-rrwx to reach generation 2... logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | + true logger.go:42: 16:39:19 | monitoring-pmm3/4-update-pmm-server-token | ++ kubectl -n kuttl-test-worthy-eagle get sts/monitoring-pmm3-instance1-rrwx -o 'jsonpath={.metadata.generation}' logger.go:42: 16:39:20 | monitoring-pmm3/4-update-pmm-server-token | + current_generation=2 logger.go:42: 16:39:20 | monitoring-pmm3/4-update-pmm-server-token | + '[' 2 -eq 2 ']' logger.go:42: 16:39:20 | monitoring-pmm3/4-update-pmm-server-token | + echo 'Resource sts/monitoring-pmm3-instance1-rrwx has reached generation 2.' logger.go:42: 16:39:20 | monitoring-pmm3/4-update-pmm-server-token | Resource sts/monitoring-pmm3-instance1-rrwx has reached generation 2. logger.go:42: 16:39:20 | monitoring-pmm3/4-update-pmm-server-token | + break logger.go:42: 16:39:20 | monitoring-pmm3/4-update-pmm-server-token | ++ generate_pmm3_server_token logger.go:42: 16:39:20 | monitoring-pmm3/4-update-pmm-server-token | ++ local key_name=2862 logger.go:42: 16:39:20 | monitoring-pmm3/4-update-pmm-server-token | ++ local ADMIN_PASSWORD logger.go:42: 16:39:20 | monitoring-pmm3/4-update-pmm-server-token | +++ kubectl -n kuttl-test-worthy-eagle get secret pmm-secret -o 'jsonpath={.data.PMM_ADMIN_PASSWORD}' logger.go:42: 16:39:20 | monitoring-pmm3/4-update-pmm-server-token | +++ base64 --decode logger.go:42: 16:39:20 | monitoring-pmm3/4-update-pmm-server-token | ++ ADMIN_PASSWORD='|Ue$9 k42fNX5&AA' logger.go:42: 16:39:20 | monitoring-pmm3/4-update-pmm-server-token | ++ [[ -z |Ue$9 k42fNX5&AA ]] logger.go:42: 16:39:20 | monitoring-pmm3/4-update-pmm-server-token | ++ local create_response create_status_code create_json_response logger.go:42: 16:39:20 | monitoring-pmm3/4-update-pmm-server-token | ++++ get_service_ip monitoring-service logger.go:42: 16:39:20 | monitoring-pmm3/4-update-pmm-server-token | ++++ local service=monitoring-service logger.go:42: 16:39:20 | monitoring-pmm3/4-update-pmm-server-token | ++++ kubectl get service/monitoring-service -n kuttl-test-worthy-eagle -o 'jsonpath={.spec.type}' logger.go:42: 16:39:20 | monitoring-pmm3/4-update-pmm-server-token | ++++ grep -q NotFound logger.go:42: 16:39:21 | monitoring-pmm3/4-update-pmm-server-token | +++++ kubectl get service/monitoring-service -n kuttl-test-worthy-eagle -o 'jsonpath={.spec.type}' logger.go:42: 16:39:21 | monitoring-pmm3/4-update-pmm-server-token | ++++ '[' LoadBalancer = ClusterIP ']' logger.go:42: 16:39:21 | monitoring-pmm3/4-update-pmm-server-token | ++++ kubectl get service/monitoring-service -n kuttl-test-worthy-eagle -o 'jsonpath={.status.loadBalancer.ingress[]}' logger.go:42: 16:39:21 | monitoring-pmm3/4-update-pmm-server-token | ++++ egrep -q 'hostname|ip' logger.go:42: 16:39:21 | monitoring-pmm3/4-update-pmm-server-token | egrep: warning: egrep is obsolescent; using grep -E logger.go:42: 16:39:22 | monitoring-pmm3/4-update-pmm-server-token | ++++ kubectl get service/monitoring-service -n kuttl-test-worthy-eagle -o 'jsonpath={.status.loadBalancer.ingress[].ip}' logger.go:42: 16:39:22 | monitoring-pmm3/4-update-pmm-server-token | ++++ kubectl get service/monitoring-service -n kuttl-test-worthy-eagle -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' logger.go:42: 16:39:22 | monitoring-pmm3/4-update-pmm-server-token | +++ curl --insecure -s -X POST -H 'Content-Type: application/json' -H 'Accept: application/json' -d '{"name":"2862", "role":"Admin", "isDisabled":false}' --user 'admin:|Ue$9 k42fNX5&AA' https://34.134.80.59/graph/api/serviceaccounts -w '\n%{http_code}' logger.go:42: 16:39:23 | monitoring-pmm3/4-update-pmm-server-token | ++ create_response='{"id":2,"uid":"aez5qxxgjrabkd","name":"2862","login":"sa-1-2862","orgId":1,"isDisabled":false,"role":"Admin","tokens":0,"avatarUrl":""} logger.go:42: 16:39:23 | monitoring-pmm3/4-update-pmm-server-token | 201' logger.go:42: 16:39:23 | monitoring-pmm3/4-update-pmm-server-token | +++ echo '{"id":2,"uid":"aez5qxxgjrabkd","name":"2862","login":"sa-1-2862","orgId":1,"isDisabled":false,"role":"Admin","tokens":0,"avatarUrl":""} logger.go:42: 16:39:23 | monitoring-pmm3/4-update-pmm-server-token | 201' logger.go:42: 16:39:23 | monitoring-pmm3/4-update-pmm-server-token | +++ tail -n1 logger.go:42: 16:39:23 | monitoring-pmm3/4-update-pmm-server-token | ++ create_status_code=201 logger.go:42: 16:39:23 | monitoring-pmm3/4-update-pmm-server-token | +++ echo '{"id":2,"uid":"aez5qxxgjrabkd","name":"2862","login":"sa-1-2862","orgId":1,"isDisabled":false,"role":"Admin","tokens":0,"avatarUrl":""} logger.go:42: 16:39:23 | monitoring-pmm3/4-update-pmm-server-token | 201' logger.go:42: 16:39:23 | monitoring-pmm3/4-update-pmm-server-token | +++ sed '$ d' logger.go:42: 16:39:23 | monitoring-pmm3/4-update-pmm-server-token | ++ create_json_response='{"id":2,"uid":"aez5qxxgjrabkd","name":"2862","login":"sa-1-2862","orgId":1,"isDisabled":false,"role":"Admin","tokens":0,"avatarUrl":""}' logger.go:42: 16:39:23 | monitoring-pmm3/4-update-pmm-server-token | ++ [[ 201 -ne 201 ]] logger.go:42: 16:39:23 | monitoring-pmm3/4-update-pmm-server-token | ++ local service_account_id logger.go:42: 16:39:23 | monitoring-pmm3/4-update-pmm-server-token | +++ echo '{"id":2,"uid":"aez5qxxgjrabkd","name":"2862","login":"sa-1-2862","orgId":1,"isDisabled":false,"role":"Admin","tokens":0,"avatarUrl":""}' logger.go:42: 16:39:23 | monitoring-pmm3/4-update-pmm-server-token | +++ jq -r .id logger.go:42: 16:39:23 | monitoring-pmm3/4-update-pmm-server-token | ++ service_account_id=2 logger.go:42: 16:39:23 | monitoring-pmm3/4-update-pmm-server-token | ++ [[ -z 2 ]] logger.go:42: 16:39:23 | monitoring-pmm3/4-update-pmm-server-token | ++ [[ 2 == \n\u\l\l ]] logger.go:42: 16:39:23 | monitoring-pmm3/4-update-pmm-server-token | ++ local token_response token_status_code token_json_response logger.go:42: 16:39:23 | monitoring-pmm3/4-update-pmm-server-token | ++++ get_service_ip monitoring-service logger.go:42: 16:39:23 | monitoring-pmm3/4-update-pmm-server-token | ++++ local service=monitoring-service logger.go:42: 16:39:23 | monitoring-pmm3/4-update-pmm-server-token | ++++ grep -q NotFound logger.go:42: 16:39:23 | monitoring-pmm3/4-update-pmm-server-token | ++++ kubectl get service/monitoring-service -n kuttl-test-worthy-eagle -o 'jsonpath={.spec.type}' logger.go:42: 16:39:23 | monitoring-pmm3/4-update-pmm-server-token | +++++ kubectl get service/monitoring-service -n kuttl-test-worthy-eagle -o 'jsonpath={.spec.type}' logger.go:42: 16:39:24 | monitoring-pmm3/4-update-pmm-server-token | ++++ '[' LoadBalancer = ClusterIP ']' logger.go:42: 16:39:24 | monitoring-pmm3/4-update-pmm-server-token | ++++ kubectl get service/monitoring-service -n kuttl-test-worthy-eagle -o 'jsonpath={.status.loadBalancer.ingress[]}' logger.go:42: 16:39:24 | monitoring-pmm3/4-update-pmm-server-token | ++++ egrep -q 'hostname|ip' logger.go:42: 16:39:24 | monitoring-pmm3/4-update-pmm-server-token | egrep: warning: egrep is obsolescent; using grep -E logger.go:42: 16:39:24 | monitoring-pmm3/4-update-pmm-server-token | ++++ kubectl get service/monitoring-service -n kuttl-test-worthy-eagle -o 'jsonpath={.status.loadBalancer.ingress[].ip}' logger.go:42: 16:39:25 | monitoring-pmm3/4-update-pmm-server-token | ++++ kubectl get service/monitoring-service -n kuttl-test-worthy-eagle -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' logger.go:42: 16:39:25 | monitoring-pmm3/4-update-pmm-server-token | +++ curl --insecure -s -X POST -H 'Content-Type: application/json' -d '{"name":"2862"}' --user 'admin:|Ue$9 k42fNX5&AA' https://34.134.80.59/graph/api/serviceaccounts/2/tokens -w '\n%{http_code}' logger.go:42: 16:39:26 | monitoring-pmm3/4-update-pmm-server-token | ++ token_response='{"id":1,"name":"2862","key":"glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9"} logger.go:42: 16:39:26 | monitoring-pmm3/4-update-pmm-server-token | 200' logger.go:42: 16:39:26 | monitoring-pmm3/4-update-pmm-server-token | +++ echo '{"id":1,"name":"2862","key":"glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9"} logger.go:42: 16:39:26 | monitoring-pmm3/4-update-pmm-server-token | 200' logger.go:42: 16:39:26 | monitoring-pmm3/4-update-pmm-server-token | +++ tail -n1 logger.go:42: 16:39:26 | monitoring-pmm3/4-update-pmm-server-token | ++ token_status_code=200 logger.go:42: 16:39:26 | monitoring-pmm3/4-update-pmm-server-token | +++ echo '{"id":1,"name":"2862","key":"glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9"} logger.go:42: 16:39:26 | monitoring-pmm3/4-update-pmm-server-token | 200' logger.go:42: 16:39:26 | monitoring-pmm3/4-update-pmm-server-token | +++ sed '$ d' logger.go:42: 16:39:26 | monitoring-pmm3/4-update-pmm-server-token | ++ token_json_response='{"id":1,"name":"2862","key":"glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9"}' logger.go:42: 16:39:26 | monitoring-pmm3/4-update-pmm-server-token | ++ [[ 200 -ne 200 ]] logger.go:42: 16:39:26 | monitoring-pmm3/4-update-pmm-server-token | ++ echo '{"id":1,"name":"2862","key":"glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9"}' logger.go:42: 16:39:26 | monitoring-pmm3/4-update-pmm-server-token | ++ jq -r .key logger.go:42: 16:39:26 | monitoring-pmm3/4-update-pmm-server-token | + token=glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9 logger.go:42: 16:39:26 | monitoring-pmm3/4-update-pmm-server-token | + [[ -n glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9 ]] logger.go:42: 16:39:26 | monitoring-pmm3/4-update-pmm-server-token | + [[ glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9 != null ]] logger.go:42: 16:39:26 | monitoring-pmm3/4-update-pmm-server-token | + kubectl -n kuttl-test-worthy-eagle patch secret monitoring-pmm3-pmm-secret --type merge --patch '{"stringData": {"PMM_SERVER_TOKEN": "glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9"}}' logger.go:42: 16:39:26 | monitoring-pmm3/4-update-pmm-server-token | secret/monitoring-pmm3-pmm-secret patched logger.go:42: 16:39:26 | monitoring-pmm3/4-update-pmm-server-token | + sleep 10 logger.go:42: 16:39:36 | monitoring-pmm3/4-update-pmm-server-token | ++ kubectl -n kuttl-test-worthy-eagle get sts --selector=postgres-operator.crunchydata.com/instance-set=instance1 -o 'jsonpath={.items[*].metadata.name}' logger.go:42: 16:39:37 | monitoring-pmm3/4-update-pmm-server-token | + sts='monitoring-pmm3-instance1-g7tq monitoring-pmm3-instance1-pt7m monitoring-pmm3-instance1-rrwx' logger.go:42: 16:39:37 | monitoring-pmm3/4-update-pmm-server-token | + for st in $sts logger.go:42: 16:39:37 | monitoring-pmm3/4-update-pmm-server-token | + wait_for_generation sts/monitoring-pmm3-instance1-g7tq 3 logger.go:42: 16:39:37 | monitoring-pmm3/4-update-pmm-server-token | + local resource=sts/monitoring-pmm3-instance1-g7tq logger.go:42: 16:39:37 | monitoring-pmm3/4-update-pmm-server-token | + local target_generation=3 logger.go:42: 16:39:37 | monitoring-pmm3/4-update-pmm-server-token | + echo 'Waiting for sts/monitoring-pmm3-instance1-g7tq to reach generation 3...' logger.go:42: 16:39:37 | monitoring-pmm3/4-update-pmm-server-token | Waiting for sts/monitoring-pmm3-instance1-g7tq to reach generation 3... logger.go:42: 16:39:37 | monitoring-pmm3/4-update-pmm-server-token | + true logger.go:42: 16:39:37 | monitoring-pmm3/4-update-pmm-server-token | ++ kubectl -n kuttl-test-worthy-eagle get sts/monitoring-pmm3-instance1-g7tq -o 'jsonpath={.metadata.generation}' logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | + current_generation=3 logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | + '[' 3 -eq 3 ']' logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | + echo 'Resource sts/monitoring-pmm3-instance1-g7tq has reached generation 3.' logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | Resource sts/monitoring-pmm3-instance1-g7tq has reached generation 3. logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | + break logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | + for st in $sts logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | + wait_for_generation sts/monitoring-pmm3-instance1-pt7m 3 logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | + local resource=sts/monitoring-pmm3-instance1-pt7m logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | + local target_generation=3 logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | + echo 'Waiting for sts/monitoring-pmm3-instance1-pt7m to reach generation 3...' logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | Waiting for sts/monitoring-pmm3-instance1-pt7m to reach generation 3... logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | + true logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | ++ kubectl -n kuttl-test-worthy-eagle get sts/monitoring-pmm3-instance1-pt7m -o 'jsonpath={.metadata.generation}' logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | + current_generation=3 logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | + '[' 3 -eq 3 ']' logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | + echo 'Resource sts/monitoring-pmm3-instance1-pt7m has reached generation 3.' logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | Resource sts/monitoring-pmm3-instance1-pt7m has reached generation 3. logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | + break logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | + for st in $sts logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | + wait_for_generation sts/monitoring-pmm3-instance1-rrwx 3 logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | + local resource=sts/monitoring-pmm3-instance1-rrwx logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | + local target_generation=3 logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | + echo 'Waiting for sts/monitoring-pmm3-instance1-rrwx to reach generation 3...' logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | Waiting for sts/monitoring-pmm3-instance1-rrwx to reach generation 3... logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | + true logger.go:42: 16:39:38 | monitoring-pmm3/4-update-pmm-server-token | ++ kubectl -n kuttl-test-worthy-eagle get sts/monitoring-pmm3-instance1-rrwx -o 'jsonpath={.metadata.generation}' logger.go:42: 16:39:39 | monitoring-pmm3/4-update-pmm-server-token | + current_generation=3 logger.go:42: 16:39:39 | monitoring-pmm3/4-update-pmm-server-token | + '[' 3 -eq 3 ']' logger.go:42: 16:39:39 | monitoring-pmm3/4-update-pmm-server-token | + echo 'Resource sts/monitoring-pmm3-instance1-rrwx has reached generation 3.' logger.go:42: 16:39:39 | monitoring-pmm3/4-update-pmm-server-token | Resource sts/monitoring-pmm3-instance1-rrwx has reached generation 3. logger.go:42: 16:39:39 | monitoring-pmm3/4-update-pmm-server-token | + break logger.go:42: 16:39:39 | monitoring-pmm3/4-update-pmm-server-token | + sleep 25 logger.go:42: 16:40:36 | monitoring-pmm3/4-update-pmm-server-token | test step completed 4-update-pmm-server-token logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | starting test step 5-check-qan logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | running command: [sh -c set -o errexit set -o xtrace source ../../functions token=$(kubectl get -n "${NAMESPACE}" secret monitoring-pmm3-pmm-secret --template='{{.data.PMM_SERVER_TOKEN | base64decode}}') instance=$(kubectl get -n "${NAMESPACE}" pod -l postgres-operator.crunchydata.com/instance-set=instance1 -o 'jsonpath={.items[].metadata.name}') get_metric_values node_boot_time_seconds ${NAMESPACE}-${instance} ${token} get_metric_values patroni_postgres_running ${NAMESPACE}-${instance} ${token} get_qan20_values_pmm3 ${NAMESPACE}-${instance} ${token}] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | + source ../../functions logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ realpath ../../.. logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | ++ CERT_MANAGER_VER=1.18.2 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | ++++ pwd logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests/tests/monitoring-pmm3 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | ++ test_name=monitoring-pmm3 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests/vars.sh logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/deploy logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/deploy logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests/conf logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/e2e-tests/conf logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export TEMP_DIR=/tmp/kuttl/pg/monitoring-pmm3 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ TEMP_DIR=/tmp/kuttl/pg/monitoring-pmm3 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export GIT_BRANCH=PR-1273 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ GIT_BRANCH=PR-1273 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export VERSION=PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ VERSION=PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export SKIP_TEST_WARNINGS=true logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ SKIP_TEST_WARNINGS=true logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ command -v oc logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export PG_VER=17 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ PG_VER=17 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export BUCKET=pg-operator-testing logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ BUCKET=pg-operator-testing logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export PGOV1_TAG=1.4.0 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ PGOV1_TAG=1.4.0 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export PGOV1_VER=14 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ PGOV1_VER=14 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export MINIO_VER=5.4.0 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ MINIO_VER=5.4.0 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export REGISTRY_NAME=docker.io logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ REGISTRY_NAME=docker.io logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export REGISTRY_NAME_FULL=docker.io/ logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ REGISTRY_NAME_FULL=docker.io/ logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | ++++ printenv logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | ++++ grep -E '^IMAGE' logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | ++++ awk -F= '{print $1}' logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | ++++ eval 'echo $IMAGE_BACKREST' logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++++ echo perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ var_value=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest17 == docker.io/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest17 == percona/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest17 == perconalab/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ echo IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ echo perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | ++++ eval 'echo $IMAGE_PMM3_CLIENT' logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++++ echo perconalab/pmm-client:3-dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ var_value=perconalab/pmm-client:3-dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/pmm-client:3-dev-latest == docker.io/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/pmm-client:3-dev-latest == percona/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/pmm-client:3-dev-latest == perconalab/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ new_value=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ echo IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3-dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ echo perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | ++++ eval 'echo $IMAGE' logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++++ echo perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ var_value=perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/percona-postgresql-operator:PR-1273-332364a9b == docker.io/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/percona-postgresql-operator:PR-1273-332364a9b == percona/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/percona-postgresql-operator:PR-1273-332364a9b == perconalab/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ new_value=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ echo IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | ++++ eval 'echo $IMAGE_UPGRADE' logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++++ echo perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ var_value=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == docker.io/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == percona/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == perconalab/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ echo IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | ++++ eval 'echo $IMAGE_BASE' logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++++ echo perconalab/percona-postgresql-operator logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ var_value=perconalab/percona-postgresql-operator logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/percona-postgresql-operator == docker.io/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/percona-postgresql-operator == percona/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/percona-postgresql-operator == perconalab/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ new_value=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ echo IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | ++++ eval 'echo $IMAGE_PMM3_SERVER' logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++++ echo perconalab/pmm-server:3-dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ var_value=perconalab/pmm-server:3-dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/pmm-server:3-dev-latest == docker.io/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/pmm-server:3-dev-latest == percona/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/pmm-server:3-dev-latest == perconalab/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ new_value=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ echo IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3-dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | ++++ eval 'echo $IMAGE_PGBOUNCER' logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++++ echo perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ var_value=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer17 == docker.io/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer17 == percona/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer17 == perconalab/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ echo IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | ++++ eval 'echo $IMAGE_POSTGRESQL' logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++++ echo perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ var_value=perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/percona-postgresql-operator:main-ppg17-postgres == docker.io/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/percona-postgresql-operator:main-ppg17-postgres == percona/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/percona-postgresql-operator:main-ppg17-postgres == perconalab/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ echo IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | ++++ eval 'echo $IMAGE_PMM_SERVER' logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++++ echo perconalab/pmm-server:dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ var_value=perconalab/pmm-server:dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/pmm-server:dev-latest == docker.io/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/pmm-server:dev-latest == percona/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/pmm-server:dev-latest == perconalab/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ new_value=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ echo IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | ++++ eval 'echo $IMAGE_PMM_CLIENT' logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++++ echo perconalab/pmm-client:dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ var_value=perconalab/pmm-client:dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/pmm-client:dev-latest == docker.io/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/pmm-client:dev-latest == percona/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ [[ perconalab/pmm-client:dev-latest == perconalab/* ]] logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ new_value=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ export IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ echo IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | ++++ which gdate logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | ++++ which date logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ date=/usr/sbin/date logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | ++++ which gsed logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1273/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | ++++ which sed logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | +++ sed=/usr/sbin/sed logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | ++ oc get projects logger.go:42: 16:40:36 | monitoring-pmm3/5-check-qan | ++ kubectl get -n kuttl-test-worthy-eagle secret monitoring-pmm3-pmm-secret '--template={{.data.PMM_SERVER_TOKEN | base64decode}}' logger.go:42: 16:40:37 | monitoring-pmm3/5-check-qan | + token=glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9 logger.go:42: 16:40:37 | monitoring-pmm3/5-check-qan | ++ kubectl get -n kuttl-test-worthy-eagle pod -l postgres-operator.crunchydata.com/instance-set=instance1 -o 'jsonpath={.items[].metadata.name}' logger.go:42: 16:40:37 | monitoring-pmm3/5-check-qan | + instance=monitoring-pmm3-instance1-g7tq-0 logger.go:42: 16:40:37 | monitoring-pmm3/5-check-qan | + get_metric_values node_boot_time_seconds kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0 glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9 logger.go:42: 16:40:37 | monitoring-pmm3/5-check-qan | + local metric=node_boot_time_seconds logger.go:42: 16:40:37 | monitoring-pmm3/5-check-qan | + local instance=kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0 logger.go:42: 16:40:37 | monitoring-pmm3/5-check-qan | + local token=glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9 logger.go:42: 16:40:37 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s -d '-5 minute' logger.go:42: 16:40:37 | monitoring-pmm3/5-check-qan | + local start=1758818137 logger.go:42: 16:40:37 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s logger.go:42: 16:40:37 | monitoring-pmm3/5-check-qan | + local end=1758818437 logger.go:42: 16:40:37 | monitoring-pmm3/5-check-qan | ++ get_service_ip monitoring-service logger.go:42: 16:40:37 | monitoring-pmm3/5-check-qan | ++ local service=monitoring-service logger.go:42: 16:40:37 | monitoring-pmm3/5-check-qan | ++ kubectl get service/monitoring-service -n kuttl-test-worthy-eagle -o 'jsonpath={.spec.type}' logger.go:42: 16:40:37 | monitoring-pmm3/5-check-qan | ++ grep -q NotFound logger.go:42: 16:40:38 | monitoring-pmm3/5-check-qan | +++ kubectl get service/monitoring-service -n kuttl-test-worthy-eagle -o 'jsonpath={.spec.type}' logger.go:42: 16:40:38 | monitoring-pmm3/5-check-qan | ++ '[' LoadBalancer = ClusterIP ']' logger.go:42: 16:40:38 | monitoring-pmm3/5-check-qan | ++ kubectl get service/monitoring-service -n kuttl-test-worthy-eagle -o 'jsonpath={.status.loadBalancer.ingress[]}' logger.go:42: 16:40:38 | monitoring-pmm3/5-check-qan | ++ egrep -q 'hostname|ip' logger.go:42: 16:40:38 | monitoring-pmm3/5-check-qan | egrep: warning: egrep is obsolescent; using grep -E logger.go:42: 16:40:39 | monitoring-pmm3/5-check-qan | ++ kubectl get service/monitoring-service -n kuttl-test-worthy-eagle -o 'jsonpath={.status.loadBalancer.ingress[].ip}' logger.go:42: 16:40:39 | monitoring-pmm3/5-check-qan | ++ kubectl get service/monitoring-service -n kuttl-test-worthy-eagle -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' logger.go:42: 16:40:40 | monitoring-pmm3/5-check-qan | + local endpoint=34.134.80.59 logger.go:42: 16:40:40 | monitoring-pmm3/5-check-qan | + local wait_count=20 logger.go:42: 16:40:40 | monitoring-pmm3/5-check-qan | + local retry=0 logger.go:42: 16:40:40 | monitoring-pmm3/5-check-qan | ++ curl -s -k -H 'Authorization: Bearer glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9' 'https://34.134.80.59/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28node_boot_time_seconds%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7d%20or%20node_boot_time_seconds%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7D%29&start=1758818137&end=1758818437&step=60' logger.go:42: 16:40:40 | monitoring-pmm3/5-check-qan | ++ jq '.data.result[0].values[][1]' logger.go:42: 16:40:40 | monitoring-pmm3/5-check-qan | ++ grep '^"[0-9]' logger.go:42: 16:40:40 | monitoring-pmm3/5-check-qan | + [[ -n "1758817415" ]] logger.go:42: 16:40:40 | monitoring-pmm3/5-check-qan | + get_metric_values patroni_postgres_running kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0 glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9 logger.go:42: 16:40:40 | monitoring-pmm3/5-check-qan | + local metric=patroni_postgres_running logger.go:42: 16:40:40 | monitoring-pmm3/5-check-qan | + local instance=kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0 logger.go:42: 16:40:40 | monitoring-pmm3/5-check-qan | + local token=glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9 logger.go:42: 16:40:40 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s -d '-5 minute' logger.go:42: 16:40:40 | monitoring-pmm3/5-check-qan | + local start=1758818140 logger.go:42: 16:40:40 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s logger.go:42: 16:40:40 | monitoring-pmm3/5-check-qan | + local end=1758818440 logger.go:42: 16:40:40 | monitoring-pmm3/5-check-qan | ++ get_service_ip monitoring-service logger.go:42: 16:40:40 | monitoring-pmm3/5-check-qan | ++ local service=monitoring-service logger.go:42: 16:40:40 | monitoring-pmm3/5-check-qan | ++ kubectl get service/monitoring-service -n kuttl-test-worthy-eagle -o 'jsonpath={.spec.type}' logger.go:42: 16:40:40 | monitoring-pmm3/5-check-qan | ++ grep -q NotFound logger.go:42: 16:40:41 | monitoring-pmm3/5-check-qan | +++ kubectl get service/monitoring-service -n kuttl-test-worthy-eagle -o 'jsonpath={.spec.type}' logger.go:42: 16:40:41 | monitoring-pmm3/5-check-qan | ++ '[' LoadBalancer = ClusterIP ']' logger.go:42: 16:40:41 | monitoring-pmm3/5-check-qan | ++ egrep -q 'hostname|ip' logger.go:42: 16:40:41 | monitoring-pmm3/5-check-qan | ++ kubectl get service/monitoring-service -n kuttl-test-worthy-eagle -o 'jsonpath={.status.loadBalancer.ingress[]}' logger.go:42: 16:40:41 | monitoring-pmm3/5-check-qan | egrep: warning: egrep is obsolescent; using grep -E logger.go:42: 16:40:42 | monitoring-pmm3/5-check-qan | ++ kubectl get service/monitoring-service -n kuttl-test-worthy-eagle -o 'jsonpath={.status.loadBalancer.ingress[].ip}' logger.go:42: 16:40:42 | monitoring-pmm3/5-check-qan | ++ kubectl get service/monitoring-service -n kuttl-test-worthy-eagle -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' logger.go:42: 16:40:42 | monitoring-pmm3/5-check-qan | + local endpoint=34.134.80.59 logger.go:42: 16:40:42 | monitoring-pmm3/5-check-qan | + local wait_count=20 logger.go:42: 16:40:42 | monitoring-pmm3/5-check-qan | + local retry=0 logger.go:42: 16:40:42 | monitoring-pmm3/5-check-qan | ++ curl -s -k -H 'Authorization: Bearer glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9' 'https://34.134.80.59/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7d%20or%20patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7D%29&start=1758818140&end=1758818440&step=60' logger.go:42: 16:40:42 | monitoring-pmm3/5-check-qan | ++ grep '^"[0-9]' logger.go:42: 16:40:42 | monitoring-pmm3/5-check-qan | ++ jq '.data.result[0].values[][1]' logger.go:42: 16:40:43 | monitoring-pmm3/5-check-qan | jq: error (at :0): Cannot iterate over null (null) logger.go:42: 16:40:43 | monitoring-pmm3/5-check-qan | + [[ -n '' ]] logger.go:42: 16:40:43 | monitoring-pmm3/5-check-qan | + sleep 2 logger.go:42: 16:40:45 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s -d '-5 minute' logger.go:42: 16:40:45 | monitoring-pmm3/5-check-qan | + local start=1758818145 logger.go:42: 16:40:45 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s logger.go:42: 16:40:45 | monitoring-pmm3/5-check-qan | + local end=1758818445 logger.go:42: 16:40:45 | monitoring-pmm3/5-check-qan | + let retry+=1 logger.go:42: 16:40:45 | monitoring-pmm3/5-check-qan | + [[ 1 -ge 20 ]] logger.go:42: 16:40:45 | monitoring-pmm3/5-check-qan | ++ curl -s -k -H 'Authorization: Bearer glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9' 'https://34.134.80.59/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7d%20or%20patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7D%29&start=1758818145&end=1758818445&step=60' logger.go:42: 16:40:45 | monitoring-pmm3/5-check-qan | ++ jq '.data.result[0].values[][1]' logger.go:42: 16:40:45 | monitoring-pmm3/5-check-qan | ++ grep '^"[0-9]' logger.go:42: 16:40:45 | monitoring-pmm3/5-check-qan | jq: error (at :0): Cannot iterate over null (null) logger.go:42: 16:40:45 | monitoring-pmm3/5-check-qan | + [[ -n '' ]] logger.go:42: 16:40:45 | monitoring-pmm3/5-check-qan | + sleep 2 logger.go:42: 16:40:47 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s -d '-5 minute' logger.go:42: 16:40:47 | monitoring-pmm3/5-check-qan | + local start=1758818147 logger.go:42: 16:40:47 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s logger.go:42: 16:40:47 | monitoring-pmm3/5-check-qan | + local end=1758818447 logger.go:42: 16:40:47 | monitoring-pmm3/5-check-qan | + let retry+=1 logger.go:42: 16:40:47 | monitoring-pmm3/5-check-qan | + [[ 2 -ge 20 ]] logger.go:42: 16:40:47 | monitoring-pmm3/5-check-qan | ++ jq '.data.result[0].values[][1]' logger.go:42: 16:40:47 | monitoring-pmm3/5-check-qan | ++ curl -s -k -H 'Authorization: Bearer glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9' 'https://34.134.80.59/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7d%20or%20patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7D%29&start=1758818147&end=1758818447&step=60' logger.go:42: 16:40:47 | monitoring-pmm3/5-check-qan | ++ grep '^"[0-9]' logger.go:42: 16:40:48 | monitoring-pmm3/5-check-qan | jq: error (at :0): Cannot iterate over null (null) logger.go:42: 16:40:48 | monitoring-pmm3/5-check-qan | + [[ -n '' ]] logger.go:42: 16:40:48 | monitoring-pmm3/5-check-qan | + sleep 2 logger.go:42: 16:40:50 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s -d '-5 minute' logger.go:42: 16:40:50 | monitoring-pmm3/5-check-qan | + local start=1758818150 logger.go:42: 16:40:50 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s logger.go:42: 16:40:50 | monitoring-pmm3/5-check-qan | + local end=1758818450 logger.go:42: 16:40:50 | monitoring-pmm3/5-check-qan | + let retry+=1 logger.go:42: 16:40:50 | monitoring-pmm3/5-check-qan | + [[ 3 -ge 20 ]] logger.go:42: 16:40:50 | monitoring-pmm3/5-check-qan | ++ jq '.data.result[0].values[][1]' logger.go:42: 16:40:50 | monitoring-pmm3/5-check-qan | ++ curl -s -k -H 'Authorization: Bearer glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9' 'https://34.134.80.59/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7d%20or%20patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7D%29&start=1758818150&end=1758818450&step=60' logger.go:42: 16:40:50 | monitoring-pmm3/5-check-qan | ++ grep '^"[0-9]' logger.go:42: 16:40:50 | monitoring-pmm3/5-check-qan | jq: error (at :0): Cannot iterate over null (null) logger.go:42: 16:40:50 | monitoring-pmm3/5-check-qan | + [[ -n '' ]] logger.go:42: 16:40:50 | monitoring-pmm3/5-check-qan | + sleep 2 logger.go:42: 16:40:52 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s -d '-5 minute' logger.go:42: 16:40:52 | monitoring-pmm3/5-check-qan | + local start=1758818152 logger.go:42: 16:40:52 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s logger.go:42: 16:40:52 | monitoring-pmm3/5-check-qan | + local end=1758818452 logger.go:42: 16:40:52 | monitoring-pmm3/5-check-qan | + let retry+=1 logger.go:42: 16:40:52 | monitoring-pmm3/5-check-qan | + [[ 4 -ge 20 ]] logger.go:42: 16:40:52 | monitoring-pmm3/5-check-qan | ++ curl -s -k -H 'Authorization: Bearer glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9' 'https://34.134.80.59/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7d%20or%20patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7D%29&start=1758818152&end=1758818452&step=60' logger.go:42: 16:40:52 | monitoring-pmm3/5-check-qan | ++ grep '^"[0-9]' logger.go:42: 16:40:52 | monitoring-pmm3/5-check-qan | ++ jq '.data.result[0].values[][1]' logger.go:42: 16:40:52 | monitoring-pmm3/5-check-qan | jq: error (at :0): Cannot iterate over null (null) logger.go:42: 16:40:52 | monitoring-pmm3/5-check-qan | + [[ -n '' ]] logger.go:42: 16:40:52 | monitoring-pmm3/5-check-qan | + sleep 2 logger.go:42: 16:40:54 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s -d '-5 minute' logger.go:42: 16:40:54 | monitoring-pmm3/5-check-qan | + local start=1758818154 logger.go:42: 16:40:54 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s logger.go:42: 16:40:54 | monitoring-pmm3/5-check-qan | + local end=1758818454 logger.go:42: 16:40:54 | monitoring-pmm3/5-check-qan | + let retry+=1 logger.go:42: 16:40:54 | monitoring-pmm3/5-check-qan | + [[ 5 -ge 20 ]] logger.go:42: 16:40:54 | monitoring-pmm3/5-check-qan | ++ curl -s -k -H 'Authorization: Bearer glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9' 'https://34.134.80.59/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7d%20or%20patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7D%29&start=1758818154&end=1758818454&step=60' logger.go:42: 16:40:54 | monitoring-pmm3/5-check-qan | ++ jq '.data.result[0].values[][1]' logger.go:42: 16:40:54 | monitoring-pmm3/5-check-qan | ++ grep '^"[0-9]' logger.go:42: 16:40:55 | monitoring-pmm3/5-check-qan | jq: error (at :0): Cannot iterate over null (null) logger.go:42: 16:40:55 | monitoring-pmm3/5-check-qan | + [[ -n '' ]] logger.go:42: 16:40:55 | monitoring-pmm3/5-check-qan | + sleep 2 logger.go:42: 16:40:57 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s -d '-5 minute' logger.go:42: 16:40:57 | monitoring-pmm3/5-check-qan | + local start=1758818157 logger.go:42: 16:40:57 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s logger.go:42: 16:40:57 | monitoring-pmm3/5-check-qan | + local end=1758818457 logger.go:42: 16:40:57 | monitoring-pmm3/5-check-qan | + let retry+=1 logger.go:42: 16:40:57 | monitoring-pmm3/5-check-qan | + [[ 6 -ge 20 ]] logger.go:42: 16:40:57 | monitoring-pmm3/5-check-qan | ++ curl -s -k -H 'Authorization: Bearer glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9' 'https://34.134.80.59/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7d%20or%20patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7D%29&start=1758818157&end=1758818457&step=60' logger.go:42: 16:40:57 | monitoring-pmm3/5-check-qan | ++ jq '.data.result[0].values[][1]' logger.go:42: 16:40:57 | monitoring-pmm3/5-check-qan | ++ grep '^"[0-9]' logger.go:42: 16:40:57 | monitoring-pmm3/5-check-qan | jq: error (at :0): Cannot iterate over null (null) logger.go:42: 16:40:57 | monitoring-pmm3/5-check-qan | + [[ -n '' ]] logger.go:42: 16:40:57 | monitoring-pmm3/5-check-qan | + sleep 2 logger.go:42: 16:40:59 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s -d '-5 minute' logger.go:42: 16:40:59 | monitoring-pmm3/5-check-qan | + local start=1758818159 logger.go:42: 16:40:59 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s logger.go:42: 16:40:59 | monitoring-pmm3/5-check-qan | + local end=1758818459 logger.go:42: 16:40:59 | monitoring-pmm3/5-check-qan | + let retry+=1 logger.go:42: 16:40:59 | monitoring-pmm3/5-check-qan | + [[ 7 -ge 20 ]] logger.go:42: 16:40:59 | monitoring-pmm3/5-check-qan | ++ curl -s -k -H 'Authorization: Bearer glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9' 'https://34.134.80.59/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7d%20or%20patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7D%29&start=1758818159&end=1758818459&step=60' logger.go:42: 16:40:59 | monitoring-pmm3/5-check-qan | ++ jq '.data.result[0].values[][1]' logger.go:42: 16:40:59 | monitoring-pmm3/5-check-qan | ++ grep '^"[0-9]' logger.go:42: 16:41:00 | monitoring-pmm3/5-check-qan | jq: error (at :0): Cannot iterate over null (null) logger.go:42: 16:41:00 | monitoring-pmm3/5-check-qan | + [[ -n '' ]] logger.go:42: 16:41:00 | monitoring-pmm3/5-check-qan | + sleep 2 logger.go:42: 16:41:02 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s -d '-5 minute' logger.go:42: 16:41:02 | monitoring-pmm3/5-check-qan | + local start=1758818162 logger.go:42: 16:41:02 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s logger.go:42: 16:41:02 | monitoring-pmm3/5-check-qan | + local end=1758818462 logger.go:42: 16:41:02 | monitoring-pmm3/5-check-qan | + let retry+=1 logger.go:42: 16:41:02 | monitoring-pmm3/5-check-qan | + [[ 8 -ge 20 ]] logger.go:42: 16:41:02 | monitoring-pmm3/5-check-qan | ++ curl -s -k -H 'Authorization: Bearer glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9' 'https://34.134.80.59/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7d%20or%20patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7D%29&start=1758818162&end=1758818462&step=60' logger.go:42: 16:41:02 | monitoring-pmm3/5-check-qan | ++ jq '.data.result[0].values[][1]' logger.go:42: 16:41:02 | monitoring-pmm3/5-check-qan | ++ grep '^"[0-9]' logger.go:42: 16:41:02 | monitoring-pmm3/5-check-qan | jq: error (at :0): Cannot iterate over null (null) logger.go:42: 16:41:02 | monitoring-pmm3/5-check-qan | + [[ -n '' ]] logger.go:42: 16:41:02 | monitoring-pmm3/5-check-qan | + sleep 2 logger.go:42: 16:41:04 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s -d '-5 minute' logger.go:42: 16:41:04 | monitoring-pmm3/5-check-qan | + local start=1758818164 logger.go:42: 16:41:04 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s logger.go:42: 16:41:04 | monitoring-pmm3/5-check-qan | + local end=1758818464 logger.go:42: 16:41:04 | monitoring-pmm3/5-check-qan | + let retry+=1 logger.go:42: 16:41:04 | monitoring-pmm3/5-check-qan | + [[ 9 -ge 20 ]] logger.go:42: 16:41:04 | monitoring-pmm3/5-check-qan | ++ curl -s -k -H 'Authorization: Bearer glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9' 'https://34.134.80.59/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7d%20or%20patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7D%29&start=1758818164&end=1758818464&step=60' logger.go:42: 16:41:04 | monitoring-pmm3/5-check-qan | ++ jq '.data.result[0].values[][1]' logger.go:42: 16:41:04 | monitoring-pmm3/5-check-qan | ++ grep '^"[0-9]' logger.go:42: 16:41:04 | monitoring-pmm3/5-check-qan | jq: error (at :0): Cannot iterate over null (null) logger.go:42: 16:41:04 | monitoring-pmm3/5-check-qan | + [[ -n '' ]] logger.go:42: 16:41:04 | monitoring-pmm3/5-check-qan | + sleep 2 logger.go:42: 16:41:06 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s -d '-5 minute' logger.go:42: 16:41:06 | monitoring-pmm3/5-check-qan | + local start=1758818166 logger.go:42: 16:41:06 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s logger.go:42: 16:41:06 | monitoring-pmm3/5-check-qan | + local end=1758818466 logger.go:42: 16:41:06 | monitoring-pmm3/5-check-qan | + let retry+=1 logger.go:42: 16:41:06 | monitoring-pmm3/5-check-qan | + [[ 10 -ge 20 ]] logger.go:42: 16:41:06 | monitoring-pmm3/5-check-qan | ++ curl -s -k -H 'Authorization: Bearer glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9' 'https://34.134.80.59/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7d%20or%20patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7D%29&start=1758818166&end=1758818466&step=60' logger.go:42: 16:41:06 | monitoring-pmm3/5-check-qan | ++ jq '.data.result[0].values[][1]' logger.go:42: 16:41:06 | monitoring-pmm3/5-check-qan | ++ grep '^"[0-9]' logger.go:42: 16:41:07 | monitoring-pmm3/5-check-qan | jq: error (at :0): Cannot iterate over null (null) logger.go:42: 16:41:07 | monitoring-pmm3/5-check-qan | + [[ -n '' ]] logger.go:42: 16:41:07 | monitoring-pmm3/5-check-qan | + sleep 2 logger.go:42: 16:41:09 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s -d '-5 minute' logger.go:42: 16:41:09 | monitoring-pmm3/5-check-qan | + local start=1758818169 logger.go:42: 16:41:09 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s logger.go:42: 16:41:09 | monitoring-pmm3/5-check-qan | + local end=1758818469 logger.go:42: 16:41:09 | monitoring-pmm3/5-check-qan | + let retry+=1 logger.go:42: 16:41:09 | monitoring-pmm3/5-check-qan | + [[ 11 -ge 20 ]] logger.go:42: 16:41:09 | monitoring-pmm3/5-check-qan | ++ jq '.data.result[0].values[][1]' logger.go:42: 16:41:09 | monitoring-pmm3/5-check-qan | ++ grep '^"[0-9]' logger.go:42: 16:41:09 | monitoring-pmm3/5-check-qan | ++ curl -s -k -H 'Authorization: Bearer glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9' 'https://34.134.80.59/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7d%20or%20patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7D%29&start=1758818169&end=1758818469&step=60' logger.go:42: 16:41:09 | monitoring-pmm3/5-check-qan | jq: error (at :0): Cannot iterate over null (null) logger.go:42: 16:41:09 | monitoring-pmm3/5-check-qan | + [[ -n '' ]] logger.go:42: 16:41:09 | monitoring-pmm3/5-check-qan | + sleep 2 logger.go:42: 16:41:11 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s -d '-5 minute' logger.go:42: 16:41:11 | monitoring-pmm3/5-check-qan | + local start=1758818171 logger.go:42: 16:41:11 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s logger.go:42: 16:41:11 | monitoring-pmm3/5-check-qan | + local end=1758818471 logger.go:42: 16:41:11 | monitoring-pmm3/5-check-qan | + let retry+=1 logger.go:42: 16:41:11 | monitoring-pmm3/5-check-qan | + [[ 12 -ge 20 ]] logger.go:42: 16:41:11 | monitoring-pmm3/5-check-qan | ++ curl -s -k -H 'Authorization: Bearer glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9' 'https://34.134.80.59/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7d%20or%20patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7D%29&start=1758818171&end=1758818471&step=60' logger.go:42: 16:41:11 | monitoring-pmm3/5-check-qan | ++ jq '.data.result[0].values[][1]' logger.go:42: 16:41:11 | monitoring-pmm3/5-check-qan | ++ grep '^"[0-9]' logger.go:42: 16:41:12 | monitoring-pmm3/5-check-qan | jq: error (at :0): Cannot iterate over null (null) logger.go:42: 16:41:12 | monitoring-pmm3/5-check-qan | + [[ -n '' ]] logger.go:42: 16:41:12 | monitoring-pmm3/5-check-qan | + sleep 2 logger.go:42: 16:41:14 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s -d '-5 minute' logger.go:42: 16:41:14 | monitoring-pmm3/5-check-qan | + local start=1758818174 logger.go:42: 16:41:14 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s logger.go:42: 16:41:14 | monitoring-pmm3/5-check-qan | + local end=1758818474 logger.go:42: 16:41:14 | monitoring-pmm3/5-check-qan | + let retry+=1 logger.go:42: 16:41:14 | monitoring-pmm3/5-check-qan | + [[ 13 -ge 20 ]] logger.go:42: 16:41:14 | monitoring-pmm3/5-check-qan | ++ curl -s -k -H 'Authorization: Bearer glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9' 'https://34.134.80.59/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7d%20or%20patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7D%29&start=1758818174&end=1758818474&step=60' logger.go:42: 16:41:14 | monitoring-pmm3/5-check-qan | ++ jq '.data.result[0].values[][1]' logger.go:42: 16:41:14 | monitoring-pmm3/5-check-qan | ++ grep '^"[0-9]' logger.go:42: 16:41:14 | monitoring-pmm3/5-check-qan | jq: error (at :0): Cannot iterate over null (null) logger.go:42: 16:41:14 | monitoring-pmm3/5-check-qan | + [[ -n '' ]] logger.go:42: 16:41:14 | monitoring-pmm3/5-check-qan | + sleep 2 logger.go:42: 16:41:16 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s -d '-5 minute' logger.go:42: 16:41:16 | monitoring-pmm3/5-check-qan | + local start=1758818176 logger.go:42: 16:41:16 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s logger.go:42: 16:41:16 | monitoring-pmm3/5-check-qan | + local end=1758818476 logger.go:42: 16:41:16 | monitoring-pmm3/5-check-qan | + let retry+=1 logger.go:42: 16:41:16 | monitoring-pmm3/5-check-qan | + [[ 14 -ge 20 ]] logger.go:42: 16:41:16 | monitoring-pmm3/5-check-qan | ++ curl -s -k -H 'Authorization: Bearer glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9' 'https://34.134.80.59/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7d%20or%20patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7D%29&start=1758818176&end=1758818476&step=60' logger.go:42: 16:41:16 | monitoring-pmm3/5-check-qan | ++ jq '.data.result[0].values[][1]' logger.go:42: 16:41:16 | monitoring-pmm3/5-check-qan | ++ grep '^"[0-9]' logger.go:42: 16:41:16 | monitoring-pmm3/5-check-qan | jq: error (at :0): Cannot iterate over null (null) logger.go:42: 16:41:16 | monitoring-pmm3/5-check-qan | + [[ -n '' ]] logger.go:42: 16:41:16 | monitoring-pmm3/5-check-qan | + sleep 2 logger.go:42: 16:41:18 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s -d '-5 minute' logger.go:42: 16:41:18 | monitoring-pmm3/5-check-qan | + local start=1758818178 logger.go:42: 16:41:18 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s logger.go:42: 16:41:18 | monitoring-pmm3/5-check-qan | + local end=1758818478 logger.go:42: 16:41:18 | monitoring-pmm3/5-check-qan | + let retry+=1 logger.go:42: 16:41:18 | monitoring-pmm3/5-check-qan | + [[ 15 -ge 20 ]] logger.go:42: 16:41:18 | monitoring-pmm3/5-check-qan | ++ curl -s -k -H 'Authorization: Bearer glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9' 'https://34.134.80.59/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7d%20or%20patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7D%29&start=1758818178&end=1758818478&step=60' logger.go:42: 16:41:18 | monitoring-pmm3/5-check-qan | ++ grep '^"[0-9]' logger.go:42: 16:41:18 | monitoring-pmm3/5-check-qan | ++ jq '.data.result[0].values[][1]' logger.go:42: 16:41:19 | monitoring-pmm3/5-check-qan | jq: error (at :0): Cannot iterate over null (null) logger.go:42: 16:41:19 | monitoring-pmm3/5-check-qan | + [[ -n '' ]] logger.go:42: 16:41:19 | monitoring-pmm3/5-check-qan | + sleep 2 logger.go:42: 16:41:21 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s -d '-5 minute' logger.go:42: 16:41:21 | monitoring-pmm3/5-check-qan | + local start=1758818181 logger.go:42: 16:41:21 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s logger.go:42: 16:41:21 | monitoring-pmm3/5-check-qan | + local end=1758818481 logger.go:42: 16:41:21 | monitoring-pmm3/5-check-qan | + let retry+=1 logger.go:42: 16:41:21 | monitoring-pmm3/5-check-qan | + [[ 16 -ge 20 ]] logger.go:42: 16:41:21 | monitoring-pmm3/5-check-qan | ++ curl -s -k -H 'Authorization: Bearer glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9' 'https://34.134.80.59/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7d%20or%20patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7D%29&start=1758818181&end=1758818481&step=60' logger.go:42: 16:41:21 | monitoring-pmm3/5-check-qan | ++ jq '.data.result[0].values[][1]' logger.go:42: 16:41:21 | monitoring-pmm3/5-check-qan | ++ grep '^"[0-9]' logger.go:42: 16:41:21 | monitoring-pmm3/5-check-qan | jq: error (at :0): Cannot iterate over null (null) logger.go:42: 16:41:21 | monitoring-pmm3/5-check-qan | + [[ -n '' ]] logger.go:42: 16:41:21 | monitoring-pmm3/5-check-qan | + sleep 2 logger.go:42: 16:41:23 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s -d '-5 minute' logger.go:42: 16:41:23 | monitoring-pmm3/5-check-qan | + local start=1758818183 logger.go:42: 16:41:23 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s logger.go:42: 16:41:23 | monitoring-pmm3/5-check-qan | + local end=1758818483 logger.go:42: 16:41:23 | monitoring-pmm3/5-check-qan | + let retry+=1 logger.go:42: 16:41:23 | monitoring-pmm3/5-check-qan | + [[ 17 -ge 20 ]] logger.go:42: 16:41:23 | monitoring-pmm3/5-check-qan | ++ jq '.data.result[0].values[][1]' logger.go:42: 16:41:23 | monitoring-pmm3/5-check-qan | ++ curl -s -k -H 'Authorization: Bearer glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9' 'https://34.134.80.59/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7d%20or%20patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7D%29&start=1758818183&end=1758818483&step=60' logger.go:42: 16:41:23 | monitoring-pmm3/5-check-qan | ++ grep '^"[0-9]' logger.go:42: 16:41:24 | monitoring-pmm3/5-check-qan | jq: error (at :0): Cannot iterate over null (null) logger.go:42: 16:41:24 | monitoring-pmm3/5-check-qan | + [[ -n '' ]] logger.go:42: 16:41:24 | monitoring-pmm3/5-check-qan | + sleep 2 logger.go:42: 16:41:26 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s -d '-5 minute' logger.go:42: 16:41:26 | monitoring-pmm3/5-check-qan | + local start=1758818186 logger.go:42: 16:41:26 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s logger.go:42: 16:41:26 | monitoring-pmm3/5-check-qan | + local end=1758818486 logger.go:42: 16:41:26 | monitoring-pmm3/5-check-qan | + let retry+=1 logger.go:42: 16:41:26 | monitoring-pmm3/5-check-qan | + [[ 18 -ge 20 ]] logger.go:42: 16:41:26 | monitoring-pmm3/5-check-qan | ++ curl -s -k -H 'Authorization: Bearer glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9' 'https://34.134.80.59/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7d%20or%20patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7D%29&start=1758818186&end=1758818486&step=60' logger.go:42: 16:41:26 | monitoring-pmm3/5-check-qan | ++ jq '.data.result[0].values[][1]' logger.go:42: 16:41:26 | monitoring-pmm3/5-check-qan | ++ grep '^"[0-9]' logger.go:42: 16:41:26 | monitoring-pmm3/5-check-qan | jq: error (at :0): Cannot iterate over null (null) logger.go:42: 16:41:26 | monitoring-pmm3/5-check-qan | + [[ -n '' ]] logger.go:42: 16:41:26 | monitoring-pmm3/5-check-qan | + sleep 2 logger.go:42: 16:41:28 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s -d '-5 minute' logger.go:42: 16:41:28 | monitoring-pmm3/5-check-qan | + local start=1758818188 logger.go:42: 16:41:28 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s logger.go:42: 16:41:28 | monitoring-pmm3/5-check-qan | + local end=1758818488 logger.go:42: 16:41:28 | monitoring-pmm3/5-check-qan | + let retry+=1 logger.go:42: 16:41:28 | monitoring-pmm3/5-check-qan | + [[ 19 -ge 20 ]] logger.go:42: 16:41:28 | monitoring-pmm3/5-check-qan | ++ jq '.data.result[0].values[][1]' logger.go:42: 16:41:28 | monitoring-pmm3/5-check-qan | ++ curl -s -k -H 'Authorization: Bearer glsa_jLC2unqXVcAQH4GRDDh6JmDt1iIkTi1i_32dcd5d9' 'https://34.134.80.59/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7d%20or%20patroni_postgres_running%7Bnode_name%3D%7E%22kuttl-test-worthy-eagle-monitoring-pmm3-instance1-g7tq-0%22%7D%29&start=1758818188&end=1758818488&step=60' logger.go:42: 16:41:28 | monitoring-pmm3/5-check-qan | ++ grep '^"[0-9]' logger.go:42: 16:41:28 | monitoring-pmm3/5-check-qan | jq: error (at :0): Cannot iterate over null (null) logger.go:42: 16:41:28 | monitoring-pmm3/5-check-qan | + [[ -n '' ]] logger.go:42: 16:41:28 | monitoring-pmm3/5-check-qan | + sleep 2 logger.go:42: 16:41:30 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s -d '-5 minute' logger.go:42: 16:41:30 | monitoring-pmm3/5-check-qan | + local start=1758818190 logger.go:42: 16:41:30 | monitoring-pmm3/5-check-qan | ++ /usr/sbin/date -u +%s logger.go:42: 16:41:30 | monitoring-pmm3/5-check-qan | + local end=1758818490 logger.go:42: 16:41:30 | monitoring-pmm3/5-check-qan | + let retry+=1 logger.go:42: 16:41:30 | monitoring-pmm3/5-check-qan | + [[ 20 -ge 20 ]] logger.go:42: 16:41:30 | monitoring-pmm3/5-check-qan | + exit 1 case.go:396: failed in step 5-check-qan case.go:398: command "set -o xtrace\\n source ../../functions\\n token=$(kubectl get -n \"${..." failed, exit status 1 logger.go:42: 16:41:31 | monitoring-pmm3 | monitoring-pmm3 events from ns kuttl-test-worthy-eagle: logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:36:17 +0000 UTC Normal Pod pg-client-84d6c45668-982v9 Binding Scheduled Successfully assigned kuttl-test-worthy-eagle/pg-client-84d6c45668-982v9 to gke-jen-pg-1273-332364a9-default-pool-dd2964b0-pg8s default-scheduler logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:36:17 +0000 UTC Normal ReplicaSet.apps pg-client-84d6c45668 SuccessfulCreate Created pod: pg-client-84d6c45668-982v9 replicaset-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:36:17 +0000 UTC Normal Deployment.apps pg-client ScalingReplicaSet Scaled up replica set pg-client-84d6c45668 to 1 deployment-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:36:18 +0000 UTC Normal Pod pg-client-84d6c45668-982v9.spec.containers{pg-client} Pulling Pulling image "perconalab/percona-distribution-postgresql:16" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:36:24 +0000 UTC Normal Service monitoring-service EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:36:24 +0000 UTC Normal StatefulSet.apps monitoring SuccessfulCreate create Claim pmm-storage-monitoring-0 Pod monitoring-0 in StatefulSet monitoring success statefulset-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:36:24 +0000 UTC Normal StatefulSet.apps monitoring SuccessfulCreate create Pod monitoring-0 in StatefulSet monitoring successful statefulset-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:36:24 +0000 UTC Normal PersistentVolumeClaim pmm-storage-monitoring-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:36:25 +0000 UTC Normal PersistentVolumeClaim pmm-storage-monitoring-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:36:25 +0000 UTC Normal PersistentVolumeClaim pmm-storage-monitoring-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-worthy-eagle/pmm-storage-monitoring-0" pd.csi.storage.gke.io_gke-ff95a4c0eb1f4faabc77-216b-b758-vm_eb6d55da-68a3-4b97-b715-87bc5a64f939 logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:36:28 +0000 UTC Normal PersistentVolumeClaim pmm-storage-monitoring-0 ProvisioningSucceeded Successfully provisioned volume pvc-7dadc4f5-557f-45dd-81c4-4eef95b6b002 pd.csi.storage.gke.io_gke-ff95a4c0eb1f4faabc77-216b-b758-vm_eb6d55da-68a3-4b97-b715-87bc5a64f939 logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:36:29 +0000 UTC Normal Pod monitoring-0 Binding Scheduled Successfully assigned kuttl-test-worthy-eagle/monitoring-0 to gke-jen-pg-1273-332364a9-default-pool-dd2964b0-pg8s default-scheduler logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:36:36 +0000 UTC Normal Pod monitoring-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-7dadc4f5-557f-45dd-81c4-4eef95b6b002" attachdetach-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:36:37 +0000 UTC Normal Pod monitoring-0.spec.containers{pmm} Pulled Container image "docker.io/perconalab/pmm-server:3-dev-latest" already present on machine kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:36:37 +0000 UTC Normal Pod monitoring-0.spec.containers{pmm} Created Created container: pmm kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:36:37 +0000 UTC Normal Pod pg-client-84d6c45668-982v9.spec.containers{pg-client} Pulled Successfully pulled image "perconalab/percona-distribution-postgresql:16" in 19.566s (19.566s including waiting). Image size: 481593668 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:36:37 +0000 UTC Normal Pod pg-client-84d6c45668-982v9.spec.containers{pg-client} Created Created container: pg-client kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:36:38 +0000 UTC Normal Pod monitoring-0.spec.containers{pmm} Started Started container pmm kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:36:38 +0000 UTC Normal Pod pg-client-84d6c45668-982v9.spec.containers{pg-client} Started Started container pg-client kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:37:01 +0000 UTC Normal Service monitoring-service EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:37:56 +0000 UTC Normal Pod monitoring-pmm3-patroni-version-check Binding Scheduled Successfully assigned kuttl-test-worthy-eagle/monitoring-pmm3-patroni-version-check to gke-jen-pg-1273-332364a9-default-pool-dd2964b0-pg8s default-scheduler logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:37:56 +0000 UTC Normal Pod monitoring-pmm3-patroni-version-check.spec.containers{patroni-version-check} Pulled Container image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" already present on machine kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:37:56 +0000 UTC Normal Pod monitoring-pmm3-patroni-version-check.spec.containers{patroni-version-check} Created Created container: patroni-version-check kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:37:57 +0000 UTC Normal Pod monitoring-pmm3-patroni-version-check.spec.containers{patroni-version-check} Started Started container patroni-version-check kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:02 +0000 UTC Normal Pod monitoring-pmm3-patroni-version-check.spec.containers{patroni-version-check} Killing Stopping container patroni-version-check kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:03 +0000 UTC Normal PersistentVolumeClaim monitoring-pmm3-instance1-g7tq-pgdata WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:03 +0000 UTC Normal PersistentVolumeClaim monitoring-pmm3-instance1-pt7m-pgdata WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:03 +0000 UTC Normal PersistentVolumeClaim monitoring-pmm3-instance1-rrwx-pgdata WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:03 +0000 UTC Normal PersistentVolumeClaim monitoring-pmm3-instance1-rrwx-pgdata ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:03 +0000 UTC Normal PersistentVolumeClaim monitoring-pmm3-instance1-rrwx-pgdata Provisioning External provisioner is provisioning volume for claim "kuttl-test-worthy-eagle/monitoring-pmm3-instance1-rrwx-pgdata" pd.csi.storage.gke.io_gke-ff95a4c0eb1f4faabc77-216b-b758-vm_eb6d55da-68a3-4b97-b715-87bc5a64f939 logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:03 +0000 UTC Normal StatefulSet.apps monitoring-pmm3-instance1-rrwx SuccessfulCreate create Pod monitoring-pmm3-instance1-rrwx-0 in StatefulSet monitoring-pmm3-instance1-rrwx successful statefulset-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:04 +0000 UTC Normal PersistentVolumeClaim monitoring-pmm3-instance1-g7tq-pgdata ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:04 +0000 UTC Normal PersistentVolumeClaim monitoring-pmm3-instance1-g7tq-pgdata Provisioning External provisioner is provisioning volume for claim "kuttl-test-worthy-eagle/monitoring-pmm3-instance1-g7tq-pgdata" pd.csi.storage.gke.io_gke-ff95a4c0eb1f4faabc77-216b-b758-vm_eb6d55da-68a3-4b97-b715-87bc5a64f939 logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:04 +0000 UTC Normal StatefulSet.apps monitoring-pmm3-instance1-g7tq SuccessfulCreate create Pod monitoring-pmm3-instance1-g7tq-0 in StatefulSet monitoring-pmm3-instance1-g7tq successful statefulset-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:04 +0000 UTC Normal PersistentVolumeClaim monitoring-pmm3-instance1-pt7m-pgdata ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:04 +0000 UTC Normal PersistentVolumeClaim monitoring-pmm3-instance1-pt7m-pgdata Provisioning External provisioner is provisioning volume for claim "kuttl-test-worthy-eagle/monitoring-pmm3-instance1-pt7m-pgdata" pd.csi.storage.gke.io_gke-ff95a4c0eb1f4faabc77-216b-b758-vm_eb6d55da-68a3-4b97-b715-87bc5a64f939 logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:04 +0000 UTC Normal StatefulSet.apps monitoring-pmm3-instance1-pt7m SuccessfulCreate create Pod monitoring-pmm3-instance1-pt7m-0 in StatefulSet monitoring-pmm3-instance1-pt7m successful statefulset-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:05 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-78slk Binding Scheduled Successfully assigned kuttl-test-worthy-eagle/monitoring-pmm3-pgbouncer-5595f6c49f-78slk to gke-jen-pg-1273-332364a9-default-pool-dd2964b0-pg8s default-scheduler logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:05 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-fkmb2 Binding Scheduled Successfully assigned kuttl-test-worthy-eagle/monitoring-pmm3-pgbouncer-5595f6c49f-fkmb2 to gke-jen-pg-1273-332364a9-default-pool-dd2964b0-qlgh default-scheduler logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:05 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-lnbtz Binding Scheduled Successfully assigned kuttl-test-worthy-eagle/monitoring-pmm3-pgbouncer-5595f6c49f-lnbtz to gke-jen-pg-1273-332364a9-default-pool-dd2964b0-gr5h default-scheduler logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:05 +0000 UTC Normal ReplicaSet.apps monitoring-pmm3-pgbouncer-5595f6c49f SuccessfulCreate Created pod: monitoring-pmm3-pgbouncer-5595f6c49f-lnbtz replicaset-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:05 +0000 UTC Normal ReplicaSet.apps monitoring-pmm3-pgbouncer-5595f6c49f SuccessfulCreate Created pod: monitoring-pmm3-pgbouncer-5595f6c49f-78slk replicaset-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:05 +0000 UTC Normal ReplicaSet.apps monitoring-pmm3-pgbouncer-5595f6c49f SuccessfulCreate Created pod: monitoring-pmm3-pgbouncer-5595f6c49f-fkmb2 replicaset-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:05 +0000 UTC Normal Deployment.apps monitoring-pmm3-pgbouncer ScalingReplicaSet Scaled up replica set monitoring-pmm3-pgbouncer-5595f6c49f to 3 deployment-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:05 +0000 UTC Normal PodDisruptionBudget.policy monitoring-pmm3-pgbouncer NoPods No matching pods found controllermanager logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:05 +0000 UTC Normal StatefulSet.apps monitoring-pmm3-repo-host SuccessfulCreate create Pod monitoring-pmm3-repo-host-0 in StatefulSet monitoring-pmm3-repo-host successful statefulset-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:05 +0000 UTC Normal PersistentVolumeClaim monitoring-pmm3-repo1 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:05 +0000 UTC Normal PersistentVolumeClaim monitoring-pmm3-repo1 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:05 +0000 UTC Normal PersistentVolumeClaim monitoring-pmm3-repo1 Provisioning External provisioner is provisioning volume for claim "kuttl-test-worthy-eagle/monitoring-pmm3-repo1" pd.csi.storage.gke.io_gke-ff95a4c0eb1f4faabc77-216b-b758-vm_eb6d55da-68a3-4b97-b715-87bc5a64f939 logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:05 +0000 UTC Normal PostgresCluster.postgres-operator.crunchydata.com monitoring-pmm3 RepoHostCreated created pgBackRest repository host StatefulSet/monitoring-pmm3-repo-host postgrescluster-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:06 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-78slk.spec.containers{pgbouncer} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:06 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-78slk.spec.containers{pgbouncer} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17" in 163ms (163ms including waiting). Image size: 82505487 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:06 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-78slk.spec.containers{pgbouncer} Created Created container: pgbouncer kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:06 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-78slk.spec.containers{pgbouncer} Started Started container pgbouncer kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:06 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-78slk.spec.containers{pgbouncer-config} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:06 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-78slk.spec.containers{pgbouncer-config} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17" in 198ms (198ms including waiting). Image size: 82505487 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:06 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-78slk.spec.containers{pgbouncer-config} Created Created container: pgbouncer-config kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:06 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-fkmb2.spec.containers{pgbouncer} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:06 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-fkmb2.spec.containers{pgbouncer} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17" in 153ms (154ms including waiting). Image size: 82505487 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:06 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-fkmb2.spec.containers{pgbouncer} Created Created container: pgbouncer kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:06 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-fkmb2.spec.containers{pgbouncer} Started Started container pgbouncer kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:06 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-fkmb2.spec.containers{pgbouncer-config} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:06 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-fkmb2.spec.containers{pgbouncer-config} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17" in 157ms (157ms including waiting). Image size: 82505487 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:06 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-fkmb2.spec.containers{pgbouncer-config} Created Created container: pgbouncer-config kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:06 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-fkmb2.spec.containers{pgbouncer-config} Started Started container pgbouncer-config kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:06 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-lnbtz.spec.containers{pgbouncer} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:06 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-lnbtz.spec.containers{pgbouncer} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17" in 170ms (170ms including waiting). Image size: 82505487 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:06 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-lnbtz.spec.containers{pgbouncer} Created Created container: pgbouncer kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:06 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-lnbtz.spec.containers{pgbouncer} Started Started container pgbouncer kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:06 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-lnbtz.spec.containers{pgbouncer-config} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:06 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-lnbtz.spec.containers{pgbouncer-config} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer17" in 176ms (176ms including waiting). Image size: 82505487 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:06 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-lnbtz.spec.containers{pgbouncer-config} Created Created container: pgbouncer-config kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:06 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-lnbtz.spec.containers{pgbouncer-config} Started Started container pgbouncer-config kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:07 +0000 UTC Normal PersistentVolumeClaim monitoring-pmm3-instance1-rrwx-pgdata ProvisioningSucceeded Successfully provisioned volume pvc-d656a95c-3bbb-4f1d-b8a2-5f05da279d48 pd.csi.storage.gke.io_gke-ff95a4c0eb1f4faabc77-216b-b758-vm_eb6d55da-68a3-4b97-b715-87bc5a64f939 logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:07 +0000 UTC Normal Pod monitoring-pmm3-pgbouncer-5595f6c49f-78slk.spec.containers{pgbouncer-config} Started Started container pgbouncer-config kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:08 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0 Binding Scheduled Successfully assigned kuttl-test-worthy-eagle/monitoring-pmm3-instance1-g7tq-0 to gke-jen-pg-1273-332364a9-default-pool-dd2964b0-pg8s default-scheduler logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:08 +0000 UTC Normal PersistentVolumeClaim monitoring-pmm3-instance1-g7tq-pgdata ProvisioningSucceeded Successfully provisioned volume pvc-df25229b-6938-4cfb-b678-0a54de700ea3 pd.csi.storage.gke.io_gke-ff95a4c0eb1f4faabc77-216b-b758-vm_eb6d55da-68a3-4b97-b715-87bc5a64f939 logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:08 +0000 UTC Normal PersistentVolumeClaim monitoring-pmm3-instance1-pt7m-pgdata ProvisioningSucceeded Successfully provisioned volume pvc-8d060951-37be-46e5-850e-9c49c3dce2db pd.csi.storage.gke.io_gke-ff95a4c0eb1f4faabc77-216b-b758-vm_eb6d55da-68a3-4b97-b715-87bc5a64f939 logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:08 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0 Binding Scheduled Successfully assigned kuttl-test-worthy-eagle/monitoring-pmm3-instance1-rrwx-0 to gke-jen-pg-1273-332364a9-default-pool-dd2964b0-qlgh default-scheduler logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:09 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0 Binding Scheduled Successfully assigned kuttl-test-worthy-eagle/monitoring-pmm3-instance1-pt7m-0 to gke-jen-pg-1273-332364a9-default-pool-dd2964b0-gr5h default-scheduler logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:09 +0000 UTC Normal PersistentVolumeClaim monitoring-pmm3-repo1 ProvisioningSucceeded Successfully provisioned volume pvc-bcb10e5c-ea5b-489e-a591-3cad217df6b2 pd.csi.storage.gke.io_gke-ff95a4c0eb1f4faabc77-216b-b758-vm_eb6d55da-68a3-4b97-b715-87bc5a64f939 logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:10 +0000 UTC Normal Pod monitoring-pmm3-repo-host-0 Binding Scheduled Successfully assigned kuttl-test-worthy-eagle/monitoring-pmm3-repo-host-0 to gke-jen-pg-1273-332364a9-default-pool-dd2964b0-qlgh default-scheduler logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:13 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0 Binding Scheduled Successfully assigned kuttl-test-worthy-eagle/monitoring-pmm3-instance1-rrwx-0 to gke-jen-pg-1273-332364a9-default-pool-dd2964b0-qlgh default-scheduler logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:14 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0 Binding Scheduled Successfully assigned kuttl-test-worthy-eagle/monitoring-pmm3-instance1-g7tq-0 to gke-jen-pg-1273-332364a9-default-pool-dd2964b0-gr5h default-scheduler logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:14 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0 Binding Scheduled Successfully assigned kuttl-test-worthy-eagle/monitoring-pmm3-instance1-pt7m-0 to gke-jen-pg-1273-332364a9-default-pool-dd2964b0-pg8s default-scheduler logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:15 +0000 UTC Warning Pod monitoring-pmm3-instance1-g7tq-0 FailedMount Unable to attach or mount volumes: unmounted volumes=[cert-volume crunchy-bin database-containerinfo dshm kube-api-access-zjkls patroni-config pgbackrest-config pgbackrest-server postgres-data tmp], unattached volumes=[], failed to process volumes=[postgres-data]: error processing PVC kuttl-test-worthy-eagle/monitoring-pmm3-instance1-g7tq-pgdata: failed to fetch PVC from API server: persistentvolumeclaims "monitoring-pmm3-instance1-g7tq-pgdata" is forbidden: User "system:node:gke-jen-pg-1273-332364a9-default-pool-dd2964b0-gr5h" cannot get resource "persistentvolumeclaims" in API group "" in the namespace "kuttl-test-worthy-eagle": no relationship found between node 'gke-jen-pg-1273-332364a9-default-pool-dd2964b0-gr5h' and this object kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:15 +0000 UTC Warning Pod monitoring-pmm3-instance1-pt7m-0 FailedMount Unable to attach or mount volumes: unmounted volumes=[cert-volume crunchy-bin database-containerinfo dshm kube-api-access-v9wnc patroni-config pgbackrest-config pgbackrest-server postgres-data tmp], unattached volumes=[], failed to process volumes=[postgres-data]: error processing PVC kuttl-test-worthy-eagle/monitoring-pmm3-instance1-pt7m-pgdata: failed to fetch PVC from API server: persistentvolumeclaims "monitoring-pmm3-instance1-pt7m-pgdata" is forbidden: User "system:node:gke-jen-pg-1273-332364a9-default-pool-dd2964b0-pg8s" cannot get resource "persistentvolumeclaims" in API group "" in the namespace "kuttl-test-worthy-eagle": no relationship found between node 'gke-jen-pg-1273-332364a9-default-pool-dd2964b0-pg8s' and this object kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:16 +0000 UTC Warning Pod monitoring-pmm3-instance1-g7tq-0 FailedMount MountVolume.SetUp failed for volume "cert-volume" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:16 +0000 UTC Warning Pod monitoring-pmm3-instance1-g7tq-0 FailedMount MountVolume.SetUp failed for volume "pgbackrest-config" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:16 +0000 UTC Warning Pod monitoring-pmm3-instance1-pt7m-0 FailedMount MountVolume.SetUp failed for volume "cert-volume" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:16 +0000 UTC Warning Pod monitoring-pmm3-instance1-pt7m-0 FailedMount MountVolume.SetUp failed for volume "pgbackrest-server" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:16 +0000 UTC Warning Pod monitoring-pmm3-instance1-pt7m-0 FailedMount MountVolume.SetUp failed for volume "patroni-config" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:16 +0000 UTC Warning Pod monitoring-pmm3-instance1-pt7m-0 FailedMount MountVolume.SetUp failed for volume "pgbackrest-config" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:16 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-8d060951-37be-46e5-850e-9c49c3dce2db" attachdetach-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:16 +0000 UTC Warning Pod monitoring-pmm3-instance1-pt7m-0 FailedAttachVolume Multi-Attach error for volume "pvc-8d060951-37be-46e5-850e-9c49c3dce2db" Volume is already exclusively attached to one node and can't be attached to another attachdetach-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:17 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-df25229b-6938-4cfb-b678-0a54de700ea3" attachdetach-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:17 +0000 UTC Warning Pod monitoring-pmm3-instance1-g7tq-0 FailedAttachVolume Multi-Attach error for volume "pvc-df25229b-6938-4cfb-b678-0a54de700ea3" Volume is already exclusively attached to one node and can't be attached to another attachdetach-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:17 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-d656a95c-3bbb-4f1d-b8a2-5f05da279d48" attachdetach-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:20 +0000 UTC Normal Pod monitoring-pmm3-repo-host-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-bcb10e5c-ea5b-489e-a591-3cad217df6b2" attachdetach-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:21 +0000 UTC Normal Pod monitoring-pmm3-repo-host-0.spec.initContainers{pgbackrest-log-dir} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:22 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.initContainers{postgres-startup} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:22 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.initContainers{postgres-startup} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" in 194ms (194ms including waiting). Image size: 535314085 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:22 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.initContainers{postgres-startup} Created Created container: postgres-startup kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:22 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.initContainers{postgres-startup} Started Started container postgres-startup kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:22 +0000 UTC Normal Pod monitoring-pmm3-repo-host-0.spec.initContainers{pgbackrest-log-dir} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17" in 142ms (142ms including waiting). Image size: 153118795 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:22 +0000 UTC Normal Pod monitoring-pmm3-repo-host-0.spec.initContainers{pgbackrest-log-dir} Created Created container: pgbackrest-log-dir kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:22 +0000 UTC Normal Pod monitoring-pmm3-repo-host-0.spec.initContainers{pgbackrest-log-dir} Started Started container pgbackrest-log-dir kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:22 +0000 UTC Normal Pod monitoring-pmm3-repo-host-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:22 +0000 UTC Normal Pod monitoring-pmm3-repo-host-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17" in 182ms (182ms including waiting). Image size: 153118795 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:22 +0000 UTC Normal Pod monitoring-pmm3-repo-host-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:22 +0000 UTC Normal Pod monitoring-pmm3-repo-host-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:23 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.initContainers{database-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:23 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.initContainers{database-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b" in 231ms (231ms including waiting). Image size: 86722161 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:23 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.initContainers{database-init} Created Created container: database-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:23 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.initContainers{database-init} Started Started container database-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:23 +0000 UTC Normal Pod monitoring-pmm3-repo-host-0.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:23 +0000 UTC Normal Pod monitoring-pmm3-repo-host-0.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17" in 206ms (206ms including waiting). Image size: 153118795 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:23 +0000 UTC Normal Pod monitoring-pmm3-repo-host-0.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:23 +0000 UTC Normal Pod monitoring-pmm3-repo-host-0.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:23 +0000 UTC Normal Pod monitoring-pmm3-repo-host-0.spec.containers{pgbackrest-config} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:23 +0000 UTC Normal Pod monitoring-pmm3-repo-host-0.spec.containers{pgbackrest-config} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17" in 295ms (295ms including waiting). Image size: 153118795 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:24 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:24 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" in 177ms (177ms including waiting). Image size: 535314085 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:24 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:24 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:24 +0000 UTC Normal Pod monitoring-pmm3-repo-host-0.spec.containers{pgbackrest-config} Created Created container: pgbackrest-config kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:24 +0000 UTC Normal Pod monitoring-pmm3-repo-host-0.spec.containers{pgbackrest-config} Started Started container pgbackrest-config kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:25 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{database} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:25 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{database} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" in 165ms (165ms including waiting). Image size: 535314085 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:25 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{database} Created Created container: database kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:25 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{database} Started Started container database kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:25 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{replication-cert-copy} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:25 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{replication-cert-copy} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" in 172ms (172ms including waiting). Image size: 535314085 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:25 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{replication-cert-copy} Created Created container: replication-cert-copy kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:25 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{replication-cert-copy} Started Started container replication-cert-copy kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:25 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:26 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17" in 147ms (147ms including waiting). Image size: 153118795 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:26 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:26 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:26 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{pgbackrest-config} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:32 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-df25229b-6938-4cfb-b678-0a54de700ea3" attachdetach-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:34 +0000 UTC Normal Pod monitoring-pmm3-backup-ng8b-9qqmc Binding Scheduled Successfully assigned kuttl-test-worthy-eagle/monitoring-pmm3-backup-ng8b-9qqmc to gke-jen-pg-1273-332364a9-default-pool-dd2964b0-pg8s default-scheduler logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:34 +0000 UTC Normal Job.batch monitoring-pmm3-backup-ng8b SuccessfulCreate Created pod: monitoring-pmm3-backup-ng8b-9qqmc job-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:34 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.initContainers{postgres-startup} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:34 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.initContainers{postgres-startup} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" in 185ms (185ms including waiting). Image size: 535314085 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:34 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.initContainers{postgres-startup} Created Created container: postgres-startup kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:34 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.initContainers{postgres-startup} Started Started container postgres-startup kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:34 +0000 UTC Normal PostgresCluster.postgres-operator.crunchydata.com monitoring-pmm3 StanzasCreated pgBackRest stanza creation completed successfully postgrescluster-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:35 +0000 UTC Normal Pod monitoring-pmm3-backup-ng8b-9qqmc.spec.initContainers{pgbackrest-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:35 +0000 UTC Normal Pod monitoring-pmm3-backup-ng8b-9qqmc.spec.initContainers{pgbackrest-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b" in 165ms (165ms including waiting). Image size: 86722161 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:35 +0000 UTC Normal Pod monitoring-pmm3-backup-ng8b-9qqmc.spec.initContainers{pgbackrest-init} Created Created container: pgbackrest-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:35 +0000 UTC Normal Pod monitoring-pmm3-backup-ng8b-9qqmc.spec.initContainers{pgbackrest-init} Started Started container pgbackrest-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:35 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.initContainers{database-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:35 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.initContainers{database-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b" in 181ms (181ms including waiting). Image size: 86722161 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:35 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.initContainers{database-init} Created Created container: database-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:35 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.initContainers{database-init} Started Started container database-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:36 +0000 UTC Normal Pod monitoring-pmm3-backup-ng8b-9qqmc.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:36 +0000 UTC Normal Pod monitoring-pmm3-backup-ng8b-9qqmc.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17" in 199ms (199ms including waiting). Image size: 153118795 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:36 +0000 UTC Normal Pod monitoring-pmm3-backup-ng8b-9qqmc.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:36 +0000 UTC Normal Pod monitoring-pmm3-backup-ng8b-9qqmc.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:36 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:36 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" in 226ms (226ms including waiting). Image size: 535314085 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:36 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:37 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:37 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.containers{database} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:37 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.containers{database} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" in 199ms (199ms including waiting). Image size: 535314085 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:37 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.containers{database} Created Created container: database kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:37 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.containers{database} Started Started container database kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:37 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.containers{replication-cert-copy} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:38 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.containers{replication-cert-copy} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" in 189ms (189ms including waiting). Image size: 535314085 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:38 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.containers{replication-cert-copy} Created Created container: replication-cert-copy kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:38 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.containers{replication-cert-copy} Started Started container replication-cert-copy kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:38 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:38 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17" in 161ms (161ms including waiting). Image size: 153118795 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:38 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-8d060951-37be-46e5-850e-9c49c3dce2db" attachdetach-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:40 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.initContainers{postgres-startup} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:40 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.initContainers{postgres-startup} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" in 182ms (182ms including waiting). Image size: 535314085 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:40 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.initContainers{postgres-startup} Created Created container: postgres-startup kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:40 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.initContainers{postgres-startup} Started Started container postgres-startup kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:41 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.initContainers{database-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:41 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.initContainers{database-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b" in 164ms (164ms including waiting). Image size: 86722161 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:41 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.initContainers{database-init} Created Created container: database-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:41 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.initContainers{database-init} Started Started container database-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:41 +0000 UTC Warning PostgresCluster.postgres-operator.crunchydata.com monitoring-pmm3 UnableToCreateStanzas command terminated with exit code 50: repo1-path = /pgbackrest/repo1 2025-09-25 16:38:41.207 P00 ERROR: [050]: raised from remote-0 tls protocol on 'monitoring-pmm3-repo-host-0.monitoring-pmm3-pods.kuttl-test-worthy-eagle.svc.cluster.local.': unable to acquire lock on file '/tmp/pgbackrest/db-backup-1.lock': Resource temporarily unavailable HINT: is another pgBackRest process running? 2025-09-25 16:38:41.315 P00 ERROR: [050]: raised from remote-0 tls protocol on 'monitoring-pmm3-repo-host-0.monitoring-pmm3-pods.kuttl-test-worthy-eagle.svc.cluster.local.': unable to acquire lock on file '/tmp/pgbackrest/db-backup-1.lock': Resource temporarily unavailable HINT: is another pgBackRest process running? postgrescluster-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:42 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:42 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" in 227ms (227ms including waiting). Image size: 535314085 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:42 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:42 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:43 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.containers{database} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:43 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.containers{database} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" in 163ms (163ms including waiting). Image size: 535314085 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:43 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.containers{database} Created Created container: database kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:43 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.containers{database} Started Started container database kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:43 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.containers{replication-cert-copy} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:43 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.containers{replication-cert-copy} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" in 178ms (178ms including waiting). Image size: 535314085 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:44 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.containers{replication-cert-copy} Created Created container: replication-cert-copy kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:38:44 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.containers{replication-cert-copy} Started Started container replication-cert-copy kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:11 +0000 UTC Normal Job.batch monitoring-pmm3-backup-ng8b Completed Job completed job-controller logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:36 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0 Binding Scheduled Successfully assigned kuttl-test-worthy-eagle/monitoring-pmm3-instance1-pt7m-0 to gke-jen-pg-1273-332364a9-default-pool-dd2964b0-pg8s default-scheduler logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:47 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.initContainers{postgres-startup} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:47 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.initContainers{postgres-startup} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" in 180ms (180ms including waiting). Image size: 535314085 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:47 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.initContainers{postgres-startup} Created Created container: postgres-startup kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:47 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.initContainers{postgres-startup} Started Started container postgres-startup kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:48 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.initContainers{database-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:48 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.initContainers{database-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b" in 157ms (158ms including waiting). Image size: 86722161 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:48 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.initContainers{database-init} Created Created container: database-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:48 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.initContainers{database-init} Started Started container database-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:49 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:49 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" in 163ms (163ms including waiting). Image size: 535314085 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:49 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:49 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:50 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.containers{database} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:50 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.containers{database} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" in 173ms (174ms including waiting). Image size: 535314085 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:50 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.containers{database} Created Created container: database kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:50 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.containers{database} Started Started container database kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:50 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.containers{replication-cert-copy} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:51 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.containers{replication-cert-copy} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" in 193ms (193ms including waiting). Image size: 535314085 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:51 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.containers{replication-cert-copy} Created Created container: replication-cert-copy kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:51 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.containers{replication-cert-copy} Started Started container replication-cert-copy kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:51 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.containers{pmm-client} Pulled Container image "docker.io/perconalab/pmm-client:3-dev-latest" already present on machine kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:51 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.containers{pmm-client} Created Created container: pmm-client kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:51 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.containers{pmm-client} Started Started container pmm-client kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:51 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:39:51 +0000 UTC Normal Pod monitoring-pmm3-instance1-pt7m-0.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17" in 183ms (183ms including waiting). Image size: 153118795 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:00 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0 Binding Scheduled Successfully assigned kuttl-test-worthy-eagle/monitoring-pmm3-instance1-g7tq-0 to gke-jen-pg-1273-332364a9-default-pool-dd2964b0-gr5h default-scheduler logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:04 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.initContainers{postgres-startup} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:05 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.initContainers{postgres-startup} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" in 205ms (205ms including waiting). Image size: 535314085 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:05 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.initContainers{postgres-startup} Created Created container: postgres-startup kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:05 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.initContainers{postgres-startup} Started Started container postgres-startup kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:05 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.initContainers{database-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:06 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.initContainers{database-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b" in 211ms (211ms including waiting). Image size: 86722161 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:06 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.initContainers{database-init} Created Created container: database-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:06 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.initContainers{database-init} Started Started container database-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:06 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:07 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" in 162ms (162ms including waiting). Image size: 535314085 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:07 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:07 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:07 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.containers{database} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:08 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.containers{database} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" in 178ms (178ms including waiting). Image size: 535314085 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:08 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.containers{database} Created Created container: database kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:08 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.containers{database} Started Started container database kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:08 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.containers{replication-cert-copy} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:08 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.containers{replication-cert-copy} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" in 201ms (201ms including waiting). Image size: 535314085 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:08 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.containers{replication-cert-copy} Created Created container: replication-cert-copy kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:08 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.containers{replication-cert-copy} Started Started container replication-cert-copy kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:08 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.containers{pmm-client} Pulled Container image "docker.io/perconalab/pmm-client:3-dev-latest" already present on machine kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:08 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.containers{pmm-client} Created Created container: pmm-client kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:08 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.containers{pmm-client} Started Started container pmm-client kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:08 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:08 +0000 UTC Normal Pod monitoring-pmm3-instance1-g7tq-0.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17" in 159ms (159ms including waiting). Image size: 153118795 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:22 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0 Binding Scheduled Successfully assigned kuttl-test-worthy-eagle/monitoring-pmm3-instance1-rrwx-0 to gke-jen-pg-1273-332364a9-default-pool-dd2964b0-qlgh default-scheduler logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:25 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.initContainers{postgres-startup} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:25 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.initContainers{postgres-startup} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" in 236ms (236ms including waiting). Image size: 535314085 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:25 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.initContainers{postgres-startup} Created Created container: postgres-startup kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:25 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.initContainers{postgres-startup} Started Started container postgres-startup kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:25 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.initContainers{database-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:25 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.initContainers{database-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1273-332364a9b" in 183ms (183ms including waiting). Image size: 86722161 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:25 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.initContainers{database-init} Created Created container: database-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:26 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.initContainers{database-init} Started Started container database-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:26 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:26 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" in 174ms (174ms including waiting). Image size: 535314085 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:26 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:27 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:27 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{database} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:27 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{database} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" in 173ms (173ms including waiting). Image size: 535314085 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:27 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{database} Created Created container: database kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:28 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{database} Started Started container database kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:28 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{replication-cert-copy} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:28 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{replication-cert-copy} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg17-postgres" in 201ms (201ms including waiting). Image size: 535314085 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:28 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{replication-cert-copy} Created Created container: replication-cert-copy kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:28 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{replication-cert-copy} Started Started container replication-cert-copy kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:28 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{pmm-client} Pulled Container image "docker.io/perconalab/pmm-client:3-dev-latest" already present on machine kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:28 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{pmm-client} Created Created container: pmm-client kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:28 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{pmm-client} Started Started container pmm-client kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:28 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | 2025-09-25 16:40:28 +0000 UTC Normal Pod monitoring-pmm3-instance1-rrwx-0.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest17" in 165ms (165ms including waiting). Image size: 153118795 bytes. kubelet logger.go:42: 16:41:31 | monitoring-pmm3 | Deleting namespace: kuttl-test-worthy-eagle === NAME kuttl harness.go:403: run tests finished harness.go:510: cleaning up harness.go:567: removing temp folder: "" --- FAIL: kuttl (389.80s) --- FAIL: kuttl/harness (0.00s) --- FAIL: kuttl/harness/monitoring-pmm3 (389.06s) FAIL