=== RUN kuttl harness.go:460: starting setup harness.go:258: running tests using configured kubeconfig. harness.go:281: Successful connection to cluster at: https://34.171.200.174 harness.go:366: running tests harness.go:77: going to run test suite with timeout of 600 seconds for each step harness.go:378: testsuite: e2e-tests/tests has 36 tests === RUN kuttl/harness === RUN kuttl/harness/demand-backup === PAUSE kuttl/harness/demand-backup === CONT kuttl/harness/demand-backup logger.go:42: 03:26:21 | demand-backup | Ignoring "conf": does not begin with a number followed by a dash. logger.go:42: 03:26:21 | demand-backup | Creating namespace "kuttl-test-growing-killdeer" logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | starting test step 0-deploy-operator logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | running command: [sh -c set -o errexit set -o xtrace source ../../functions init_temp_dir # do this only in the first TestStep deploy_operator deploy_client deploy_s3_secrets] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | + source ../../functions logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ realpath ../../.. logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | ++ CERT_MANAGER_VER=1.19.1 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | ++++ pwd logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | ++ test_name=demand-backup logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/vars.sh logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/conf logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/conf logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export TEST_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ TEST_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export TEMP_DIR=/tmp/kuttl/pg/demand-backup logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ TEMP_DIR=/tmp/kuttl/pg/demand-backup logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export GIT_BRANCH=PR-1506 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ GIT_BRANCH=PR-1506 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export VERSION=PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ VERSION=PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export SKIP_TEST_WARNINGS=true logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ SKIP_TEST_WARNINGS=true logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ command -v oc logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ ! -n '' ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ -n '' ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export PG_VER=18 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ PG_VER=18 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export BUCKET=pg-operator-testing logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ BUCKET=pg-operator-testing logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export IMAGE_PMM3_CLIENT=perconalab/pmm-client:3.4 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ IMAGE_PMM3_CLIENT=perconalab/pmm-client:3.4 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export IMAGE_PMM3_SERVER=perconalab/pmm-server:3.4 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ IMAGE_PMM3_SERVER=perconalab/pmm-server:3.4 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export PGOV1_TAG=1.4.0 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ PGOV1_TAG=1.4.0 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export PGOV1_VER=14 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ PGOV1_VER=14 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export MINIO_VER=5.4.0 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ MINIO_VER=5.4.0 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export VAULT_VER=0.32.0 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ VAULT_VER=0.32.0 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export REGISTRY_NAME=docker.io logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ REGISTRY_NAME=docker.io logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export REGISTRY_NAME_FULL=docker.io/ logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ REGISTRY_NAME_FULL=docker.io/ logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | ++++ printenv logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | ++++ grep -E '^IMAGE' logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | ++++ awk -F= '{print $1}' logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | ++++ eval 'echo $IMAGE_BACKREST' logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++++ echo perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ var_value=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == docker.io/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == percona/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == perconalab/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ echo IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | ++++ eval 'echo $IMAGE_PMM3_CLIENT' logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++++ echo perconalab/pmm-client:3.4 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ var_value=perconalab/pmm-client:3.4 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/pmm-client:3.4 == docker.io/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/pmm-client:3.4 == percona/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/pmm-client:3.4 == perconalab/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ new_value=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ echo IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | ++++ eval 'echo $IMAGE' logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ var_value=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == docker.io/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == percona/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == perconalab/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ new_value=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ echo IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | ++++ eval 'echo $IMAGE_UPGRADE' logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++++ echo perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ var_value=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == docker.io/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == percona/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == perconalab/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ echo IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | ++++ eval 'echo $IMAGE_BASE' logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++++ echo perconalab/percona-postgresql-operator logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ var_value=perconalab/percona-postgresql-operator logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator == docker.io/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator == percona/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator == perconalab/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ new_value=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ echo IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | ++++ eval 'echo $IMAGE_PMM3_SERVER' logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++++ echo perconalab/pmm-server:3.4 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ var_value=perconalab/pmm-server:3.4 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/pmm-server:3.4 == docker.io/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/pmm-server:3.4 == percona/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/pmm-server:3.4 == perconalab/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ new_value=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ echo IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | ++++ eval 'echo $IMAGE_PGBOUNCER' logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++++ echo perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ var_value=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == docker.io/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == percona/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == perconalab/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ echo IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | ++++ eval 'echo $IMAGE_POSTGRESQL' logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++++ echo perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ var_value=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == docker.io/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == percona/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == perconalab/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ echo IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | ++++ eval 'echo $IMAGE_PMM_SERVER' logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++++ echo perconalab/pmm-server:dev-latest logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ var_value=perconalab/pmm-server:dev-latest logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/pmm-server:dev-latest == docker.io/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/pmm-server:dev-latest == percona/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/pmm-server:dev-latest == perconalab/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ new_value=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ echo IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | ++++ eval 'echo $IMAGE_PMM_CLIENT' logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++++ echo perconalab/pmm-client:dev-latest logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ var_value=perconalab/pmm-client:dev-latest logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/pmm-client:dev-latest == docker.io/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/pmm-client:dev-latest == percona/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ [[ perconalab/pmm-client:dev-latest == perconalab/* ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ new_value=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ export IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ echo IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | ++++ which gdate logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | ++++ which date logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ date=/usr/sbin/date logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | ++++ which gsed logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | ++++ which sed logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | +++ sed=/usr/sbin/sed logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | ++ oc get projects logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | + init_temp_dir logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | + rm -rf /tmp/kuttl/pg/demand-backup logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | + mkdir -p /tmp/kuttl/pg/demand-backup logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | + deploy_operator logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | + local cw_prefix= logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | + destroy_operator logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | + kubectl -n pg-operator delete deployment percona-postgresql-operator --force --grace-period=0 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | Error from server (NotFound): deployments.apps "percona-postgresql-operator" not found logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | + true logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | + [[ -n pg-operator ]] logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | + kubectl delete namespace pg-operator --force --grace-period=0 logger.go:42: 03:26:22 | demand-backup/0-deploy-operator | Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. logger.go:42: 03:26:23 | demand-backup/0-deploy-operator | Error from server (NotFound): namespaces "pg-operator" not found logger.go:42: 03:26:23 | demand-backup/0-deploy-operator | + true logger.go:42: 03:26:23 | demand-backup/0-deploy-operator | + [[ -n pg-operator ]] logger.go:42: 03:26:23 | demand-backup/0-deploy-operator | + create_namespace pg-operator logger.go:42: 03:26:23 | demand-backup/0-deploy-operator | + local namespace=pg-operator logger.go:42: 03:26:23 | demand-backup/0-deploy-operator | + [[ -n '' ]] logger.go:42: 03:26:23 | demand-backup/0-deploy-operator | + kubectl delete namespace pg-operator --ignore-not-found logger.go:42: 03:26:23 | demand-backup/0-deploy-operator | + kubectl wait --for=delete namespace pg-operator logger.go:42: 03:26:24 | demand-backup/0-deploy-operator | + kubectl create namespace pg-operator logger.go:42: 03:26:24 | demand-backup/0-deploy-operator | namespace/pg-operator created logger.go:42: 03:26:24 | demand-backup/0-deploy-operator | + cw_prefix=cw- logger.go:42: 03:26:24 | demand-backup/0-deploy-operator | + kubectl -n pg-operator apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy/crd.yaml logger.go:42: 03:26:26 | demand-backup/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/crunchybridgeclusters.postgres-operator.crunchydata.com serverside-applied logger.go:42: 03:26:26 | demand-backup/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconapgbackups.pgv2.percona.com serverside-applied logger.go:42: 03:26:29 | demand-backup/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconapgclusters.pgv2.percona.com serverside-applied logger.go:42: 03:26:29 | demand-backup/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconapgrestores.pgv2.percona.com serverside-applied logger.go:42: 03:26:30 | demand-backup/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconapgupgrades.pgv2.percona.com serverside-applied logger.go:42: 03:26:30 | demand-backup/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/pgadmins.postgres-operator.crunchydata.com serverside-applied logger.go:42: 03:26:30 | demand-backup/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/pgupgrades.postgres-operator.crunchydata.com serverside-applied logger.go:42: 03:26:33 | demand-backup/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/postgresclusters.postgres-operator.crunchydata.com serverside-applied logger.go:42: 03:26:33 | demand-backup/0-deploy-operator | + kubectl -n pg-operator apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy/cw-rbac.yaml logger.go:42: 03:26:34 | demand-backup/0-deploy-operator | serviceaccount/percona-postgresql-operator serverside-applied logger.go:42: 03:26:34 | demand-backup/0-deploy-operator | clusterrole.rbac.authorization.k8s.io/percona-postgresql-operator serverside-applied logger.go:42: 03:26:34 | demand-backup/0-deploy-operator | clusterrolebinding.rbac.authorization.k8s.io/percona-postgresql-operator serverside-applied logger.go:42: 03:26:34 | demand-backup/0-deploy-operator | + local disable_telemetry=true logger.go:42: 03:26:34 | demand-backup/0-deploy-operator | + '[' demand-backup == telemetry-transfer ']' logger.go:42: 03:26:34 | demand-backup/0-deploy-operator | + yq eval '.spec.template.spec.containers[0].image = "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c"' /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy/cw-operator.yaml logger.go:42: 03:26:34 | demand-backup/0-deploy-operator | + yq eval '(.spec.template.spec.containers[] | select(.name=="operator") | .env[] | select(.name=="DISABLE_TELEMETRY") | .value) = "true"' - logger.go:42: 03:26:34 | demand-backup/0-deploy-operator | + yq eval '(.spec.template.spec.containers[] | select(.name=="operator") | .env[] | select(.name=="LOG_LEVEL") | .value) = "DEBUG"' - logger.go:42: 03:26:34 | demand-backup/0-deploy-operator | + yq eval '(.spec.template.spec.containers[] | select(.name=="operator") | .env[] | select(.name=="PGO_FEATURE_GATES") | .value) = ""' - logger.go:42: 03:26:34 | demand-backup/0-deploy-operator | + kubectl -n pg-operator apply -f - logger.go:42: 03:26:35 | demand-backup/0-deploy-operator | deployment.apps/percona-postgresql-operator created logger.go:42: 03:26:35 | demand-backup/0-deploy-operator | + deploy_client logger.go:42: 03:26:35 | demand-backup/0-deploy-operator | + kubectl -n kuttl-test-growing-killdeer apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/conf/client.yaml logger.go:42: 03:26:36 | demand-backup/0-deploy-operator | deployment.apps/pg-client created logger.go:42: 03:26:36 | demand-backup/0-deploy-operator | + deploy_s3_secrets logger.go:42: 03:26:36 | demand-backup/0-deploy-operator | + set +o xtrace logger.go:42: 03:26:38 | demand-backup/0-deploy-operator | secret/demand-backup-pgbackrest-secrets created logger.go:42: 03:26:39 | demand-backup/0-deploy-operator | running command: [sh -c kubectl assert exist-enhanced deployment percona-postgresql-operator -n ${OPERATOR_NS:-$NAMESPACE} --field-selector status.readyReplicas=1] logger.go:42: 03:26:39 | demand-backup/0-deploy-operator | ASSERT deployment percona-postgresql-operator matching field criteria 'status.readyReplicas=1' should exist. logger.go:42: 03:26:39 | demand-backup/0-deploy-operator | INFO Found 1 resource(s). logger.go:42: 03:26:39 | demand-backup/0-deploy-operator | NAME NAMESPACE COL0 logger.go:42: 03:26:39 | demand-backup/0-deploy-operator | percona-postgresql-operator pg-operator 1 logger.go:42: 03:26:39 | demand-backup/0-deploy-operator | ASSERT PASS logger.go:42: 03:26:39 | demand-backup/0-deploy-operator | test step completed 0-deploy-operator logger.go:42: 03:26:39 | demand-backup/1-create-cluster | starting test step 1-create-cluster logger.go:42: 03:26:39 | demand-backup/1-create-cluster | running command: [sh -c set -o errexit set -o xtrace source ../../functions get_cr "demand-backup" ${RANDOM} \ | yq '.metadata.finalizers=["percona.com/delete-backups"]' \ | yq '.spec.backups.pgbackrest.global.log-level-console="debug"' \ | yq '.spec.backups.pgbackrest.global.repo1-retention-full="2"' \ | yq '.spec.backups.pgbackrest.global.repo1-retention-full-type="count"' \ | yq '.spec.backups.pgbackrest.global.repo3-retention-full="2"' \ | yq '.spec.backups.pgbackrest.global.repo3-retention-full-type="count"' \ | yq '.spec.backups.pgbackrest.jobs.backoffLimit=20' \ | yq '.spec.backups.pgbackrest.jobs.restartPolicy="OnFailure"' \ | kubectl -n "${NAMESPACE}" apply -f -] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | + source ../../functions logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ realpath ../../.. logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++ CERT_MANAGER_VER=1.19.1 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++++ pwd logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++ test_name=demand-backup logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/vars.sh logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/conf logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/conf logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export TEST_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ TEST_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export TEMP_DIR=/tmp/kuttl/pg/demand-backup logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ TEMP_DIR=/tmp/kuttl/pg/demand-backup logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export GIT_BRANCH=PR-1506 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ GIT_BRANCH=PR-1506 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export VERSION=PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ VERSION=PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export SKIP_TEST_WARNINGS=true logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ SKIP_TEST_WARNINGS=true logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ command -v oc logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ ! -n '' ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ -n '' ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export PG_VER=18 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ PG_VER=18 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export BUCKET=pg-operator-testing logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ BUCKET=pg-operator-testing logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export IMAGE_PMM3_CLIENT=perconalab/pmm-client:3.4 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ IMAGE_PMM3_CLIENT=perconalab/pmm-client:3.4 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export IMAGE_PMM3_SERVER=perconalab/pmm-server:3.4 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ IMAGE_PMM3_SERVER=perconalab/pmm-server:3.4 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export PGOV1_TAG=1.4.0 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ PGOV1_TAG=1.4.0 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export PGOV1_VER=14 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ PGOV1_VER=14 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export MINIO_VER=5.4.0 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ MINIO_VER=5.4.0 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export VAULT_VER=0.32.0 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ VAULT_VER=0.32.0 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export REGISTRY_NAME=docker.io logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ REGISTRY_NAME=docker.io logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export REGISTRY_NAME_FULL=docker.io/ logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ REGISTRY_NAME_FULL=docker.io/ logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++++ printenv logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++++ grep -E '^IMAGE' logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++++ awk -F= '{print $1}' logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++++ eval 'echo $IMAGE_BACKREST' logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++++ echo perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ var_value=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == docker.io/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == percona/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == perconalab/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ echo IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++++ eval 'echo $IMAGE_PMM3_CLIENT' logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++++ echo perconalab/pmm-client:3.4 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ var_value=perconalab/pmm-client:3.4 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/pmm-client:3.4 == docker.io/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/pmm-client:3.4 == percona/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/pmm-client:3.4 == perconalab/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ new_value=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ echo IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++++ eval 'echo $IMAGE' logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ var_value=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == docker.io/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == percona/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == perconalab/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ new_value=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ echo IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++++ eval 'echo $IMAGE_UPGRADE' logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++++ echo perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ var_value=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == docker.io/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == percona/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == perconalab/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ echo IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:26:39 | demand-backup/1-create-cluster | IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++++ eval 'echo $IMAGE_BASE' logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++++ echo perconalab/percona-postgresql-operator logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ var_value=perconalab/percona-postgresql-operator logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/percona-postgresql-operator == docker.io/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/percona-postgresql-operator == percona/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/percona-postgresql-operator == perconalab/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ new_value=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ echo IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:26:39 | demand-backup/1-create-cluster | IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++++ eval 'echo $IMAGE_PMM3_SERVER' logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++++ echo perconalab/pmm-server:3.4 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ var_value=perconalab/pmm-server:3.4 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/pmm-server:3.4 == docker.io/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/pmm-server:3.4 == percona/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/pmm-server:3.4 == perconalab/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ new_value=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ echo IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++++ eval 'echo $IMAGE_PGBOUNCER' logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++++ echo perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ var_value=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == docker.io/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == percona/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == perconalab/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ echo IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++++ eval 'echo $IMAGE_POSTGRESQL' logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++++ echo perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ var_value=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == docker.io/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == percona/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == perconalab/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ echo IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:26:39 | demand-backup/1-create-cluster | IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++++ eval 'echo $IMAGE_PMM_SERVER' logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++++ echo perconalab/pmm-server:dev-latest logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ var_value=perconalab/pmm-server:dev-latest logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/pmm-server:dev-latest == docker.io/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/pmm-server:dev-latest == percona/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/pmm-server:dev-latest == perconalab/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ new_value=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ echo IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:26:39 | demand-backup/1-create-cluster | IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++++ eval 'echo $IMAGE_PMM_CLIENT' logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++++ echo perconalab/pmm-client:dev-latest logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ var_value=perconalab/pmm-client:dev-latest logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/pmm-client:dev-latest == docker.io/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/pmm-client:dev-latest == percona/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ [[ perconalab/pmm-client:dev-latest == perconalab/* ]] logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ new_value=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ export IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ echo IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:26:39 | demand-backup/1-create-cluster | IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++++ which gdate logger.go:42: 03:26:39 | demand-backup/1-create-cluster | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++++ which date logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ date=/usr/sbin/date logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++++ which gsed logger.go:42: 03:26:39 | demand-backup/1-create-cluster | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++++ which sed logger.go:42: 03:26:39 | demand-backup/1-create-cluster | +++ sed=/usr/sbin/sed logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++ oc get projects logger.go:42: 03:26:39 | demand-backup/1-create-cluster | + get_cr demand-backup 22413 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | + local cr_name=demand-backup logger.go:42: 03:26:39 | demand-backup/1-create-cluster | + local repo_path=22413 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | + local source_path= logger.go:42: 03:26:39 | demand-backup/1-create-cluster | + '[' -z demand-backup ']' logger.go:42: 03:26:39 | demand-backup/1-create-cluster | + yq '.metadata.finalizers=["percona.com/delete-backups"]' logger.go:42: 03:26:39 | demand-backup/1-create-cluster | + yq '.spec.backups.pgbackrest.global.log-level-console="debug"' logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++ detect_k8s_platform logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++ local platform= logger.go:42: 03:26:39 | demand-backup/1-create-cluster | + yq '.spec.backups.pgbackrest.global.repo1-retention-full="2"' logger.go:42: 03:26:39 | demand-backup/1-create-cluster | + yq '.spec.backups.pgbackrest.global.repo1-retention-full-type="count"' logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++ kubectl get nodes -o json logger.go:42: 03:26:39 | demand-backup/1-create-cluster | + yq '.spec.backups.pgbackrest.global.repo3-retention-full="2"' logger.go:42: 03:26:39 | demand-backup/1-create-cluster | + yq '.spec.backups.pgbackrest.global.repo3-retention-full-type="count"' logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++ jq -r '.items[0].spec.providerID' logger.go:42: 03:26:39 | demand-backup/1-create-cluster | + yq .spec.backups.pgbackrest.jobs.backoffLimit=20 logger.go:42: 03:26:39 | demand-backup/1-create-cluster | + kubectl -n kuttl-test-growing-killdeer apply -f - logger.go:42: 03:26:39 | demand-backup/1-create-cluster | ++ grep -q gce:// logger.go:42: 03:26:39 | demand-backup/1-create-cluster | + yq '.spec.backups.pgbackrest.jobs.restartPolicy="OnFailure"' logger.go:42: 03:26:40 | demand-backup/1-create-cluster | ++ platform=gke logger.go:42: 03:26:40 | demand-backup/1-create-cluster | ++ echo gke logger.go:42: 03:26:40 | demand-backup/1-create-cluster | + local platform=gke logger.go:42: 03:26:40 | demand-backup/1-create-cluster | + local cr_file=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf/demand-backup.yaml logger.go:42: 03:26:40 | demand-backup/1-create-cluster | + local spaces_cr_file=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf/demand-backup-spaces.yaml logger.go:42: 03:26:40 | demand-backup/1-create-cluster | + local aks_cr_file=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf/demand-backup-aks.yaml logger.go:42: 03:26:40 | demand-backup/1-create-cluster | + [[ gke == \d\i\g\i\t\a\l\o\c\e\a\n ]] logger.go:42: 03:26:40 | demand-backup/1-create-cluster | + [[ gke == \a\k\s ]] logger.go:42: 03:26:40 | demand-backup/1-create-cluster | + crs=('/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy/cr.yaml') logger.go:42: 03:26:40 | demand-backup/1-create-cluster | + local crs logger.go:42: 03:26:40 | demand-backup/1-create-cluster | + [[ -f /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf/demand-backup.yaml ]] logger.go:42: 03:26:40 | demand-backup/1-create-cluster | + crs+=($cr_file) logger.go:42: 03:26:40 | demand-backup/1-create-cluster | + yq eval-all ' logger.go:42: 03:26:40 | demand-backup/1-create-cluster | select(fileIndex == 0) * (select(fileIndex == 1) // {}) | logger.go:42: 03:26:40 | demand-backup/1-create-cluster | .metadata.name = "demand-backup" | logger.go:42: 03:26:40 | demand-backup/1-create-cluster | .metadata.labels = {"e2e":"demand-backup"} | logger.go:42: 03:26:40 | demand-backup/1-create-cluster | .spec.postgresVersion = 18 | logger.go:42: 03:26:40 | demand-backup/1-create-cluster | .spec.users += [{"name":"postgres","password":{"type":"AlphaNumeric"}}] | logger.go:42: 03:26:40 | demand-backup/1-create-cluster | .spec.users += [{"name":"demand-backup","password":{"type":"AlphaNumeric"}}] | logger.go:42: 03:26:40 | demand-backup/1-create-cluster | .spec.image = "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" | logger.go:42: 03:26:40 | demand-backup/1-create-cluster | .spec.initContainer.image = "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" | logger.go:42: 03:26:40 | demand-backup/1-create-cluster | .spec.backups.pgbackrest.image = "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" | logger.go:42: 03:26:40 | demand-backup/1-create-cluster | .spec.proxy.pgBouncer.image = "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18" | logger.go:42: 03:26:40 | demand-backup/1-create-cluster | .spec.pmm.image = "docker.io/perconalab/pmm-client:dev-latest" | logger.go:42: 03:26:40 | demand-backup/1-create-cluster | .spec.pmm.secret = "demand-backup-pmm-secret" | logger.go:42: 03:26:40 | demand-backup/1-create-cluster | .spec.pmm.customClusterName = "demand-backup-pmm-custom-name" | logger.go:42: 03:26:40 | demand-backup/1-create-cluster | .spec.pmm.postgresParams = "--environment=dev-postgres" logger.go:42: 03:26:40 | demand-backup/1-create-cluster | ' /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy/cr.yaml /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf/demand-backup.yaml logger.go:42: 03:26:40 | demand-backup/1-create-cluster | + /usr/sbin/sed -i 's||22413|g' /tmp/kuttl/pg/demand-backup/cr.yaml logger.go:42: 03:26:40 | demand-backup/1-create-cluster | + /usr/sbin/sed -i 's|||g' /tmp/kuttl/pg/demand-backup/cr.yaml logger.go:42: 03:26:40 | demand-backup/1-create-cluster | + /usr/sbin/sed -i 's||pg-operator-testing|g' /tmp/kuttl/pg/demand-backup/cr.yaml logger.go:42: 03:26:40 | demand-backup/1-create-cluster | + /usr/sbin/sed -i 's||docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c|g' /tmp/kuttl/pg/demand-backup/cr.yaml logger.go:42: 03:26:40 | demand-backup/1-create-cluster | + [[ -n '' ]] logger.go:42: 03:26:40 | demand-backup/1-create-cluster | + cat /tmp/kuttl/pg/demand-backup/cr.yaml logger.go:42: 03:26:41 | demand-backup/1-create-cluster | perconapgcluster.pgv2.percona.com/demand-backup created logger.go:42: 03:28:52 | demand-backup/1-create-cluster | test step completed 1-create-cluster logger.go:42: 03:28:52 | demand-backup/2-write-data | starting test step 2-write-data logger.go:42: 03:28:52 | demand-backup/2-write-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_psql_local \ 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' \ "postgres:$(get_psql_user_pass demand-backup-pguser-postgres)@$(get_psql_user_host demand-backup-pguser-postgres)" run_psql_local \ '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' \ "postgres:$(get_psql_user_pass demand-backup-pguser-postgres)@$(get_psql_user_host demand-backup-pguser-postgres)"] logger.go:42: 03:28:52 | demand-backup/2-write-data | + source ../../functions logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ realpath ../../.. logger.go:42: 03:28:52 | demand-backup/2-write-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:28:52 | demand-backup/2-write-data | ++ CERT_MANAGER_VER=1.19.1 logger.go:42: 03:28:52 | demand-backup/2-write-data | ++++ pwd logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup logger.go:42: 03:28:52 | demand-backup/2-write-data | ++ test_name=demand-backup logger.go:42: 03:28:52 | demand-backup/2-write-data | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/vars.sh logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/conf logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/conf logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export TEST_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ TEST_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export TEMP_DIR=/tmp/kuttl/pg/demand-backup logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ TEMP_DIR=/tmp/kuttl/pg/demand-backup logger.go:42: 03:28:52 | demand-backup/2-write-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export GIT_BRANCH=PR-1506 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ GIT_BRANCH=PR-1506 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export VERSION=PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ VERSION=PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export SKIP_TEST_WARNINGS=true logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ SKIP_TEST_WARNINGS=true logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ command -v oc logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ ! -n '' ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ -n '' ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export PG_VER=18 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ PG_VER=18 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export BUCKET=pg-operator-testing logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ BUCKET=pg-operator-testing logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export IMAGE_PMM3_CLIENT=perconalab/pmm-client:3.4 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ IMAGE_PMM3_CLIENT=perconalab/pmm-client:3.4 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export IMAGE_PMM3_SERVER=perconalab/pmm-server:3.4 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ IMAGE_PMM3_SERVER=perconalab/pmm-server:3.4 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export PGOV1_TAG=1.4.0 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ PGOV1_TAG=1.4.0 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export PGOV1_VER=14 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ PGOV1_VER=14 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export MINIO_VER=5.4.0 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ MINIO_VER=5.4.0 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export VAULT_VER=0.32.0 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ VAULT_VER=0.32.0 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export REGISTRY_NAME=docker.io logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ REGISTRY_NAME=docker.io logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export REGISTRY_NAME_FULL=docker.io/ logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ REGISTRY_NAME_FULL=docker.io/ logger.go:42: 03:28:52 | demand-backup/2-write-data | ++++ printenv logger.go:42: 03:28:52 | demand-backup/2-write-data | ++++ grep -E '^IMAGE' logger.go:42: 03:28:52 | demand-backup/2-write-data | ++++ awk -F= '{print $1}' logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:28:52 | demand-backup/2-write-data | ++++ eval 'echo $IMAGE_BACKREST' logger.go:42: 03:28:52 | demand-backup/2-write-data | +++++ echo perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ var_value=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == docker.io/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == percona/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == perconalab/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ echo IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:28:52 | demand-backup/2-write-data | IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:28:52 | demand-backup/2-write-data | ++++ eval 'echo $IMAGE_PMM3_CLIENT' logger.go:42: 03:28:52 | demand-backup/2-write-data | +++++ echo perconalab/pmm-client:3.4 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ var_value=perconalab/pmm-client:3.4 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/pmm-client:3.4 == docker.io/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/pmm-client:3.4 == percona/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/pmm-client:3.4 == perconalab/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ new_value=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ echo IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:28:52 | demand-backup/2-write-data | IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:28:52 | demand-backup/2-write-data | ++++ eval 'echo $IMAGE' logger.go:42: 03:28:52 | demand-backup/2-write-data | +++++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ var_value=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == docker.io/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == percona/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == perconalab/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ new_value=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ echo IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:28:52 | demand-backup/2-write-data | ++++ eval 'echo $IMAGE_UPGRADE' logger.go:42: 03:28:52 | demand-backup/2-write-data | +++++ echo perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ var_value=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == docker.io/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == percona/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == perconalab/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ echo IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:28:52 | demand-backup/2-write-data | IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:28:52 | demand-backup/2-write-data | ++++ eval 'echo $IMAGE_BASE' logger.go:42: 03:28:52 | demand-backup/2-write-data | +++++ echo perconalab/percona-postgresql-operator logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ var_value=perconalab/percona-postgresql-operator logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/percona-postgresql-operator == docker.io/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/percona-postgresql-operator == percona/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/percona-postgresql-operator == perconalab/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ new_value=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ echo IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:28:52 | demand-backup/2-write-data | IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:28:52 | demand-backup/2-write-data | ++++ eval 'echo $IMAGE_PMM3_SERVER' logger.go:42: 03:28:52 | demand-backup/2-write-data | +++++ echo perconalab/pmm-server:3.4 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ var_value=perconalab/pmm-server:3.4 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/pmm-server:3.4 == docker.io/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/pmm-server:3.4 == percona/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/pmm-server:3.4 == perconalab/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ new_value=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ echo IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:28:52 | demand-backup/2-write-data | IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:28:52 | demand-backup/2-write-data | ++++ eval 'echo $IMAGE_PGBOUNCER' logger.go:42: 03:28:52 | demand-backup/2-write-data | +++++ echo perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ var_value=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == docker.io/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == percona/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == perconalab/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ echo IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:28:52 | demand-backup/2-write-data | IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:28:52 | demand-backup/2-write-data | ++++ eval 'echo $IMAGE_POSTGRESQL' logger.go:42: 03:28:52 | demand-backup/2-write-data | +++++ echo perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ var_value=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == docker.io/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == percona/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == perconalab/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ echo IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:28:52 | demand-backup/2-write-data | IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:28:52 | demand-backup/2-write-data | ++++ eval 'echo $IMAGE_PMM_SERVER' logger.go:42: 03:28:52 | demand-backup/2-write-data | +++++ echo perconalab/pmm-server:dev-latest logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ var_value=perconalab/pmm-server:dev-latest logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/pmm-server:dev-latest == docker.io/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/pmm-server:dev-latest == percona/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/pmm-server:dev-latest == perconalab/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ new_value=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ echo IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:28:52 | demand-backup/2-write-data | IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:28:52 | demand-backup/2-write-data | ++++ eval 'echo $IMAGE_PMM_CLIENT' logger.go:42: 03:28:52 | demand-backup/2-write-data | +++++ echo perconalab/pmm-client:dev-latest logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ var_value=perconalab/pmm-client:dev-latest logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/pmm-client:dev-latest == docker.io/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/pmm-client:dev-latest == percona/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ [[ perconalab/pmm-client:dev-latest == perconalab/* ]] logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ new_value=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ export IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ echo IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:28:52 | demand-backup/2-write-data | IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:52 | demand-backup/2-write-data | ++++ which gdate logger.go:42: 03:28:52 | demand-backup/2-write-data | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 03:28:52 | demand-backup/2-write-data | ++++ which date logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ date=/usr/sbin/date logger.go:42: 03:28:52 | demand-backup/2-write-data | ++++ which gsed logger.go:42: 03:28:52 | demand-backup/2-write-data | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 03:28:52 | demand-backup/2-write-data | ++++ which sed logger.go:42: 03:28:52 | demand-backup/2-write-data | +++ sed=/usr/sbin/sed logger.go:42: 03:28:52 | demand-backup/2-write-data | ++ oc get projects logger.go:42: 03:28:52 | demand-backup/2-write-data | ++ get_psql_user_pass demand-backup-pguser-postgres logger.go:42: 03:28:52 | demand-backup/2-write-data | ++ local secret_name=demand-backup-pguser-postgres logger.go:42: 03:28:52 | demand-backup/2-write-data | ++ kubectl -n kuttl-test-growing-killdeer get secret/demand-backup-pguser-postgres '--template={{.data.password | base64decode}}' logger.go:42: 03:28:52 | demand-backup/2-write-data | ++ get_psql_user_host demand-backup-pguser-postgres logger.go:42: 03:28:52 | demand-backup/2-write-data | ++ local secret_name=demand-backup-pguser-postgres logger.go:42: 03:28:52 | demand-backup/2-write-data | ++ kubectl -n kuttl-test-growing-killdeer get secret/demand-backup-pguser-postgres '--template={{.data.host | base64decode }}' logger.go:42: 03:28:53 | demand-backup/2-write-data | + run_psql_local 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:N7HeCfaqv9X8trS8vJrCEsWf@demand-backup-primary.kuttl-test-growing-killdeer.svc logger.go:42: 03:28:53 | demand-backup/2-write-data | + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' logger.go:42: 03:28:53 | demand-backup/2-write-data | + local uri=postgres:N7HeCfaqv9X8trS8vJrCEsWf@demand-backup-primary.kuttl-test-growing-killdeer.svc logger.go:42: 03:28:53 | demand-backup/2-write-data | + local driver=postgres logger.go:42: 03:28:53 | demand-backup/2-write-data | ++ get_client_pod logger.go:42: 03:28:53 | demand-backup/2-write-data | ++ kubectl -n kuttl-test-growing-killdeer get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 03:28:53 | demand-backup/2-write-data | + kubectl -n kuttl-test-growing-killdeer exec pg-client-65d98588cc-8jjx9 -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://'\''postgres:N7HeCfaqv9X8trS8vJrCEsWf@demand-backup-primary.kuttl-test-growing-killdeer.svc'\''' logger.go:42: 03:28:54 | demand-backup/2-write-data | ++ get_psql_user_pass demand-backup-pguser-postgres logger.go:42: 03:28:54 | demand-backup/2-write-data | ++ local secret_name=demand-backup-pguser-postgres logger.go:42: 03:28:54 | demand-backup/2-write-data | ++ kubectl -n kuttl-test-growing-killdeer get secret/demand-backup-pguser-postgres '--template={{.data.password | base64decode}}' logger.go:42: 03:28:55 | demand-backup/2-write-data | ++ get_psql_user_host demand-backup-pguser-postgres logger.go:42: 03:28:55 | demand-backup/2-write-data | ++ local secret_name=demand-backup-pguser-postgres logger.go:42: 03:28:55 | demand-backup/2-write-data | ++ kubectl -n kuttl-test-growing-killdeer get secret/demand-backup-pguser-postgres '--template={{.data.host | base64decode }}' logger.go:42: 03:28:55 | demand-backup/2-write-data | + run_psql_local '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' postgres:N7HeCfaqv9X8trS8vJrCEsWf@demand-backup-primary.kuttl-test-growing-killdeer.svc logger.go:42: 03:28:55 | demand-backup/2-write-data | + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)' logger.go:42: 03:28:55 | demand-backup/2-write-data | + local uri=postgres:N7HeCfaqv9X8trS8vJrCEsWf@demand-backup-primary.kuttl-test-growing-killdeer.svc logger.go:42: 03:28:55 | demand-backup/2-write-data | + local driver=postgres logger.go:42: 03:28:55 | demand-backup/2-write-data | ++ get_client_pod logger.go:42: 03:28:55 | demand-backup/2-write-data | ++ kubectl -n kuttl-test-growing-killdeer get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 03:28:56 | demand-backup/2-write-data | + kubectl -n kuttl-test-growing-killdeer exec pg-client-65d98588cc-8jjx9 -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://'\''postgres:N7HeCfaqv9X8trS8vJrCEsWf@demand-backup-primary.kuttl-test-growing-killdeer.svc'\''' logger.go:42: 03:28:57 | demand-backup/2-write-data | test step completed 2-write-data logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | starting test step 3-read-from-primary logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | running command: [sh -c set -o errexit set -o xtrace source ../../functions data=$(run_psql_local '\c myapp \\\ SELECT * from myApp;' "postgres:$(get_psql_user_pass demand-backup-pguser-postgres)@$(get_psql_user_host demand-backup-pguser-postgres)") kubectl create configmap -n "${NAMESPACE}" 03-read-from-primary --from-literal=data="${data}"] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | + source ../../functions logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ realpath ../../.. logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | ++ CERT_MANAGER_VER=1.19.1 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | ++++ pwd logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | ++ test_name=demand-backup logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/vars.sh logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/conf logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/conf logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export TEST_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ TEST_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export TEMP_DIR=/tmp/kuttl/pg/demand-backup logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ TEMP_DIR=/tmp/kuttl/pg/demand-backup logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export GIT_BRANCH=PR-1506 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ GIT_BRANCH=PR-1506 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export VERSION=PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ VERSION=PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export SKIP_TEST_WARNINGS=true logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ SKIP_TEST_WARNINGS=true logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ command -v oc logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ ! -n '' ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ -n '' ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export PG_VER=18 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ PG_VER=18 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export BUCKET=pg-operator-testing logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ BUCKET=pg-operator-testing logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export IMAGE_PMM3_CLIENT=perconalab/pmm-client:3.4 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ IMAGE_PMM3_CLIENT=perconalab/pmm-client:3.4 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export IMAGE_PMM3_SERVER=perconalab/pmm-server:3.4 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ IMAGE_PMM3_SERVER=perconalab/pmm-server:3.4 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export PGOV1_TAG=1.4.0 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ PGOV1_TAG=1.4.0 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export PGOV1_VER=14 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ PGOV1_VER=14 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export MINIO_VER=5.4.0 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ MINIO_VER=5.4.0 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export VAULT_VER=0.32.0 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ VAULT_VER=0.32.0 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export REGISTRY_NAME=docker.io logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ REGISTRY_NAME=docker.io logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export REGISTRY_NAME_FULL=docker.io/ logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ REGISTRY_NAME_FULL=docker.io/ logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | ++++ printenv logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | ++++ grep -E '^IMAGE' logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | ++++ awk -F= '{print $1}' logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | ++++ eval 'echo $IMAGE_BACKREST' logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++++ echo perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ var_value=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == docker.io/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == percona/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == perconalab/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ echo IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | ++++ eval 'echo $IMAGE_PMM3_CLIENT' logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++++ echo perconalab/pmm-client:3.4 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ var_value=perconalab/pmm-client:3.4 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/pmm-client:3.4 == docker.io/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/pmm-client:3.4 == percona/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/pmm-client:3.4 == perconalab/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ new_value=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ echo IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | ++++ eval 'echo $IMAGE' logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ var_value=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == docker.io/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == percona/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == perconalab/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ new_value=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ echo IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | ++++ eval 'echo $IMAGE_UPGRADE' logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++++ echo perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ var_value=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == docker.io/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == percona/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == perconalab/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ echo IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | ++++ eval 'echo $IMAGE_BASE' logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++++ echo perconalab/percona-postgresql-operator logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ var_value=perconalab/percona-postgresql-operator logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/percona-postgresql-operator == docker.io/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/percona-postgresql-operator == percona/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/percona-postgresql-operator == perconalab/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ new_value=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ echo IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | ++++ eval 'echo $IMAGE_PMM3_SERVER' logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++++ echo perconalab/pmm-server:3.4 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ var_value=perconalab/pmm-server:3.4 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/pmm-server:3.4 == docker.io/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/pmm-server:3.4 == percona/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/pmm-server:3.4 == perconalab/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ new_value=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ echo IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | ++++ eval 'echo $IMAGE_PGBOUNCER' logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++++ echo perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ var_value=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == docker.io/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == percona/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == perconalab/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ echo IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | ++++ eval 'echo $IMAGE_POSTGRESQL' logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++++ echo perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ var_value=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == docker.io/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == percona/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == perconalab/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ echo IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | ++++ eval 'echo $IMAGE_PMM_SERVER' logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++++ echo perconalab/pmm-server:dev-latest logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ var_value=perconalab/pmm-server:dev-latest logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/pmm-server:dev-latest == docker.io/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/pmm-server:dev-latest == percona/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/pmm-server:dev-latest == perconalab/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ new_value=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ echo IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | ++++ eval 'echo $IMAGE_PMM_CLIENT' logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++++ echo perconalab/pmm-client:dev-latest logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ var_value=perconalab/pmm-client:dev-latest logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/pmm-client:dev-latest == docker.io/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/pmm-client:dev-latest == percona/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ [[ perconalab/pmm-client:dev-latest == perconalab/* ]] logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ new_value=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ export IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ echo IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | ++++ which gdate logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | ++++ which date logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ date=/usr/sbin/date logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | ++++ which gsed logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | ++++ which sed logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ sed=/usr/sbin/sed logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | ++ oc get projects logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ get_psql_user_pass demand-backup-pguser-postgres logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ local secret_name=demand-backup-pguser-postgres logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ kubectl -n kuttl-test-growing-killdeer get secret/demand-backup-pguser-postgres '--template={{.data.password | base64decode}}' logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ get_psql_user_host demand-backup-pguser-postgres logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ local secret_name=demand-backup-pguser-postgres logger.go:42: 03:28:57 | demand-backup/3-read-from-primary | +++ kubectl -n kuttl-test-growing-killdeer get secret/demand-backup-pguser-postgres '--template={{.data.host | base64decode }}' logger.go:42: 03:28:58 | demand-backup/3-read-from-primary | ++ run_psql_local '\c myapp \\\ SELECT * from myApp;' postgres:N7HeCfaqv9X8trS8vJrCEsWf@demand-backup-primary.kuttl-test-growing-killdeer.svc logger.go:42: 03:28:58 | demand-backup/3-read-from-primary | ++ local 'command=\c myapp \\\ SELECT * from myApp;' logger.go:42: 03:28:58 | demand-backup/3-read-from-primary | ++ local uri=postgres:N7HeCfaqv9X8trS8vJrCEsWf@demand-backup-primary.kuttl-test-growing-killdeer.svc logger.go:42: 03:28:58 | demand-backup/3-read-from-primary | ++ local driver=postgres logger.go:42: 03:28:58 | demand-backup/3-read-from-primary | +++ get_client_pod logger.go:42: 03:28:58 | demand-backup/3-read-from-primary | +++ kubectl -n kuttl-test-growing-killdeer get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 03:28:58 | demand-backup/3-read-from-primary | ++ kubectl -n kuttl-test-growing-killdeer exec pg-client-65d98588cc-8jjx9 -- bash -c 'printf '\''\c myapp \\\ SELECT * from myApp;\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://'\''postgres:N7HeCfaqv9X8trS8vJrCEsWf@demand-backup-primary.kuttl-test-growing-killdeer.svc'\''' logger.go:42: 03:28:59 | demand-backup/3-read-from-primary | + data=' 100500' logger.go:42: 03:28:59 | demand-backup/3-read-from-primary | + kubectl create configmap -n kuttl-test-growing-killdeer 03-read-from-primary '--from-literal=data= 100500' logger.go:42: 03:29:00 | demand-backup/3-read-from-primary | configmap/03-read-from-primary created logger.go:42: 03:29:00 | demand-backup/3-read-from-primary | test step completed 3-read-from-primary logger.go:42: 03:29:00 | demand-backup/4-create-backup-repo1 | starting test step 4-create-backup-repo1 logger.go:42: 03:29:01 | demand-backup/4-create-backup-repo1 | PerconaPGBackup:kuttl-test-growing-killdeer/demand-backup-full-repo1 created logger.go:42: 03:32:53 | demand-backup/4-create-backup-repo1 | test step completed 4-create-backup-repo1 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | starting test step 5-check-pgbackrest-info-repo1 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | running command: [sh -c set -o errexit set -o xtrace source ../../functions instance=$(kubectl get -n "${NAMESPACE}" pod -l postgres-operator.crunchydata.com/instance-set=instance1 -o 'jsonpath={.items[].metadata.name}') pgbackrest_info_backups=$(kubectl exec -n "${NAMESPACE}" "$instance" -c database -- pgbackrest info --output json --log-level-console=info | jq '.[0].backup[]') check_backup() { local backup_name=$1 local pgbackrest_annotation=$2 local pgbackrest_annotation_value=$3 status_backup_name=$(kubectl get -n "${NAMESPACE}" pg-backup "$backup_name" -o jsonpath='{.status.backupName}') if [[ -z $status_backup_name ]]; then echo ".status.backupName is empty in $backup_name" exit 1 fi backup_info=$(echo "$pgbackrest_info_backups" | jq "select(.annotation.\"$pgbackrest_annotation\" == \"$pgbackrest_annotation_value\")") if [[ $status_backup_name != $(echo "$backup_info" | jq ".label" --raw-output) ]]; then echo ".status.backupName doesn't equal to label in pgbackrest info" exit 1 fi backup_job_name=$(kubectl get -n "${NAMESPACE}" pg-backup "$backup_name" -o jsonpath='{.status.jobName}') backup_job_annotation=$(echo "$backup_info" | jq '.annotation."percona.com/backup-job-name"' --raw-output) if [[ $backup_job_name != "$backup_job_annotation" ]]; then echo "Failed to get job name annotation from pgbackrest" exit 1 fi } manual_backup_name="demand-backup-full-repo1" check_backup "$manual_backup_name" "percona.com/backup-name" "$manual_backup_name" replica_backup_name=$(kubectl get -n "${NAMESPACE}" pg-backup -o jsonpath='{.items[?(@.metadata.annotations.pgv2\.percona\.com/pgbackrest-backup-job-type=="replica-create")].metadata.name}') check_backup "$replica_backup_name" "percona.com/backup-job-type" "replica-create"] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | + source ../../functions logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ realpath ../../.. logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | ++ CERT_MANAGER_VER=1.19.1 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | ++++ pwd logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | ++ test_name=demand-backup logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/vars.sh logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/conf logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/conf logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export TEST_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ TEST_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export TEMP_DIR=/tmp/kuttl/pg/demand-backup logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ TEMP_DIR=/tmp/kuttl/pg/demand-backup logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export GIT_BRANCH=PR-1506 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ GIT_BRANCH=PR-1506 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export VERSION=PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ VERSION=PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export SKIP_TEST_WARNINGS=true logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ SKIP_TEST_WARNINGS=true logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ command -v oc logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ ! -n '' ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ -n '' ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export PG_VER=18 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ PG_VER=18 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export BUCKET=pg-operator-testing logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ BUCKET=pg-operator-testing logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export IMAGE_PMM3_CLIENT=perconalab/pmm-client:3.4 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ IMAGE_PMM3_CLIENT=perconalab/pmm-client:3.4 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export IMAGE_PMM3_SERVER=perconalab/pmm-server:3.4 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ IMAGE_PMM3_SERVER=perconalab/pmm-server:3.4 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export PGOV1_TAG=1.4.0 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ PGOV1_TAG=1.4.0 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export PGOV1_VER=14 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ PGOV1_VER=14 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export MINIO_VER=5.4.0 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ MINIO_VER=5.4.0 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export VAULT_VER=0.32.0 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ VAULT_VER=0.32.0 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export REGISTRY_NAME=docker.io logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ REGISTRY_NAME=docker.io logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export REGISTRY_NAME_FULL=docker.io/ logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ REGISTRY_NAME_FULL=docker.io/ logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | ++++ printenv logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | ++++ grep -E '^IMAGE' logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | ++++ awk -F= '{print $1}' logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | ++++ eval 'echo $IMAGE_BACKREST' logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++++ echo perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ var_value=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == docker.io/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == percona/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == perconalab/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ echo IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | ++++ eval 'echo $IMAGE_PMM3_CLIENT' logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++++ echo perconalab/pmm-client:3.4 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ var_value=perconalab/pmm-client:3.4 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/pmm-client:3.4 == docker.io/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/pmm-client:3.4 == percona/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/pmm-client:3.4 == perconalab/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ new_value=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ echo IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | ++++ eval 'echo $IMAGE' logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ var_value=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == docker.io/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == percona/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == perconalab/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ new_value=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ echo IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | ++++ eval 'echo $IMAGE_UPGRADE' logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++++ echo perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ var_value=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == docker.io/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == percona/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == perconalab/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ echo IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | ++++ eval 'echo $IMAGE_BASE' logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++++ echo perconalab/percona-postgresql-operator logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ var_value=perconalab/percona-postgresql-operator logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/percona-postgresql-operator == docker.io/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/percona-postgresql-operator == percona/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/percona-postgresql-operator == perconalab/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ new_value=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ echo IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | ++++ eval 'echo $IMAGE_PMM3_SERVER' logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++++ echo perconalab/pmm-server:3.4 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ var_value=perconalab/pmm-server:3.4 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/pmm-server:3.4 == docker.io/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/pmm-server:3.4 == percona/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/pmm-server:3.4 == perconalab/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ new_value=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ echo IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | ++++ eval 'echo $IMAGE_PGBOUNCER' logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++++ echo perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ var_value=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == docker.io/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == percona/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == perconalab/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ echo IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | ++++ eval 'echo $IMAGE_POSTGRESQL' logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++++ echo perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ var_value=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == docker.io/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == percona/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == perconalab/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ echo IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | ++++ eval 'echo $IMAGE_PMM_SERVER' logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++++ echo perconalab/pmm-server:dev-latest logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ var_value=perconalab/pmm-server:dev-latest logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/pmm-server:dev-latest == docker.io/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/pmm-server:dev-latest == percona/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/pmm-server:dev-latest == perconalab/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ new_value=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ echo IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | ++++ eval 'echo $IMAGE_PMM_CLIENT' logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++++ echo perconalab/pmm-client:dev-latest logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ var_value=perconalab/pmm-client:dev-latest logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/pmm-client:dev-latest == docker.io/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/pmm-client:dev-latest == percona/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ [[ perconalab/pmm-client:dev-latest == perconalab/* ]] logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ new_value=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ export IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ echo IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | ++++ which gdate logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | ++++ which date logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ date=/usr/sbin/date logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | ++++ which gsed logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | ++++ which sed logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | +++ sed=/usr/sbin/sed logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | ++ oc get projects logger.go:42: 03:32:53 | demand-backup/5-check-pgbackrest-info-repo1 | ++ kubectl get -n kuttl-test-growing-killdeer pod -l postgres-operator.crunchydata.com/instance-set=instance1 -o 'jsonpath={.items[].metadata.name}' logger.go:42: 03:32:54 | demand-backup/5-check-pgbackrest-info-repo1 | + instance=demand-backup-instance1-5kbg-0 logger.go:42: 03:32:54 | demand-backup/5-check-pgbackrest-info-repo1 | ++ kubectl exec -n kuttl-test-growing-killdeer demand-backup-instance1-5kbg-0 -c database -- pgbackrest info --output json --log-level-console=info logger.go:42: 03:32:54 | demand-backup/5-check-pgbackrest-info-repo1 | ++ jq '.[0].backup[]' logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | + pgbackrest_info_backups='{ logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "annotation": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "percona.com/backup-job-name": "demand-backup-backup-dn8b", logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "percona.com/backup-job-type": "replica-create" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "archive": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "start": "000000010000000000000005", logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": "000000010000000000000006" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "backrest": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "format": 5, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "version": "2.58.0" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "database": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "id": 1, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "repo-key": 1 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "error": false, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "info": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "delta": 24276563, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "repository": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "delta": 3133041, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "size": 3133041 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "size": 24276563 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "label": "20260319-032710F", logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "lsn": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "start": "0/5000028", logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": "0/6000088" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "prior": null, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "reference": null, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "timestamp": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "start": 1773890830, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": 1773890924 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "type": "full" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | } logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "annotation": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "percona.com/backup-job-name": "demand-backup-backup-877f", logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "percona.com/backup-name": "demand-backup-full-repo1" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "archive": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "start": "000000010000000000000008", logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": "00000001000000000000000A" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "backrest": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "format": 5, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "version": "2.58.0" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "database": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "id": 1, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "repo-key": 1 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "error": false, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "info": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "delta": 32411538, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "repository": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "delta": 4214354, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "size": 4214354 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "size": 32411538 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "label": "20260319-032906F", logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "lsn": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "start": "0/8000028", logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": "0/A000050" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "prior": null, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "reference": null, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "timestamp": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "start": 1773890946, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": 1773891152 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "type": "full" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }' logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | + manual_backup_name=demand-backup-full-repo1 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | + check_backup demand-backup-full-repo1 percona.com/backup-name demand-backup-full-repo1 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | + local backup_name=demand-backup-full-repo1 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | + local pgbackrest_annotation=percona.com/backup-name logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | + local pgbackrest_annotation_value=demand-backup-full-repo1 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | ++ kubectl get -n kuttl-test-growing-killdeer pg-backup demand-backup-full-repo1 -o 'jsonpath={.status.backupName}' logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | + status_backup_name=20260319-032906F logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | + [[ -z 20260319-032906F ]] logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | ++ echo '{ logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "annotation": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "percona.com/backup-job-name": "demand-backup-backup-dn8b", logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "percona.com/backup-job-type": "replica-create" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "archive": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "start": "000000010000000000000005", logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": "000000010000000000000006" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "backrest": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "format": 5, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "version": "2.58.0" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "database": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "id": 1, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "repo-key": 1 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "error": false, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "info": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "delta": 24276563, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "repository": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "delta": 3133041, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "size": 3133041 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "size": 24276563 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "label": "20260319-032710F", logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "lsn": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "start": "0/5000028", logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": "0/6000088" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "prior": null, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "reference": null, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "timestamp": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "start": 1773890830, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": 1773890924 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "type": "full" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | } logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "annotation": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "percona.com/backup-job-name": "demand-backup-backup-877f", logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "percona.com/backup-name": "demand-backup-full-repo1" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "archive": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "start": "000000010000000000000008", logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": "00000001000000000000000A" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "backrest": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "format": 5, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "version": "2.58.0" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "database": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "id": 1, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "repo-key": 1 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "error": false, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "info": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "delta": 32411538, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "repository": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "delta": 4214354, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "size": 4214354 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "size": 32411538 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "label": "20260319-032906F", logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "lsn": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "start": "0/8000028", logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": "0/A000050" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "prior": null, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "reference": null, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "timestamp": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "start": 1773890946, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": 1773891152 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "type": "full" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }' logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | ++ jq 'select(.annotation."percona.com/backup-name" == "demand-backup-full-repo1")' logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | + backup_info='{ logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "annotation": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "percona.com/backup-job-name": "demand-backup-backup-877f", logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "percona.com/backup-name": "demand-backup-full-repo1" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "archive": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "start": "000000010000000000000008", logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": "00000001000000000000000A" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "backrest": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "format": 5, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "version": "2.58.0" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "database": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "id": 1, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "repo-key": 1 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "error": false, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "info": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "delta": 32411538, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "repository": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "delta": 4214354, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "size": 4214354 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "size": 32411538 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "label": "20260319-032906F", logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "lsn": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "start": "0/8000028", logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": "0/A000050" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "prior": null, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "reference": null, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "timestamp": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "start": 1773890946, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": 1773891152 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "type": "full" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }' logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | ++ echo '{ logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "annotation": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "percona.com/backup-job-name": "demand-backup-backup-877f", logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "percona.com/backup-name": "demand-backup-full-repo1" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "archive": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "start": "000000010000000000000008", logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": "00000001000000000000000A" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "backrest": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "format": 5, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "version": "2.58.0" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "database": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "id": 1, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "repo-key": 1 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "error": false, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "info": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "delta": 32411538, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "repository": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "delta": 4214354, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "size": 4214354 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "size": 32411538 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "label": "20260319-032906F", logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "lsn": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "start": "0/8000028", logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": "0/A000050" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "prior": null, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "reference": null, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "timestamp": { logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "start": 1773890946, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": 1773891152 logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | "type": "full" logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | }' logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | ++ jq .label --raw-output logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | + [[ 20260319-032906F != 20260319-032906F ]] logger.go:42: 03:32:56 | demand-backup/5-check-pgbackrest-info-repo1 | ++ kubectl get -n kuttl-test-growing-killdeer pg-backup demand-backup-full-repo1 -o 'jsonpath={.status.jobName}' logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | + backup_job_name=demand-backup-backup-877f logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | ++ echo '{ logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "annotation": { logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "percona.com/backup-job-name": "demand-backup-backup-877f", logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "percona.com/backup-name": "demand-backup-full-repo1" logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "archive": { logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "start": "000000010000000000000008", logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": "00000001000000000000000A" logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "backrest": { logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "format": 5, logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "version": "2.58.0" logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "database": { logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "id": 1, logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "repo-key": 1 logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "error": false, logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "info": { logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "delta": 32411538, logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "repository": { logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "delta": 4214354, logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "size": 4214354 logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "size": 32411538 logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "label": "20260319-032906F", logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "lsn": { logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "start": "0/8000028", logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": "0/A000050" logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "prior": null, logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "reference": null, logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "timestamp": { logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "start": 1773890946, logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": 1773891152 logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | "type": "full" logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | }' logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | ++ jq '.annotation."percona.com/backup-job-name"' --raw-output logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | + backup_job_annotation=demand-backup-backup-877f logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | + [[ demand-backup-backup-877f != \d\e\m\a\n\d\-\b\a\c\k\u\p\-\b\a\c\k\u\p\-\8\7\7\f ]] logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | ++ kubectl get -n kuttl-test-growing-killdeer pg-backup -o 'jsonpath={.items[?(@.metadata.annotations.pgv2\.percona\.com/pgbackrest-backup-job-type=="replica-create")].metadata.name}' logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | + replica_backup_name=demand-backup-backup-dn8b-n2nxr logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | + check_backup demand-backup-backup-dn8b-n2nxr percona.com/backup-job-type replica-create logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | + local backup_name=demand-backup-backup-dn8b-n2nxr logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | + local pgbackrest_annotation=percona.com/backup-job-type logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | + local pgbackrest_annotation_value=replica-create logger.go:42: 03:32:57 | demand-backup/5-check-pgbackrest-info-repo1 | ++ kubectl get -n kuttl-test-growing-killdeer pg-backup demand-backup-backup-dn8b-n2nxr -o 'jsonpath={.status.backupName}' logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | + status_backup_name=20260319-032710F logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | + [[ -z 20260319-032710F ]] logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | ++ echo '{ logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "annotation": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "percona.com/backup-job-name": "demand-backup-backup-dn8b", logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "percona.com/backup-job-type": "replica-create" logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "archive": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "start": "000000010000000000000005", logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": "000000010000000000000006" logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "backrest": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "format": 5, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "version": "2.58.0" logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "database": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "id": 1, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "repo-key": 1 logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "error": false, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "info": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "delta": 24276563, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "repository": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "delta": 3133041, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "size": 3133041 logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "size": 24276563 logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "label": "20260319-032710F", logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "lsn": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "start": "0/5000028", logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": "0/6000088" logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "prior": null, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "reference": null, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "timestamp": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "start": 1773890830, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": 1773890924 logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "type": "full" logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | } logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "annotation": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "percona.com/backup-job-name": "demand-backup-backup-877f", logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "percona.com/backup-name": "demand-backup-full-repo1" logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "archive": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "start": "000000010000000000000008", logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": "00000001000000000000000A" logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "backrest": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "format": 5, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "version": "2.58.0" logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "database": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "id": 1, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "repo-key": 1 logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "error": false, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "info": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "delta": 32411538, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "repository": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "delta": 4214354, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "size": 4214354 logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "size": 32411538 logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "label": "20260319-032906F", logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "lsn": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "start": "0/8000028", logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": "0/A000050" logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "prior": null, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "reference": null, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "timestamp": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "start": 1773890946, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": 1773891152 logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "type": "full" logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }' logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | ++ jq 'select(.annotation."percona.com/backup-job-type" == "replica-create")' logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | + backup_info='{ logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "annotation": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "percona.com/backup-job-name": "demand-backup-backup-dn8b", logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "percona.com/backup-job-type": "replica-create" logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "archive": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "start": "000000010000000000000005", logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": "000000010000000000000006" logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "backrest": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "format": 5, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "version": "2.58.0" logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "database": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "id": 1, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "repo-key": 1 logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "error": false, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "info": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "delta": 24276563, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "repository": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "delta": 3133041, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "size": 3133041 logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "size": 24276563 logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "label": "20260319-032710F", logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "lsn": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "start": "0/5000028", logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": "0/6000088" logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "prior": null, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "reference": null, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "timestamp": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "start": 1773890830, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": 1773890924 logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "type": "full" logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }' logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | ++ echo '{ logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "annotation": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "percona.com/backup-job-name": "demand-backup-backup-dn8b", logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "percona.com/backup-job-type": "replica-create" logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "archive": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "start": "000000010000000000000005", logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": "000000010000000000000006" logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "backrest": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "format": 5, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "version": "2.58.0" logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "database": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "id": 1, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "repo-key": 1 logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "error": false, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "info": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "delta": 24276563, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "repository": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "delta": 3133041, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "size": 3133041 logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "size": 24276563 logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "label": "20260319-032710F", logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "lsn": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "start": "0/5000028", logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": "0/6000088" logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "prior": null, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "reference": null, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "timestamp": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "start": 1773890830, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": 1773890924 logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "type": "full" logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }' logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | ++ jq .label --raw-output logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | + [[ 20260319-032710F != 20260319-032710F ]] logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | ++ kubectl get -n kuttl-test-growing-killdeer pg-backup demand-backup-backup-dn8b-n2nxr -o 'jsonpath={.status.jobName}' logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | + backup_job_name=demand-backup-backup-dn8b logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | ++ echo '{ logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "annotation": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "percona.com/backup-job-name": "demand-backup-backup-dn8b", logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "percona.com/backup-job-type": "replica-create" logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "archive": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "start": "000000010000000000000005", logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": "000000010000000000000006" logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "backrest": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "format": 5, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "version": "2.58.0" logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "database": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "id": 1, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "repo-key": 1 logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "error": false, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "info": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "delta": 24276563, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "repository": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "delta": 3133041, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "size": 3133041 logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "size": 24276563 logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "label": "20260319-032710F", logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "lsn": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "start": "0/5000028", logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": "0/6000088" logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "prior": null, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "reference": null, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "timestamp": { logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "start": 1773890830, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "stop": 1773890924 logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }, logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | "type": "full" logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | }' logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | ++ jq '.annotation."percona.com/backup-job-name"' --raw-output logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | + backup_job_annotation=demand-backup-backup-dn8b logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | + [[ demand-backup-backup-dn8b != \d\e\m\a\n\d\-\b\a\c\k\u\p\-\b\a\c\k\u\p\-\d\n\8\b ]] logger.go:42: 03:32:58 | demand-backup/5-check-pgbackrest-info-repo1 | test step completed 5-check-pgbackrest-info-repo1 logger.go:42: 03:32:58 | demand-backup/6-create-backup-repo3 | starting test step 6-create-backup-repo3 logger.go:42: 03:32:59 | demand-backup/6-create-backup-repo3 | PerconaPGBackup:kuttl-test-growing-killdeer/demand-backup-full-repo3 created logger.go:42: 03:34:27 | demand-backup/6-create-backup-repo3 | test step completed 6-create-backup-repo3 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | starting test step 7-check-pgbackrest-info-repo3 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | running command: [sh -c set -o errexit set -o xtrace source ../../functions instance=$(kubectl get -n "${NAMESPACE}" pod -l postgres-operator.crunchydata.com/instance-set=instance1 -o 'jsonpath={.items[].metadata.name}') pgbackrest_info_backups=$(kubectl exec -n "${NAMESPACE}" "$instance" -c database -- pgbackrest info --output json --log-level-console=info | jq '.[0].backup[]') check_backup() { local backup_name=$1 local pgbackrest_annotation=$2 local pgbackrest_annotation_value=$3 status_backup_name=$(kubectl get -n "${NAMESPACE}" pg-backup "$backup_name" -o jsonpath='{.status.backupName}') if [[ -z $status_backup_name ]]; then echo ".status.backupName is empty in $backup_name" exit 1 fi backup_info=$(echo "$pgbackrest_info_backups" | jq "select(.annotation.\"$pgbackrest_annotation\" == \"$pgbackrest_annotation_value\")") if [[ $status_backup_name != $(echo "$backup_info" | jq ".label" --raw-output) ]]; then echo ".status.backupName doesn't equal to label in pgbackrest info" exit 1 fi backup_job_name=$(kubectl get -n "${NAMESPACE}" pg-backup "$backup_name" -o jsonpath='{.status.jobName}') backup_job_annotation=$(echo "$backup_info" | jq '.annotation."percona.com/backup-job-name"' --raw-output) if [[ $backup_job_name != "$backup_job_annotation" ]]; then echo "Failed to get job name annotation from pgbackrest" exit 1 fi } manual_backup_name="demand-backup-full-repo3" check_backup "$manual_backup_name" "percona.com/backup-name" "$manual_backup_name" replica_backup_name=$(kubectl get -n "${NAMESPACE}" pg-backup -o jsonpath='{.items[?(@.metadata.annotations.pgv2\.percona\.com/pgbackrest-backup-job-type=="replica-create")].metadata.name}') check_backup "$replica_backup_name" "percona.com/backup-job-type" "replica-create"] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | + source ../../functions logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ realpath ../../.. logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | ++ CERT_MANAGER_VER=1.19.1 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | ++++ pwd logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | ++ test_name=demand-backup logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/vars.sh logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/conf logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/conf logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export TEST_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ TEST_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export TEMP_DIR=/tmp/kuttl/pg/demand-backup logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ TEMP_DIR=/tmp/kuttl/pg/demand-backup logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export GIT_BRANCH=PR-1506 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ GIT_BRANCH=PR-1506 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export VERSION=PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ VERSION=PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export SKIP_TEST_WARNINGS=true logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ SKIP_TEST_WARNINGS=true logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ command -v oc logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ ! -n '' ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ -n '' ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export PG_VER=18 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ PG_VER=18 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export BUCKET=pg-operator-testing logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ BUCKET=pg-operator-testing logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export IMAGE_PMM3_CLIENT=perconalab/pmm-client:3.4 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ IMAGE_PMM3_CLIENT=perconalab/pmm-client:3.4 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export IMAGE_PMM3_SERVER=perconalab/pmm-server:3.4 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ IMAGE_PMM3_SERVER=perconalab/pmm-server:3.4 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export PGOV1_TAG=1.4.0 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ PGOV1_TAG=1.4.0 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export PGOV1_VER=14 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ PGOV1_VER=14 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export MINIO_VER=5.4.0 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ MINIO_VER=5.4.0 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export VAULT_VER=0.32.0 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ VAULT_VER=0.32.0 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export REGISTRY_NAME=docker.io logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ REGISTRY_NAME=docker.io logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export REGISTRY_NAME_FULL=docker.io/ logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ REGISTRY_NAME_FULL=docker.io/ logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | ++++ printenv logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | ++++ grep -E '^IMAGE' logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | ++++ awk -F= '{print $1}' logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | ++++ eval 'echo $IMAGE_BACKREST' logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++++ echo perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ var_value=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == docker.io/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == percona/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == perconalab/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ echo IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | ++++ eval 'echo $IMAGE_PMM3_CLIENT' logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++++ echo perconalab/pmm-client:3.4 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ var_value=perconalab/pmm-client:3.4 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/pmm-client:3.4 == docker.io/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/pmm-client:3.4 == percona/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/pmm-client:3.4 == perconalab/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ new_value=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ echo IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | ++++ eval 'echo $IMAGE' logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ var_value=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == docker.io/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == percona/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == perconalab/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ new_value=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ echo IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | ++++ eval 'echo $IMAGE_UPGRADE' logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++++ echo perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ var_value=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == docker.io/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == percona/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == perconalab/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ echo IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | ++++ eval 'echo $IMAGE_BASE' logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++++ echo perconalab/percona-postgresql-operator logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ var_value=perconalab/percona-postgresql-operator logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/percona-postgresql-operator == docker.io/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/percona-postgresql-operator == percona/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/percona-postgresql-operator == perconalab/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ new_value=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ echo IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | ++++ eval 'echo $IMAGE_PMM3_SERVER' logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++++ echo perconalab/pmm-server:3.4 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ var_value=perconalab/pmm-server:3.4 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/pmm-server:3.4 == docker.io/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/pmm-server:3.4 == percona/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/pmm-server:3.4 == perconalab/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ new_value=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ echo IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | ++++ eval 'echo $IMAGE_PGBOUNCER' logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++++ echo perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ var_value=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == docker.io/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == percona/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == perconalab/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ echo IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | ++++ eval 'echo $IMAGE_POSTGRESQL' logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++++ echo perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ var_value=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == docker.io/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == percona/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == perconalab/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ echo IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | ++++ eval 'echo $IMAGE_PMM_SERVER' logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++++ echo perconalab/pmm-server:dev-latest logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ var_value=perconalab/pmm-server:dev-latest logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/pmm-server:dev-latest == docker.io/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/pmm-server:dev-latest == percona/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/pmm-server:dev-latest == perconalab/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ new_value=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ echo IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | ++++ eval 'echo $IMAGE_PMM_CLIENT' logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++++ echo perconalab/pmm-client:dev-latest logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ var_value=perconalab/pmm-client:dev-latest logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/pmm-client:dev-latest == docker.io/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/pmm-client:dev-latest == percona/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ [[ perconalab/pmm-client:dev-latest == perconalab/* ]] logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ new_value=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ export IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ echo IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:27 | demand-backup/7-check-pgbackrest-info-repo3 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:34:28 | demand-backup/7-check-pgbackrest-info-repo3 | ++++ which gdate logger.go:42: 03:34:28 | demand-backup/7-check-pgbackrest-info-repo3 | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 03:34:28 | demand-backup/7-check-pgbackrest-info-repo3 | ++++ which date logger.go:42: 03:34:28 | demand-backup/7-check-pgbackrest-info-repo3 | +++ date=/usr/sbin/date logger.go:42: 03:34:28 | demand-backup/7-check-pgbackrest-info-repo3 | ++++ which gsed logger.go:42: 03:34:28 | demand-backup/7-check-pgbackrest-info-repo3 | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 03:34:28 | demand-backup/7-check-pgbackrest-info-repo3 | ++++ which sed logger.go:42: 03:34:28 | demand-backup/7-check-pgbackrest-info-repo3 | +++ sed=/usr/sbin/sed logger.go:42: 03:34:28 | demand-backup/7-check-pgbackrest-info-repo3 | ++ oc get projects logger.go:42: 03:34:28 | demand-backup/7-check-pgbackrest-info-repo3 | ++ kubectl get -n kuttl-test-growing-killdeer pod -l postgres-operator.crunchydata.com/instance-set=instance1 -o 'jsonpath={.items[].metadata.name}' logger.go:42: 03:34:28 | demand-backup/7-check-pgbackrest-info-repo3 | + instance=demand-backup-instance1-5kbg-0 logger.go:42: 03:34:28 | demand-backup/7-check-pgbackrest-info-repo3 | ++ kubectl exec -n kuttl-test-growing-killdeer demand-backup-instance1-5kbg-0 -c database -- pgbackrest info --output json --log-level-console=info logger.go:42: 03:34:28 | demand-backup/7-check-pgbackrest-info-repo3 | ++ jq '.[0].backup[]' logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | + pgbackrest_info_backups='{ logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "annotation": { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-job-name": "demand-backup-backup-dn8b", logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-job-type": "replica-create" logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "archive": { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "000000010000000000000005", logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "000000010000000000000006" logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "backrest": { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "format": 5, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "version": "2.58.0" logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "database": { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "id": 1, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "repo-key": 1 logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "error": false, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "info": { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 24276563, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "repository": { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 3133041, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 3133041 logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 24276563 logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "label": "20260319-032710F", logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "lsn": { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "0/5000028", logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "0/6000088" logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "prior": null, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "reference": null, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "timestamp": { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "start": 1773890830, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": 1773890924 logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "type": "full" logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | } logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "annotation": { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-job-name": "demand-backup-backup-877f", logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-name": "demand-backup-full-repo1" logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "archive": { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "000000010000000000000008", logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "00000001000000000000000A" logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "backrest": { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "format": 5, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "version": "2.58.0" logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "database": { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "id": 1, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "repo-key": 1 logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "error": false, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "info": { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 32411538, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "repository": { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 4214354, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 4214354 logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 32411538 logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "label": "20260319-032906F", logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "lsn": { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "0/8000028", logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "0/A000050" logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "prior": null, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "reference": null, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "timestamp": { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "start": 1773890946, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": 1773891152 logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "type": "full" logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | } logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "annotation": { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-job-name": "demand-backup-backup-7z25", logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-name": "demand-backup-full-repo3" logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "archive": { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "00000001000000000000000C", logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "00000001000000000000000C" logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "backrest": { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "format": 5, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "version": "2.58.0" logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "database": { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "id": 1, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "repo-key": 3 logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "error": false, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "info": { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 32480856, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "repository": { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 4216184, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 4216184 logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 32480856 logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "label": "20260319-033307F", logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "lsn": { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "0/C000028", logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "0/C000158" logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "prior": null, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "reference": null, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "timestamp": { logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "start": 1773891187, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": 1773891242 logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | "type": "full" logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | }' logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | + manual_backup_name=demand-backup-full-repo3 logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | + check_backup demand-backup-full-repo3 percona.com/backup-name demand-backup-full-repo3 logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | + local backup_name=demand-backup-full-repo3 logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | + local pgbackrest_annotation=percona.com/backup-name logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | + local pgbackrest_annotation_value=demand-backup-full-repo3 logger.go:42: 03:34:30 | demand-backup/7-check-pgbackrest-info-repo3 | ++ kubectl get -n kuttl-test-growing-killdeer pg-backup demand-backup-full-repo3 -o 'jsonpath={.status.backupName}' logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | + status_backup_name=20260319-033307F logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | + [[ -z 20260319-033307F ]] logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | ++ echo '{ logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "annotation": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-job-name": "demand-backup-backup-dn8b", logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-job-type": "replica-create" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "archive": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "000000010000000000000005", logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "000000010000000000000006" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "backrest": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "format": 5, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "version": "2.58.0" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "database": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "id": 1, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "repo-key": 1 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "error": false, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "info": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 24276563, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "repository": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 3133041, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 3133041 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 24276563 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "label": "20260319-032710F", logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "lsn": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "0/5000028", logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "0/6000088" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "prior": null, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "reference": null, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "timestamp": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "start": 1773890830, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": 1773890924 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "type": "full" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | } logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "annotation": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-job-name": "demand-backup-backup-877f", logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-name": "demand-backup-full-repo1" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "archive": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "000000010000000000000008", logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "00000001000000000000000A" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "backrest": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "format": 5, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "version": "2.58.0" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "database": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "id": 1, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "repo-key": 1 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "error": false, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "info": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 32411538, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "repository": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 4214354, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 4214354 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 32411538 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "label": "20260319-032906F", logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "lsn": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "0/8000028", logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "0/A000050" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "prior": null, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "reference": null, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "timestamp": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "start": 1773890946, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": 1773891152 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "type": "full" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | } logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "annotation": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-job-name": "demand-backup-backup-7z25", logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-name": "demand-backup-full-repo3" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "archive": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "00000001000000000000000C", logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "00000001000000000000000C" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "backrest": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "format": 5, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "version": "2.58.0" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "database": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "id": 1, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "repo-key": 3 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "error": false, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "info": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 32480856, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "repository": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 4216184, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 4216184 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 32480856 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "label": "20260319-033307F", logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "lsn": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "0/C000028", logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "0/C000158" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "prior": null, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "reference": null, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "timestamp": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "start": 1773891187, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": 1773891242 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "type": "full" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }' logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | ++ jq 'select(.annotation."percona.com/backup-name" == "demand-backup-full-repo3")' logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | + backup_info='{ logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "annotation": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-job-name": "demand-backup-backup-7z25", logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-name": "demand-backup-full-repo3" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "archive": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "00000001000000000000000C", logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "00000001000000000000000C" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "backrest": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "format": 5, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "version": "2.58.0" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "database": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "id": 1, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "repo-key": 3 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "error": false, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "info": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 32480856, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "repository": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 4216184, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 4216184 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 32480856 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "label": "20260319-033307F", logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "lsn": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "0/C000028", logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "0/C000158" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "prior": null, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "reference": null, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "timestamp": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "start": 1773891187, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": 1773891242 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "type": "full" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }' logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | ++ echo '{ logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "annotation": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-job-name": "demand-backup-backup-7z25", logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-name": "demand-backup-full-repo3" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "archive": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "00000001000000000000000C", logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "00000001000000000000000C" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "backrest": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "format": 5, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "version": "2.58.0" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "database": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "id": 1, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "repo-key": 3 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "error": false, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "info": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 32480856, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "repository": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 4216184, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 4216184 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 32480856 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "label": "20260319-033307F", logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "lsn": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "0/C000028", logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "0/C000158" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "prior": null, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "reference": null, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "timestamp": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "start": 1773891187, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": 1773891242 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "type": "full" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }' logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | ++ jq .label --raw-output logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | + [[ 20260319-033307F != 20260319-033307F ]] logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | ++ kubectl get -n kuttl-test-growing-killdeer pg-backup demand-backup-full-repo3 -o 'jsonpath={.status.jobName}' logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | + backup_job_name=demand-backup-backup-7z25 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | ++ echo '{ logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "annotation": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-job-name": "demand-backup-backup-7z25", logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-name": "demand-backup-full-repo3" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "archive": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "00000001000000000000000C", logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "00000001000000000000000C" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "backrest": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "format": 5, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "version": "2.58.0" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "database": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "id": 1, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "repo-key": 3 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "error": false, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "info": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 32480856, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "repository": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 4216184, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 4216184 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 32480856 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "label": "20260319-033307F", logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "lsn": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "0/C000028", logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "0/C000158" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "prior": null, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "reference": null, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "timestamp": { logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "start": 1773891187, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": 1773891242 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | "type": "full" logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | }' logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | ++ jq '.annotation."percona.com/backup-job-name"' --raw-output logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | + backup_job_annotation=demand-backup-backup-7z25 logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | + [[ demand-backup-backup-7z25 != \d\e\m\a\n\d\-\b\a\c\k\u\p\-\b\a\c\k\u\p\-\7\z\2\5 ]] logger.go:42: 03:34:31 | demand-backup/7-check-pgbackrest-info-repo3 | ++ kubectl get -n kuttl-test-growing-killdeer pg-backup -o 'jsonpath={.items[?(@.metadata.annotations.pgv2\.percona\.com/pgbackrest-backup-job-type=="replica-create")].metadata.name}' logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | + replica_backup_name=demand-backup-backup-dn8b-n2nxr logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | + check_backup demand-backup-backup-dn8b-n2nxr percona.com/backup-job-type replica-create logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | + local backup_name=demand-backup-backup-dn8b-n2nxr logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | + local pgbackrest_annotation=percona.com/backup-job-type logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | + local pgbackrest_annotation_value=replica-create logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | ++ kubectl get -n kuttl-test-growing-killdeer pg-backup demand-backup-backup-dn8b-n2nxr -o 'jsonpath={.status.backupName}' logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | + status_backup_name=20260319-032710F logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | + [[ -z 20260319-032710F ]] logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | ++ echo '{ logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "annotation": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-job-name": "demand-backup-backup-dn8b", logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-job-type": "replica-create" logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "archive": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "000000010000000000000005", logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "000000010000000000000006" logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "backrest": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "format": 5, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "version": "2.58.0" logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "database": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "id": 1, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "repo-key": 1 logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "error": false, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "info": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 24276563, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "repository": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 3133041, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 3133041 logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 24276563 logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "label": "20260319-032710F", logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "lsn": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "0/5000028", logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "0/6000088" logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "prior": null, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "reference": null, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "timestamp": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "start": 1773890830, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": 1773890924 logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "type": "full" logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | } logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "annotation": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-job-name": "demand-backup-backup-877f", logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-name": "demand-backup-full-repo1" logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "archive": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "000000010000000000000008", logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "00000001000000000000000A" logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "backrest": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "format": 5, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "version": "2.58.0" logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "database": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "id": 1, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "repo-key": 1 logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "error": false, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "info": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 32411538, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "repository": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 4214354, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 4214354 logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 32411538 logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "label": "20260319-032906F", logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "lsn": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "0/8000028", logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "0/A000050" logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "prior": null, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "reference": null, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "timestamp": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "start": 1773890946, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": 1773891152 logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "type": "full" logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | } logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "annotation": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-job-name": "demand-backup-backup-7z25", logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-name": "demand-backup-full-repo3" logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "archive": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "00000001000000000000000C", logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "00000001000000000000000C" logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "backrest": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "format": 5, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "version": "2.58.0" logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "database": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "id": 1, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "repo-key": 3 logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "error": false, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "info": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 32480856, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "repository": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 4216184, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 4216184 logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 32480856 logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "label": "20260319-033307F", logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "lsn": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "0/C000028", logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "0/C000158" logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "prior": null, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "reference": null, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "timestamp": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "start": 1773891187, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": 1773891242 logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "type": "full" logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }' logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | ++ jq 'select(.annotation."percona.com/backup-job-type" == "replica-create")' logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | + backup_info='{ logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "annotation": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-job-name": "demand-backup-backup-dn8b", logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-job-type": "replica-create" logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "archive": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "000000010000000000000005", logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "000000010000000000000006" logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "backrest": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "format": 5, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "version": "2.58.0" logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "database": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "id": 1, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "repo-key": 1 logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "error": false, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "info": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 24276563, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "repository": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 3133041, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 3133041 logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 24276563 logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "label": "20260319-032710F", logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "lsn": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "0/5000028", logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "0/6000088" logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "prior": null, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "reference": null, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "timestamp": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "start": 1773890830, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": 1773890924 logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "type": "full" logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }' logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | ++ echo '{ logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "annotation": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-job-name": "demand-backup-backup-dn8b", logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-job-type": "replica-create" logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "archive": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "000000010000000000000005", logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "000000010000000000000006" logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "backrest": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "format": 5, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "version": "2.58.0" logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "database": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "id": 1, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "repo-key": 1 logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "error": false, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "info": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 24276563, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "repository": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 3133041, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 3133041 logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 24276563 logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "label": "20260319-032710F", logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "lsn": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "0/5000028", logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "0/6000088" logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "prior": null, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "reference": null, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "timestamp": { logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "start": 1773890830, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": 1773890924 logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | "type": "full" logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | }' logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | ++ jq .label --raw-output logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | + [[ 20260319-032710F != 20260319-032710F ]] logger.go:42: 03:34:32 | demand-backup/7-check-pgbackrest-info-repo3 | ++ kubectl get -n kuttl-test-growing-killdeer pg-backup demand-backup-backup-dn8b-n2nxr -o 'jsonpath={.status.jobName}' logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | + backup_job_name=demand-backup-backup-dn8b logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | ++ echo '{ logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "annotation": { logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-job-name": "demand-backup-backup-dn8b", logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "percona.com/backup-job-type": "replica-create" logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "archive": { logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "000000010000000000000005", logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "000000010000000000000006" logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "backrest": { logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "format": 5, logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "version": "2.58.0" logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "database": { logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "id": 1, logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "repo-key": 1 logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "error": false, logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "info": { logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 24276563, logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "repository": { logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "delta": 3133041, logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 3133041 logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "size": 24276563 logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "label": "20260319-032710F", logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "lsn": { logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "start": "0/5000028", logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": "0/6000088" logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "prior": null, logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "reference": null, logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "timestamp": { logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "start": 1773890830, logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "stop": 1773890924 logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | }, logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | "type": "full" logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | }' logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | ++ jq '.annotation."percona.com/backup-job-name"' --raw-output logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | + backup_job_annotation=demand-backup-backup-dn8b logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | + [[ demand-backup-backup-dn8b != \d\e\m\a\n\d\-\b\a\c\k\u\p\-\b\a\c\k\u\p\-\d\n\8\b ]] logger.go:42: 03:34:33 | demand-backup/7-check-pgbackrest-info-repo3 | test step completed 7-check-pgbackrest-info-repo3 logger.go:42: 03:34:33 | demand-backup/8-create-second-backup-repo1 | starting test step 8-create-second-backup-repo1 logger.go:42: 03:34:34 | demand-backup/8-create-second-backup-repo1 | PerconaPGBackup:kuttl-test-growing-killdeer/demand-backup-full-repo1-2 created logger.go:42: 03:40:57 | demand-backup/8-create-second-backup-repo1 | test step completed 8-create-second-backup-repo1 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | starting test step 9-check-retention-repo1 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | running command: [sh -c set -o errexit set -o xtrace source ../../functions retention_count=2 backups_count=$(kubectl -n "${NAMESPACE}" get pg-backup -o yaml | yq '.items | map(select(.metadata.name != "demand-backup-full-repo3*")) | length') jobs_count=$(kubectl -n "${NAMESPACE}" get jobs -o yaml | yq '.items | map(select(.metadata.ownerReferences.[].name != "demand-backup-full-repo3*")) | length') if [[ $backups_count != $retention_count ]]; then echo "There are $backups_count backups, but our retention is set to $retention_count" exit 1 fi if [[ $jobs_count != $retention_count ]]; then echo "There are $jobs_count jobs, but our retention is set to $retention_count" exit 1 fi] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | + source ../../functions logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ realpath ../../.. logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++ CERT_MANAGER_VER=1.19.1 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++++ pwd logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++ test_name=demand-backup logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/vars.sh logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/conf logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/conf logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export TEST_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ TEST_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export TEMP_DIR=/tmp/kuttl/pg/demand-backup logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ TEMP_DIR=/tmp/kuttl/pg/demand-backup logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export GIT_BRANCH=PR-1506 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ GIT_BRANCH=PR-1506 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export VERSION=PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ VERSION=PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export SKIP_TEST_WARNINGS=true logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ SKIP_TEST_WARNINGS=true logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ command -v oc logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ ! -n '' ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ -n '' ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export PG_VER=18 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ PG_VER=18 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export BUCKET=pg-operator-testing logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ BUCKET=pg-operator-testing logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export IMAGE_PMM3_CLIENT=perconalab/pmm-client:3.4 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ IMAGE_PMM3_CLIENT=perconalab/pmm-client:3.4 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export IMAGE_PMM3_SERVER=perconalab/pmm-server:3.4 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ IMAGE_PMM3_SERVER=perconalab/pmm-server:3.4 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export PGOV1_TAG=1.4.0 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ PGOV1_TAG=1.4.0 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export PGOV1_VER=14 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ PGOV1_VER=14 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export MINIO_VER=5.4.0 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ MINIO_VER=5.4.0 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export VAULT_VER=0.32.0 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ VAULT_VER=0.32.0 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export REGISTRY_NAME=docker.io logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ REGISTRY_NAME=docker.io logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export REGISTRY_NAME_FULL=docker.io/ logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ REGISTRY_NAME_FULL=docker.io/ logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++++ printenv logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++++ grep -E '^IMAGE' logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++++ awk -F= '{print $1}' logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++++ eval 'echo $IMAGE_BACKREST' logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++++ echo perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ var_value=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == docker.io/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == percona/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == perconalab/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ echo IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++++ eval 'echo $IMAGE_PMM3_CLIENT' logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++++ echo perconalab/pmm-client:3.4 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ var_value=perconalab/pmm-client:3.4 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/pmm-client:3.4 == docker.io/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/pmm-client:3.4 == percona/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/pmm-client:3.4 == perconalab/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ new_value=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ echo IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++++ eval 'echo $IMAGE' logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ var_value=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == docker.io/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == percona/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == perconalab/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ new_value=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ echo IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++++ eval 'echo $IMAGE_UPGRADE' logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++++ echo perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ var_value=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == docker.io/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == percona/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == perconalab/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ echo IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++++ eval 'echo $IMAGE_BASE' logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++++ echo perconalab/percona-postgresql-operator logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ var_value=perconalab/percona-postgresql-operator logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/percona-postgresql-operator == docker.io/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/percona-postgresql-operator == percona/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/percona-postgresql-operator == perconalab/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ new_value=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ echo IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++++ eval 'echo $IMAGE_PMM3_SERVER' logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++++ echo perconalab/pmm-server:3.4 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ var_value=perconalab/pmm-server:3.4 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/pmm-server:3.4 == docker.io/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/pmm-server:3.4 == percona/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/pmm-server:3.4 == perconalab/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ new_value=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ echo IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++++ eval 'echo $IMAGE_PGBOUNCER' logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++++ echo perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ var_value=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == docker.io/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == percona/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == perconalab/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ echo IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++++ eval 'echo $IMAGE_POSTGRESQL' logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++++ echo perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ var_value=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == docker.io/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == percona/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == perconalab/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ echo IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++++ eval 'echo $IMAGE_PMM_SERVER' logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++++ echo perconalab/pmm-server:dev-latest logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ var_value=perconalab/pmm-server:dev-latest logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/pmm-server:dev-latest == docker.io/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/pmm-server:dev-latest == percona/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/pmm-server:dev-latest == perconalab/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ new_value=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ echo IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++++ eval 'echo $IMAGE_PMM_CLIENT' logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++++ echo perconalab/pmm-client:dev-latest logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ var_value=perconalab/pmm-client:dev-latest logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/pmm-client:dev-latest == docker.io/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/pmm-client:dev-latest == percona/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ [[ perconalab/pmm-client:dev-latest == perconalab/* ]] logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ new_value=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ export IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ echo IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++++ which gdate logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++++ which date logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ date=/usr/sbin/date logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++++ which gsed logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++++ which sed logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | +++ sed=/usr/sbin/sed logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++ oc get projects logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | + retention_count=2 logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++ kubectl -n kuttl-test-growing-killdeer get pg-backup -o yaml logger.go:42: 03:40:57 | demand-backup/9-check-retention-repo1 | ++ yq '.items | map(select(.metadata.name != "demand-backup-full-repo3*")) | length' logger.go:42: 03:40:58 | demand-backup/9-check-retention-repo1 | + backups_count=2 logger.go:42: 03:40:58 | demand-backup/9-check-retention-repo1 | ++ kubectl -n kuttl-test-growing-killdeer get jobs -o yaml logger.go:42: 03:40:58 | demand-backup/9-check-retention-repo1 | ++ yq '.items | map(select(.metadata.ownerReferences.[].name != "demand-backup-full-repo3*")) | length' logger.go:42: 03:40:58 | demand-backup/9-check-retention-repo1 | + jobs_count=2 logger.go:42: 03:40:58 | demand-backup/9-check-retention-repo1 | + [[ 2 != 2 ]] logger.go:42: 03:40:58 | demand-backup/9-check-retention-repo1 | + [[ 2 != 2 ]] logger.go:42: 03:40:58 | demand-backup/9-check-retention-repo1 | test step completed 9-check-retention-repo1 logger.go:42: 03:40:58 | demand-backup/10-create-second-backup-repo3 | starting test step 10-create-second-backup-repo3 logger.go:42: 03:40:59 | demand-backup/10-create-second-backup-repo3 | PerconaPGBackup:kuttl-test-growing-killdeer/demand-backup-full-repo3-2 created logger.go:42: 03:42:42 | demand-backup/10-create-second-backup-repo3 | test step completed 10-create-second-backup-repo3 logger.go:42: 03:42:42 | demand-backup/11-create-third-backup-repo3 | starting test step 11-create-third-backup-repo3 logger.go:42: 03:42:43 | demand-backup/11-create-third-backup-repo3 | PerconaPGBackup:kuttl-test-growing-killdeer/demand-backup-full-repo3-3 created logger.go:42: 03:46:42 | demand-backup/11-create-third-backup-repo3 | test step completed 11-create-third-backup-repo3 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | starting test step 12-check-retention-repo3 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | running command: [sh -c set -o errexit set -o xtrace source ../../functions retention_count=2 backups_count=$(kubectl -n "${NAMESPACE}" get pg-backup -o yaml | yq '.items | map(select(.metadata.name == "demand-backup-full-repo3*")) | length') jobs_count=$(kubectl -n "${NAMESPACE}" get jobs -o yaml | yq '.items | map(select(.metadata.ownerReferences.[].name == "demand-backup-full-repo3*")) | length') if [[ $backups_count != $retention_count ]]; then echo "There are $backups_count backups, but our retention is set to $retention_count" exit 1 fi if [[ $jobs_count != $retention_count ]]; then echo "There are $jobs_count jobs, but our retention is set to $retention_count" exit 1 fi] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | + source ../../functions logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ realpath ../../.. logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++ CERT_MANAGER_VER=1.19.1 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++++ pwd logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++ test_name=demand-backup logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/vars.sh logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/conf logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/conf logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export TEST_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ TEST_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export TEMP_DIR=/tmp/kuttl/pg/demand-backup logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ TEMP_DIR=/tmp/kuttl/pg/demand-backup logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export GIT_BRANCH=PR-1506 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ GIT_BRANCH=PR-1506 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export VERSION=PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ VERSION=PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export SKIP_TEST_WARNINGS=true logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ SKIP_TEST_WARNINGS=true logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ command -v oc logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ ! -n '' ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ -n '' ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export PG_VER=18 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ PG_VER=18 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export BUCKET=pg-operator-testing logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ BUCKET=pg-operator-testing logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export IMAGE_PMM3_CLIENT=perconalab/pmm-client:3.4 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ IMAGE_PMM3_CLIENT=perconalab/pmm-client:3.4 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export IMAGE_PMM3_SERVER=perconalab/pmm-server:3.4 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ IMAGE_PMM3_SERVER=perconalab/pmm-server:3.4 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export PGOV1_TAG=1.4.0 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ PGOV1_TAG=1.4.0 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export PGOV1_VER=14 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ PGOV1_VER=14 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export MINIO_VER=5.4.0 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ MINIO_VER=5.4.0 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export VAULT_VER=0.32.0 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ VAULT_VER=0.32.0 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export REGISTRY_NAME=docker.io logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ REGISTRY_NAME=docker.io logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export REGISTRY_NAME_FULL=docker.io/ logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ REGISTRY_NAME_FULL=docker.io/ logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++++ printenv logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++++ grep -E '^IMAGE' logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++++ awk -F= '{print $1}' logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++++ eval 'echo $IMAGE_BACKREST' logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++++ echo perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ var_value=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == docker.io/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == percona/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == perconalab/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ echo IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++++ eval 'echo $IMAGE_PMM3_CLIENT' logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++++ echo perconalab/pmm-client:3.4 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ var_value=perconalab/pmm-client:3.4 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/pmm-client:3.4 == docker.io/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/pmm-client:3.4 == percona/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/pmm-client:3.4 == perconalab/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ new_value=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ echo IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++++ eval 'echo $IMAGE' logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ var_value=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == docker.io/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == percona/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == perconalab/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ new_value=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ echo IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++++ eval 'echo $IMAGE_UPGRADE' logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++++ echo perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ var_value=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == docker.io/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == percona/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == perconalab/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ echo IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++++ eval 'echo $IMAGE_BASE' logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++++ echo perconalab/percona-postgresql-operator logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ var_value=perconalab/percona-postgresql-operator logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/percona-postgresql-operator == docker.io/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/percona-postgresql-operator == percona/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/percona-postgresql-operator == perconalab/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ new_value=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ echo IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++++ eval 'echo $IMAGE_PMM3_SERVER' logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++++ echo perconalab/pmm-server:3.4 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ var_value=perconalab/pmm-server:3.4 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/pmm-server:3.4 == docker.io/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/pmm-server:3.4 == percona/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/pmm-server:3.4 == perconalab/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ new_value=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ echo IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++++ eval 'echo $IMAGE_PGBOUNCER' logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++++ echo perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ var_value=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == docker.io/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == percona/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == perconalab/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ echo IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++++ eval 'echo $IMAGE_POSTGRESQL' logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++++ echo perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ var_value=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == docker.io/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == percona/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == perconalab/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ echo IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++++ eval 'echo $IMAGE_PMM_SERVER' logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++++ echo perconalab/pmm-server:dev-latest logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ var_value=perconalab/pmm-server:dev-latest logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/pmm-server:dev-latest == docker.io/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/pmm-server:dev-latest == percona/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/pmm-server:dev-latest == perconalab/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ new_value=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ echo IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++++ eval 'echo $IMAGE_PMM_CLIENT' logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++++ echo perconalab/pmm-client:dev-latest logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ var_value=perconalab/pmm-client:dev-latest logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/pmm-client:dev-latest == docker.io/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/pmm-client:dev-latest == percona/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ [[ perconalab/pmm-client:dev-latest == perconalab/* ]] logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ new_value=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ export IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ echo IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++++ which gdate logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++++ which date logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ date=/usr/sbin/date logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++++ which gsed logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++++ which sed logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | +++ sed=/usr/sbin/sed logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++ oc get projects logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | + retention_count=2 logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++ kubectl -n kuttl-test-growing-killdeer get pg-backup -o yaml logger.go:42: 03:46:42 | demand-backup/12-check-retention-repo3 | ++ yq '.items | map(select(.metadata.name == "demand-backup-full-repo3*")) | length' logger.go:42: 03:46:43 | demand-backup/12-check-retention-repo3 | + backups_count=2 logger.go:42: 03:46:43 | demand-backup/12-check-retention-repo3 | ++ kubectl -n kuttl-test-growing-killdeer get jobs -o yaml logger.go:42: 03:46:43 | demand-backup/12-check-retention-repo3 | ++ yq '.items | map(select(.metadata.ownerReferences.[].name == "demand-backup-full-repo3*")) | length' logger.go:42: 03:46:43 | demand-backup/12-check-retention-repo3 | + jobs_count=2 logger.go:42: 03:46:43 | demand-backup/12-check-retention-repo3 | + [[ 2 != 2 ]] logger.go:42: 03:46:43 | demand-backup/12-check-retention-repo3 | + [[ 2 != 2 ]] logger.go:42: 03:46:43 | demand-backup/12-check-retention-repo3 | test step completed 12-check-retention-repo3 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | starting test step 13-delete-backup logger.go:42: 03:46:43 | demand-backup/13-delete-backup | running command: [sh -c set -o errexit set -o xtrace source ../../functions kubectl delete pg-backup -n "${NAMESPACE}" demand-backup-full-repo1 kubectl delete pg-backup -n "${NAMESPACE}" demand-backup-full-repo3-2] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | + source ../../functions logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ realpath ../../.. logger.go:42: 03:46:43 | demand-backup/13-delete-backup | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | ++ CERT_MANAGER_VER=1.19.1 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | ++++ pwd logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup logger.go:42: 03:46:43 | demand-backup/13-delete-backup | ++ test_name=demand-backup logger.go:42: 03:46:43 | demand-backup/13-delete-backup | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/vars.sh logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/conf logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/conf logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export TEST_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ TEST_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export TEMP_DIR=/tmp/kuttl/pg/demand-backup logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ TEMP_DIR=/tmp/kuttl/pg/demand-backup logger.go:42: 03:46:43 | demand-backup/13-delete-backup | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export GIT_BRANCH=PR-1506 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ GIT_BRANCH=PR-1506 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export VERSION=PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ VERSION=PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export SKIP_TEST_WARNINGS=true logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ SKIP_TEST_WARNINGS=true logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ command -v oc logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ ! -n '' ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ -n '' ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export PG_VER=18 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ PG_VER=18 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export BUCKET=pg-operator-testing logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ BUCKET=pg-operator-testing logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export IMAGE_PMM3_CLIENT=perconalab/pmm-client:3.4 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ IMAGE_PMM3_CLIENT=perconalab/pmm-client:3.4 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export IMAGE_PMM3_SERVER=perconalab/pmm-server:3.4 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ IMAGE_PMM3_SERVER=perconalab/pmm-server:3.4 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export PGOV1_TAG=1.4.0 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ PGOV1_TAG=1.4.0 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export PGOV1_VER=14 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ PGOV1_VER=14 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export MINIO_VER=5.4.0 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ MINIO_VER=5.4.0 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export VAULT_VER=0.32.0 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ VAULT_VER=0.32.0 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export REGISTRY_NAME=docker.io logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ REGISTRY_NAME=docker.io logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export REGISTRY_NAME_FULL=docker.io/ logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ REGISTRY_NAME_FULL=docker.io/ logger.go:42: 03:46:43 | demand-backup/13-delete-backup | ++++ printenv logger.go:42: 03:46:43 | demand-backup/13-delete-backup | ++++ grep -E '^IMAGE' logger.go:42: 03:46:43 | demand-backup/13-delete-backup | ++++ awk -F= '{print $1}' logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:46:43 | demand-backup/13-delete-backup | ++++ eval 'echo $IMAGE_BACKREST' logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++++ echo perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ var_value=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == docker.io/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == percona/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == perconalab/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ echo IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:46:43 | demand-backup/13-delete-backup | ++++ eval 'echo $IMAGE_PMM3_CLIENT' logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++++ echo perconalab/pmm-client:3.4 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ var_value=perconalab/pmm-client:3.4 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/pmm-client:3.4 == docker.io/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/pmm-client:3.4 == percona/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/pmm-client:3.4 == perconalab/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ new_value=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ echo IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:46:43 | demand-backup/13-delete-backup | ++++ eval 'echo $IMAGE' logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ var_value=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == docker.io/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == percona/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == perconalab/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ new_value=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ echo IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:46:43 | demand-backup/13-delete-backup | ++++ eval 'echo $IMAGE_UPGRADE' logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++++ echo perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ var_value=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == docker.io/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == percona/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == perconalab/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ echo IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:46:43 | demand-backup/13-delete-backup | IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:46:43 | demand-backup/13-delete-backup | ++++ eval 'echo $IMAGE_BASE' logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++++ echo perconalab/percona-postgresql-operator logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ var_value=perconalab/percona-postgresql-operator logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/percona-postgresql-operator == docker.io/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/percona-postgresql-operator == percona/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/percona-postgresql-operator == perconalab/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ new_value=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ echo IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:46:43 | demand-backup/13-delete-backup | IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:46:43 | demand-backup/13-delete-backup | ++++ eval 'echo $IMAGE_PMM3_SERVER' logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++++ echo perconalab/pmm-server:3.4 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ var_value=perconalab/pmm-server:3.4 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/pmm-server:3.4 == docker.io/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/pmm-server:3.4 == percona/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/pmm-server:3.4 == perconalab/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ new_value=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ echo IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:46:43 | demand-backup/13-delete-backup | ++++ eval 'echo $IMAGE_PGBOUNCER' logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++++ echo perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ var_value=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == docker.io/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == percona/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == perconalab/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ echo IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:46:43 | demand-backup/13-delete-backup | ++++ eval 'echo $IMAGE_POSTGRESQL' logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++++ echo perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ var_value=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == docker.io/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == percona/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == perconalab/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ echo IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:46:43 | demand-backup/13-delete-backup | IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:46:43 | demand-backup/13-delete-backup | ++++ eval 'echo $IMAGE_PMM_SERVER' logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++++ echo perconalab/pmm-server:dev-latest logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ var_value=perconalab/pmm-server:dev-latest logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/pmm-server:dev-latest == docker.io/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/pmm-server:dev-latest == percona/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/pmm-server:dev-latest == perconalab/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ new_value=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ echo IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:46:43 | demand-backup/13-delete-backup | IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:46:43 | demand-backup/13-delete-backup | ++++ eval 'echo $IMAGE_PMM_CLIENT' logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++++ echo perconalab/pmm-client:dev-latest logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ var_value=perconalab/pmm-client:dev-latest logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/pmm-client:dev-latest == docker.io/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/pmm-client:dev-latest == percona/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ [[ perconalab/pmm-client:dev-latest == perconalab/* ]] logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ new_value=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ export IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ echo IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:46:43 | demand-backup/13-delete-backup | IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:46:43 | demand-backup/13-delete-backup | ++++ which gdate logger.go:42: 03:46:43 | demand-backup/13-delete-backup | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 03:46:43 | demand-backup/13-delete-backup | ++++ which date logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ date=/usr/sbin/date logger.go:42: 03:46:43 | demand-backup/13-delete-backup | ++++ which gsed logger.go:42: 03:46:43 | demand-backup/13-delete-backup | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 03:46:43 | demand-backup/13-delete-backup | ++++ which sed logger.go:42: 03:46:43 | demand-backup/13-delete-backup | +++ sed=/usr/sbin/sed logger.go:42: 03:46:43 | demand-backup/13-delete-backup | ++ oc get projects logger.go:42: 03:46:43 | demand-backup/13-delete-backup | + kubectl delete pg-backup -n kuttl-test-growing-killdeer demand-backup-full-repo1 logger.go:42: 03:46:44 | demand-backup/13-delete-backup | perconapgbackup.pgv2.percona.com "demand-backup-full-repo1" deleted from kuttl-test-growing-killdeer namespace logger.go:42: 03:46:44 | demand-backup/13-delete-backup | + kubectl delete pg-backup -n kuttl-test-growing-killdeer demand-backup-full-repo3-2 logger.go:42: 03:46:45 | demand-backup/13-delete-backup | perconapgbackup.pgv2.percona.com "demand-backup-full-repo3-2" deleted from kuttl-test-growing-killdeer namespace logger.go:42: 03:46:45 | demand-backup/13-delete-backup | test step completed 13-delete-backup logger.go:42: 03:46:45 | demand-backup/14-recreate-backup-repo1 | starting test step 14-recreate-backup-repo1 logger.go:42: 03:46:45 | demand-backup/14-recreate-backup-repo1 | PerconaPGBackup:kuttl-test-growing-killdeer/demand-backup-full-repo1 created logger.go:42: 03:49:13 | demand-backup/14-recreate-backup-repo1 | test step completed 14-recreate-backup-repo1 logger.go:42: 03:49:13 | demand-backup/15-recreate-backup-repo3 | starting test step 15-recreate-backup-repo3 logger.go:42: 03:49:14 | demand-backup/15-recreate-backup-repo3 | PerconaPGBackup:kuttl-test-growing-killdeer/demand-backup-full-repo3 created logger.go:42: 03:51:44 | demand-backup/15-recreate-backup-repo3 | test step completed 15-recreate-backup-repo3 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | starting test step 16-check-password-leak logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | running command: [sh -c set -o errexit set -o xtrace source ../../functions # Temporarily skipping this check # check_passwords_leak] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | + source ../../functions logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ realpath ../../.. logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | ++ CERT_MANAGER_VER=1.19.1 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | ++++ pwd logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | ++ test_name=demand-backup logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/vars.sh logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/conf logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/conf logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export TEST_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ TEST_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export TEMP_DIR=/tmp/kuttl/pg/demand-backup logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ TEMP_DIR=/tmp/kuttl/pg/demand-backup logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export GIT_BRANCH=PR-1506 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ GIT_BRANCH=PR-1506 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export VERSION=PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ VERSION=PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export SKIP_TEST_WARNINGS=true logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ SKIP_TEST_WARNINGS=true logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ command -v oc logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ ! -n '' ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ -n '' ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export PG_VER=18 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ PG_VER=18 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export BUCKET=pg-operator-testing logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ BUCKET=pg-operator-testing logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export IMAGE_PMM3_CLIENT=perconalab/pmm-client:3.4 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ IMAGE_PMM3_CLIENT=perconalab/pmm-client:3.4 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export IMAGE_PMM3_SERVER=perconalab/pmm-server:3.4 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ IMAGE_PMM3_SERVER=perconalab/pmm-server:3.4 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export PGOV1_TAG=1.4.0 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ PGOV1_TAG=1.4.0 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export PGOV1_VER=14 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ PGOV1_VER=14 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export MINIO_VER=5.4.0 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ MINIO_VER=5.4.0 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export VAULT_VER=0.32.0 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ VAULT_VER=0.32.0 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export REGISTRY_NAME=docker.io logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ REGISTRY_NAME=docker.io logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export REGISTRY_NAME_FULL=docker.io/ logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ REGISTRY_NAME_FULL=docker.io/ logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | ++++ printenv logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | ++++ grep -E '^IMAGE' logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | ++++ awk -F= '{print $1}' logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | ++++ eval 'echo $IMAGE_BACKREST' logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++++ echo perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ var_value=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == docker.io/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == percona/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == perconalab/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ echo IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | ++++ eval 'echo $IMAGE_PMM3_CLIENT' logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++++ echo perconalab/pmm-client:3.4 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ var_value=perconalab/pmm-client:3.4 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/pmm-client:3.4 == docker.io/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/pmm-client:3.4 == percona/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/pmm-client:3.4 == perconalab/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ new_value=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ echo IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | ++++ eval 'echo $IMAGE' logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ var_value=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == docker.io/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == percona/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == perconalab/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ new_value=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ echo IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | ++++ eval 'echo $IMAGE_UPGRADE' logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++++ echo perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ var_value=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == docker.io/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == percona/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == perconalab/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ echo IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | ++++ eval 'echo $IMAGE_BASE' logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++++ echo perconalab/percona-postgresql-operator logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ var_value=perconalab/percona-postgresql-operator logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/percona-postgresql-operator == docker.io/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/percona-postgresql-operator == percona/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/percona-postgresql-operator == perconalab/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ new_value=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ echo IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | ++++ eval 'echo $IMAGE_PMM3_SERVER' logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++++ echo perconalab/pmm-server:3.4 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ var_value=perconalab/pmm-server:3.4 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/pmm-server:3.4 == docker.io/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/pmm-server:3.4 == percona/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/pmm-server:3.4 == perconalab/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ new_value=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ echo IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | ++++ eval 'echo $IMAGE_PGBOUNCER' logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++++ echo perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ var_value=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == docker.io/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == percona/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == perconalab/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ echo IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | ++++ eval 'echo $IMAGE_POSTGRESQL' logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++++ echo perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ var_value=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == docker.io/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == percona/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == perconalab/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ echo IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | ++++ eval 'echo $IMAGE_PMM_SERVER' logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++++ echo perconalab/pmm-server:dev-latest logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ var_value=perconalab/pmm-server:dev-latest logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/pmm-server:dev-latest == docker.io/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/pmm-server:dev-latest == percona/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/pmm-server:dev-latest == perconalab/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ new_value=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ echo IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | ++++ eval 'echo $IMAGE_PMM_CLIENT' logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++++ echo perconalab/pmm-client:dev-latest logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ var_value=perconalab/pmm-client:dev-latest logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/pmm-client:dev-latest == docker.io/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/pmm-client:dev-latest == percona/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ [[ perconalab/pmm-client:dev-latest == perconalab/* ]] logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ new_value=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ export IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ echo IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | ++++ which gdate logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | ++++ which date logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ date=/usr/sbin/date logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | ++++ which gsed logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | ++++ which sed logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | +++ sed=/usr/sbin/sed logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | ++ oc get projects logger.go:42: 03:51:44 | demand-backup/16-check-password-leak | test step completed 16-check-password-leak logger.go:42: 03:51:44 | demand-backup/17-delete-data | starting test step 17-delete-data logger.go:42: 03:51:44 | demand-backup/17-delete-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_psql_local \ '\c myapp \\\ TRUNCATE TABLE myApp' \ "postgres:$(get_psql_user_pass demand-backup-pguser-postgres)@$(get_psql_user_host demand-backup-pguser-postgres)"] logger.go:42: 03:51:44 | demand-backup/17-delete-data | + source ../../functions logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ realpath ../../.. logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++ CERT_MANAGER_VER=1.19.1 logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++++ pwd logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++ test_name=demand-backup logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/vars.sh logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/conf logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/conf logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export TEST_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ TEST_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export TEMP_DIR=/tmp/kuttl/pg/demand-backup logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ TEMP_DIR=/tmp/kuttl/pg/demand-backup logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export GIT_BRANCH=PR-1506 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ GIT_BRANCH=PR-1506 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export VERSION=PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ VERSION=PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export SKIP_TEST_WARNINGS=true logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ SKIP_TEST_WARNINGS=true logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ command -v oc logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ ! -n '' ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ -n '' ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export PG_VER=18 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ PG_VER=18 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export BUCKET=pg-operator-testing logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ BUCKET=pg-operator-testing logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export IMAGE_PMM3_CLIENT=perconalab/pmm-client:3.4 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ IMAGE_PMM3_CLIENT=perconalab/pmm-client:3.4 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export IMAGE_PMM3_SERVER=perconalab/pmm-server:3.4 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ IMAGE_PMM3_SERVER=perconalab/pmm-server:3.4 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export PGOV1_TAG=1.4.0 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ PGOV1_TAG=1.4.0 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export PGOV1_VER=14 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ PGOV1_VER=14 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export MINIO_VER=5.4.0 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ MINIO_VER=5.4.0 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export VAULT_VER=0.32.0 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ VAULT_VER=0.32.0 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export REGISTRY_NAME=docker.io logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ REGISTRY_NAME=docker.io logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export REGISTRY_NAME_FULL=docker.io/ logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ REGISTRY_NAME_FULL=docker.io/ logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++++ printenv logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++++ grep -E '^IMAGE' logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++++ awk -F= '{print $1}' logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++++ eval 'echo $IMAGE_BACKREST' logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++++ echo perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ var_value=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == docker.io/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == percona/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == perconalab/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ echo IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:51:44 | demand-backup/17-delete-data | IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++++ eval 'echo $IMAGE_PMM3_CLIENT' logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++++ echo perconalab/pmm-client:3.4 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ var_value=perconalab/pmm-client:3.4 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/pmm-client:3.4 == docker.io/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/pmm-client:3.4 == percona/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/pmm-client:3.4 == perconalab/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ new_value=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ echo IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:51:44 | demand-backup/17-delete-data | IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++++ eval 'echo $IMAGE' logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ var_value=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == docker.io/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == percona/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == perconalab/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ new_value=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ echo IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++++ eval 'echo $IMAGE_UPGRADE' logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++++ echo perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ var_value=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == docker.io/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == percona/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == perconalab/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ echo IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:51:44 | demand-backup/17-delete-data | IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++++ eval 'echo $IMAGE_BASE' logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++++ echo perconalab/percona-postgresql-operator logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ var_value=perconalab/percona-postgresql-operator logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/percona-postgresql-operator == docker.io/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/percona-postgresql-operator == percona/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/percona-postgresql-operator == perconalab/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ new_value=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ echo IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:51:44 | demand-backup/17-delete-data | IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++++ eval 'echo $IMAGE_PMM3_SERVER' logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++++ echo perconalab/pmm-server:3.4 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ var_value=perconalab/pmm-server:3.4 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/pmm-server:3.4 == docker.io/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/pmm-server:3.4 == percona/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/pmm-server:3.4 == perconalab/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ new_value=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ echo IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:51:44 | demand-backup/17-delete-data | IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++++ eval 'echo $IMAGE_PGBOUNCER' logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++++ echo perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ var_value=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == docker.io/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == percona/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == perconalab/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ echo IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:51:44 | demand-backup/17-delete-data | IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++++ eval 'echo $IMAGE_POSTGRESQL' logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++++ echo perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ var_value=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == docker.io/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == percona/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == perconalab/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ echo IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:51:44 | demand-backup/17-delete-data | IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++++ eval 'echo $IMAGE_PMM_SERVER' logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++++ echo perconalab/pmm-server:dev-latest logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ var_value=perconalab/pmm-server:dev-latest logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/pmm-server:dev-latest == docker.io/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/pmm-server:dev-latest == percona/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/pmm-server:dev-latest == perconalab/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ new_value=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ echo IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:51:44 | demand-backup/17-delete-data | IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++++ eval 'echo $IMAGE_PMM_CLIENT' logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++++ echo perconalab/pmm-client:dev-latest logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ var_value=perconalab/pmm-client:dev-latest logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/pmm-client:dev-latest == docker.io/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/pmm-client:dev-latest == percona/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ [[ perconalab/pmm-client:dev-latest == perconalab/* ]] logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ new_value=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ export IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ echo IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:51:44 | demand-backup/17-delete-data | IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++++ which gdate logger.go:42: 03:51:44 | demand-backup/17-delete-data | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++++ which date logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ date=/usr/sbin/date logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++++ which gsed logger.go:42: 03:51:44 | demand-backup/17-delete-data | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++++ which sed logger.go:42: 03:51:44 | demand-backup/17-delete-data | +++ sed=/usr/sbin/sed logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++ oc get projects logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++ get_psql_user_pass demand-backup-pguser-postgres logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++ local secret_name=demand-backup-pguser-postgres logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++ kubectl -n kuttl-test-growing-killdeer get secret/demand-backup-pguser-postgres '--template={{.data.password | base64decode}}' logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++ get_psql_user_host demand-backup-pguser-postgres logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++ local secret_name=demand-backup-pguser-postgres logger.go:42: 03:51:44 | demand-backup/17-delete-data | ++ kubectl -n kuttl-test-growing-killdeer get secret/demand-backup-pguser-postgres '--template={{.data.host | base64decode }}' logger.go:42: 03:51:45 | demand-backup/17-delete-data | + run_psql_local '\c myapp \\\ TRUNCATE TABLE myApp' postgres:N7HeCfaqv9X8trS8vJrCEsWf@demand-backup-primary.kuttl-test-growing-killdeer.svc logger.go:42: 03:51:45 | demand-backup/17-delete-data | + local 'command=\c myapp \\\ TRUNCATE TABLE myApp' logger.go:42: 03:51:45 | demand-backup/17-delete-data | + local uri=postgres:N7HeCfaqv9X8trS8vJrCEsWf@demand-backup-primary.kuttl-test-growing-killdeer.svc logger.go:42: 03:51:45 | demand-backup/17-delete-data | + local driver=postgres logger.go:42: 03:51:45 | demand-backup/17-delete-data | ++ get_client_pod logger.go:42: 03:51:45 | demand-backup/17-delete-data | ++ kubectl -n kuttl-test-growing-killdeer get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 03:51:45 | demand-backup/17-delete-data | + kubectl -n kuttl-test-growing-killdeer exec pg-client-65d98588cc-qfkgp -- bash -c 'printf '\''\c myapp \\\ TRUNCATE TABLE myApp\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://'\''postgres:N7HeCfaqv9X8trS8vJrCEsWf@demand-backup-primary.kuttl-test-growing-killdeer.svc'\''' logger.go:42: 03:51:46 | demand-backup/17-delete-data | test step completed 17-delete-data logger.go:42: 03:51:46 | demand-backup/18-create-restore-repo1 | starting test step 18-create-restore-repo1 logger.go:42: 03:51:46 | demand-backup/18-create-restore-repo1 | running command: [sh -c set -o errexit set -o xtrace source ../../functions primary=$(get_pod_by_role demand-backup primary name) latest_full_repo1_backup=$(kubectl -n ${NAMESPACE} exec ${primary} -- pgbackrest info --output json --log-level-console=info | jq '[.[] | .backup[] | select(.type == "full") | select(.database.["repo-key"] == 1)][-1].label') cat <&1 >/dev/null) if [[ $res == *$(echo "No resources found in ${NAMESPACE} namespace.")* ]]; then data=0 fi kubectl create configmap -n "${NAMESPACE}" 25-pg-backup-objects --from-literal=data="${data}"] logger.go:42: 03:55:47 | demand-backup/25-delete-cluster-with-finalizer | + kubectl delete pg -n kuttl-test-growing-killdeer demand-backup logger.go:42: 03:55:47 | demand-backup/25-delete-cluster-with-finalizer | perconapgcluster.pgv2.percona.com "demand-backup" deleted from kuttl-test-growing-killdeer namespace logger.go:42: 03:57:38 | demand-backup/25-delete-cluster-with-finalizer | + sleep 15 logger.go:42: 03:57:53 | demand-backup/25-delete-cluster-with-finalizer | + data=1 logger.go:42: 03:57:53 | demand-backup/25-delete-cluster-with-finalizer | ++ kubectl -n kuttl-test-growing-killdeer get pg-backup logger.go:42: 03:57:55 | demand-backup/25-delete-cluster-with-finalizer | + res='No resources found in kuttl-test-growing-killdeer namespace.' logger.go:42: 03:57:55 | demand-backup/25-delete-cluster-with-finalizer | ++ echo 'No resources found in kuttl-test-growing-killdeer namespace.' logger.go:42: 03:57:55 | demand-backup/25-delete-cluster-with-finalizer | + [[ No resources found in kuttl-test-growing-killdeer namespace. == *No resources found in kuttl-test-growing-killdeer namespace.* ]] logger.go:42: 03:57:55 | demand-backup/25-delete-cluster-with-finalizer | + data=0 logger.go:42: 03:57:55 | demand-backup/25-delete-cluster-with-finalizer | + kubectl create configmap -n kuttl-test-growing-killdeer 25-pg-backup-objects --from-literal=data=0 logger.go:42: 03:57:55 | demand-backup/25-delete-cluster-with-finalizer | configmap/25-pg-backup-objects created logger.go:42: 03:57:56 | demand-backup/25-delete-cluster-with-finalizer | test step completed 25-delete-cluster-with-finalizer logger.go:42: 03:57:56 | demand-backup/99-remove-cluster-gracefully | starting test step 99-remove-cluster-gracefully logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | running command: [sh -c set -o errexit set -o xtrace source ../../functions remove_all_finalizers check_operator_panic destroy_operator] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | + source ../../functions logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ realpath ../../.. logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | ++ CERT_MANAGER_VER=1.19.1 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | ++++ pwd logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | ++ test_name=demand-backup logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/vars.sh logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/deploy logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/conf logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/conf logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export TEST_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ TEST_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/e2e-tests/tests/demand-backup/conf logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export TEMP_DIR=/tmp/kuttl/pg/demand-backup logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ TEMP_DIR=/tmp/kuttl/pg/demand-backup logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export GIT_BRANCH=PR-1506 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ GIT_BRANCH=PR-1506 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export VERSION=PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ VERSION=PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export SKIP_TEST_WARNINGS=true logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ SKIP_TEST_WARNINGS=true logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ command -v oc logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ ! -n '' ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ -n '' ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export PG_VER=18 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ PG_VER=18 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export BUCKET=pg-operator-testing logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ BUCKET=pg-operator-testing logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_PMM3_CLIENT=perconalab/pmm-client:3.4 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ IMAGE_PMM3_CLIENT=perconalab/pmm-client:3.4 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_PMM3_SERVER=perconalab/pmm-server:3.4 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ IMAGE_PMM3_SERVER=perconalab/pmm-server:3.4 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export PGOV1_TAG=1.4.0 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ PGOV1_TAG=1.4.0 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export PGOV1_VER=14 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ PGOV1_VER=14 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export MINIO_VER=5.4.0 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ MINIO_VER=5.4.0 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export VAULT_VER=0.32.0 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ VAULT_VER=0.32.0 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export REGISTRY_NAME=docker.io logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ REGISTRY_NAME=docker.io logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export REGISTRY_NAME_FULL=docker.io/ logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ REGISTRY_NAME_FULL=docker.io/ logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | ++++ printenv logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | ++++ grep -E '^IMAGE' logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | ++++ awk -F= '{print $1}' logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | ++++ eval 'echo $IMAGE_BACKREST' logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++++ echo perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ var_value=perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == docker.io/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == percona/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/percona-postgresql-operator:main-pgbackrest18 == perconalab/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ echo IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | IMAGE_BACKREST=docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | ++++ eval 'echo $IMAGE_PMM3_CLIENT' logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++++ echo perconalab/pmm-client:3.4 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ var_value=perconalab/pmm-client:3.4 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/pmm-client:3.4 == docker.io/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/pmm-client:3.4 == percona/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/pmm-client:3.4 == perconalab/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ new_value=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ echo IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | IMAGE_PMM3_CLIENT=docker.io/perconalab/pmm-client:3.4 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | ++++ eval 'echo $IMAGE' logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++++ echo perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ var_value=perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == docker.io/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == percona/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/percona-postgresql-operator:PR-1506-e8248869c == perconalab/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ new_value=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ echo IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | IMAGE=docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | ++++ eval 'echo $IMAGE_UPGRADE' logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++++ echo perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ var_value=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == docker.io/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == percona/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/percona-postgresql-operator:main-upgrade == perconalab/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ echo IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | IMAGE_UPGRADE=docker.io/perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | ++++ eval 'echo $IMAGE_BASE' logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++++ echo perconalab/percona-postgresql-operator logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ var_value=perconalab/percona-postgresql-operator logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/percona-postgresql-operator == docker.io/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/percona-postgresql-operator == percona/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/percona-postgresql-operator == perconalab/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ new_value=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ echo IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | IMAGE_BASE=docker.io/perconalab/percona-postgresql-operator logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | ++++ eval 'echo $IMAGE_PMM3_SERVER' logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++++ echo perconalab/pmm-server:3.4 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ var_value=perconalab/pmm-server:3.4 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/pmm-server:3.4 == docker.io/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/pmm-server:3.4 == percona/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/pmm-server:3.4 == perconalab/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ new_value=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ echo IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | IMAGE_PMM3_SERVER=docker.io/perconalab/pmm-server:3.4 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | ++++ eval 'echo $IMAGE_PGBOUNCER' logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++++ echo perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ var_value=perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == docker.io/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == percona/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/percona-postgresql-operator:main-pgbouncer18 == perconalab/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ echo IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | IMAGE_PGBOUNCER=docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18 logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | ++++ eval 'echo $IMAGE_POSTGRESQL' logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++++ echo perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ var_value=perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == docker.io/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == percona/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/percona-postgresql-operator:main-ppg18-postgres == perconalab/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ new_value=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ echo IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | IMAGE_POSTGRESQL=docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | ++++ eval 'echo $IMAGE_PMM_SERVER' logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++++ echo perconalab/pmm-server:dev-latest logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ var_value=perconalab/pmm-server:dev-latest logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/pmm-server:dev-latest == docker.io/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/pmm-server:dev-latest == percona/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/pmm-server:dev-latest == perconalab/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ new_value=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ echo IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | IMAGE_PMM_SERVER=docker.io/perconalab/pmm-server:dev-latest logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ for var in $(printenv | grep -E '^IMAGE' | awk -F'=' '{print $1}') logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | ++++ eval 'echo $IMAGE_PMM_CLIENT' logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++++ echo perconalab/pmm-client:dev-latest logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ var_value=perconalab/pmm-client:dev-latest logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/pmm-client:dev-latest == docker.io/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/pmm-client:dev-latest == percona/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ [[ perconalab/pmm-client:dev-latest == perconalab/* ]] logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ new_value=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ export IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ echo IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | IMAGE_PMM_CLIENT=docker.io/perconalab/pmm-client:dev-latest logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ echo docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | ++++ which gdate logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | ++++ which date logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ date=/usr/sbin/date logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | ++++ which gsed logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1506/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | ++++ which sed logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | +++ sed=/usr/sbin/sed logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | ++ oc get projects logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | + remove_all_finalizers logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | + resource_types=("pg-restore" "pg-backup" "pg") logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | + for resource in "${resource_types[@]}" logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | + echo 'removing all finalizers for pg-restore resources' logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | removing all finalizers for pg-restore resources logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | + kubectl -n kuttl-test-growing-killdeer get pg-restore -o json logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | + jq '.items[] | .metadata.name' -r logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | + IFS= logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | + read -r name logger.go:42: 03:57:57 | demand-backup/99-remove-cluster-gracefully | + kubectl -n kuttl-test-growing-killdeer delete pg-restore demand-backup-restore-repo1 --wait=0 logger.go:42: 03:57:58 | demand-backup/99-remove-cluster-gracefully | perconapgrestore.pgv2.percona.com "demand-backup-restore-repo1" deleted from kuttl-test-growing-killdeer namespace logger.go:42: 03:57:58 | demand-backup/99-remove-cluster-gracefully | ++ kubectl -n kuttl-test-growing-killdeer get pg-restore demand-backup-restore-repo1 -o yaml logger.go:42: 03:57:58 | demand-backup/99-remove-cluster-gracefully | ++ yq '.metadata.finalizers | length' logger.go:42: 03:57:58 | demand-backup/99-remove-cluster-gracefully | + [[ 1 == \0 ]] logger.go:42: 03:57:58 | demand-backup/99-remove-cluster-gracefully | + kubectl -n kuttl-test-growing-killdeer patch pg-restore demand-backup-restore-repo1 --type=json '-p=[{"op": "remove", "path": "/metadata/finalizers"}]' logger.go:42: 03:57:59 | demand-backup/99-remove-cluster-gracefully | perconapgrestore.pgv2.percona.com/demand-backup-restore-repo1 patched logger.go:42: 03:57:59 | demand-backup/99-remove-cluster-gracefully | + IFS= logger.go:42: 03:57:59 | demand-backup/99-remove-cluster-gracefully | + read -r name logger.go:42: 03:57:59 | demand-backup/99-remove-cluster-gracefully | + kubectl -n kuttl-test-growing-killdeer delete pg-restore demand-backup-restore-repo3 --wait=0 logger.go:42: 03:58:00 | demand-backup/99-remove-cluster-gracefully | perconapgrestore.pgv2.percona.com "demand-backup-restore-repo3" deleted from kuttl-test-growing-killdeer namespace logger.go:42: 03:58:00 | demand-backup/99-remove-cluster-gracefully | ++ kubectl -n kuttl-test-growing-killdeer get pg-restore demand-backup-restore-repo3 -o yaml logger.go:42: 03:58:00 | demand-backup/99-remove-cluster-gracefully | ++ yq '.metadata.finalizers | length' logger.go:42: 03:58:00 | demand-backup/99-remove-cluster-gracefully | + [[ 1 == \0 ]] logger.go:42: 03:58:00 | demand-backup/99-remove-cluster-gracefully | + kubectl -n kuttl-test-growing-killdeer patch pg-restore demand-backup-restore-repo3 --type=json '-p=[{"op": "remove", "path": "/metadata/finalizers"}]' logger.go:42: 03:58:01 | demand-backup/99-remove-cluster-gracefully | perconapgrestore.pgv2.percona.com/demand-backup-restore-repo3 patched logger.go:42: 03:58:01 | demand-backup/99-remove-cluster-gracefully | + IFS= logger.go:42: 03:58:01 | demand-backup/99-remove-cluster-gracefully | + read -r name logger.go:42: 03:58:01 | demand-backup/99-remove-cluster-gracefully | + for resource in "${resource_types[@]}" logger.go:42: 03:58:01 | demand-backup/99-remove-cluster-gracefully | + echo 'removing all finalizers for pg-backup resources' logger.go:42: 03:58:01 | demand-backup/99-remove-cluster-gracefully | removing all finalizers for pg-backup resources logger.go:42: 03:58:01 | demand-backup/99-remove-cluster-gracefully | + kubectl -n kuttl-test-growing-killdeer get pg-backup -o json logger.go:42: 03:58:01 | demand-backup/99-remove-cluster-gracefully | + jq '.items[] | .metadata.name' -r logger.go:42: 03:58:01 | demand-backup/99-remove-cluster-gracefully | + IFS= logger.go:42: 03:58:01 | demand-backup/99-remove-cluster-gracefully | + read -r name logger.go:42: 03:58:01 | demand-backup/99-remove-cluster-gracefully | + for resource in "${resource_types[@]}" logger.go:42: 03:58:01 | demand-backup/99-remove-cluster-gracefully | + echo 'removing all finalizers for pg resources' logger.go:42: 03:58:01 | demand-backup/99-remove-cluster-gracefully | removing all finalizers for pg resources logger.go:42: 03:58:01 | demand-backup/99-remove-cluster-gracefully | + kubectl -n kuttl-test-growing-killdeer get pg -o json logger.go:42: 03:58:01 | demand-backup/99-remove-cluster-gracefully | + jq '.items[] | .metadata.name' -r logger.go:42: 03:58:01 | demand-backup/99-remove-cluster-gracefully | + IFS= logger.go:42: 03:58:01 | demand-backup/99-remove-cluster-gracefully | + read -r name logger.go:42: 03:58:02 | demand-backup/99-remove-cluster-gracefully | + check_operator_panic logger.go:42: 03:58:02 | demand-backup/99-remove-cluster-gracefully | ++ get_operator_pod logger.go:42: 03:58:02 | demand-backup/99-remove-cluster-gracefully | +++ kubectl get pods -n pg-operator --selector=app.kubernetes.io/name=percona-postgresql-operator -o 'jsonpath={.items[].metadata.name}' logger.go:42: 03:58:02 | demand-backup/99-remove-cluster-gracefully | ++ echo percona-postgresql-operator-6446fff4bd-82g8d logger.go:42: 03:58:02 | demand-backup/99-remove-cluster-gracefully | + local operator_pod=percona-postgresql-operator-6446fff4bd-82g8d logger.go:42: 03:58:02 | demand-backup/99-remove-cluster-gracefully | + local panic_log logger.go:42: 03:58:02 | demand-backup/99-remove-cluster-gracefully | ++ kubectl logs -n pg-operator percona-postgresql-operator-6446fff4bd-82g8d -c operator logger.go:42: 03:58:02 | demand-backup/99-remove-cluster-gracefully | ++ grep -A 100 'Observed a panic' logger.go:42: 03:58:03 | demand-backup/99-remove-cluster-gracefully | ++ true logger.go:42: 03:58:03 | demand-backup/99-remove-cluster-gracefully | + panic_log= logger.go:42: 03:58:03 | demand-backup/99-remove-cluster-gracefully | + '[' -n '' ']' logger.go:42: 03:58:03 | demand-backup/99-remove-cluster-gracefully | + destroy_operator logger.go:42: 03:58:03 | demand-backup/99-remove-cluster-gracefully | + kubectl -n pg-operator delete deployment percona-postgresql-operator --force --grace-period=0 logger.go:42: 03:58:03 | demand-backup/99-remove-cluster-gracefully | Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. logger.go:42: 03:58:04 | demand-backup/99-remove-cluster-gracefully | deployment.apps "percona-postgresql-operator" force deleted from pg-operator namespace logger.go:42: 03:58:04 | demand-backup/99-remove-cluster-gracefully | + [[ -n pg-operator ]] logger.go:42: 03:58:04 | demand-backup/99-remove-cluster-gracefully | + kubectl delete namespace pg-operator --force --grace-period=0 logger.go:42: 03:58:04 | demand-backup/99-remove-cluster-gracefully | Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. logger.go:42: 03:58:04 | demand-backup/99-remove-cluster-gracefully | namespace "pg-operator" force deleted logger.go:42: 03:58:16 | demand-backup/99-remove-cluster-gracefully | test step completed 99-remove-cluster-gracefully logger.go:42: 03:58:16 | demand-backup | demand-backup events from ns kuttl-test-growing-killdeer: logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:36 +0000 UTC Normal Pod pg-client-65d98588cc-8jjx9 Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/pg-client-65d98588cc-8jjx9 to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-5f1h default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:36 +0000 UTC Normal ReplicaSet.apps pg-client-65d98588cc SuccessfulCreate Created pod: pg-client-65d98588cc-8jjx9 replicaset-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:36 +0000 UTC Normal Deployment.apps pg-client ScalingReplicaSet Scaled up replica set pg-client-65d98588cc from 0 to 1 deployment-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:37 +0000 UTC Normal Pod pg-client-65d98588cc-8jjx9.spec.containers{pg-client} Pulling Pulling image "perconalab/percona-distribution-postgresql:16" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:37 +0000 UTC Normal Pod pg-client-65d98588cc-8jjx9.spec.containers{pg-client} Pulled Successfully pulled image "perconalab/percona-distribution-postgresql:16" in 241ms (241ms including waiting). Image size: 564259787 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:37 +0000 UTC Normal Pod pg-client-65d98588cc-8jjx9.spec.containers{pg-client} Created Created container: pg-client kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:37 +0000 UTC Normal Pod pg-client-65d98588cc-8jjx9.spec.containers{pg-client} Started Started container pg-client kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:43 +0000 UTC Normal PersistentVolumeClaim demand-backup-instance1-5kbg-pgdata WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:43 +0000 UTC Normal PersistentVolumeClaim demand-backup-instance1-r2k8-pgdata WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:43 +0000 UTC Normal PersistentVolumeClaim demand-backup-instance1-tv29-pgdata WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:43 +0000 UTC Normal PersistentVolumeClaim demand-backup-instance1-tv29-pgdata ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:43 +0000 UTC Normal PersistentVolumeClaim demand-backup-instance1-tv29-pgdata Provisioning External provisioner is provisioning volume for claim "kuttl-test-growing-killdeer/demand-backup-instance1-tv29-pgdata" pd.csi.storage.gke.io_gke-5df0b62e6b744c5ab008-1867-5d70-vm_44ec33d2-f0f8-4b5c-af89-26efd3f7b41b logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:43 +0000 UTC Normal StatefulSet.apps demand-backup-instance1-tv29 SuccessfulCreate create Pod demand-backup-instance1-tv29-0 in StatefulSet demand-backup-instance1-tv29 successful statefulset-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:44 +0000 UTC Normal PersistentVolumeClaim demand-backup-instance1-5kbg-pgdata ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:44 +0000 UTC Normal PersistentVolumeClaim demand-backup-instance1-5kbg-pgdata Provisioning External provisioner is provisioning volume for claim "kuttl-test-growing-killdeer/demand-backup-instance1-5kbg-pgdata" pd.csi.storage.gke.io_gke-5df0b62e6b744c5ab008-1867-5d70-vm_44ec33d2-f0f8-4b5c-af89-26efd3f7b41b logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:44 +0000 UTC Normal StatefulSet.apps demand-backup-instance1-5kbg SuccessfulCreate create Pod demand-backup-instance1-5kbg-0 in StatefulSet demand-backup-instance1-5kbg successful statefulset-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:44 +0000 UTC Normal PersistentVolumeClaim demand-backup-instance1-r2k8-pgdata ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:44 +0000 UTC Normal PersistentVolumeClaim demand-backup-instance1-r2k8-pgdata Provisioning External provisioner is provisioning volume for claim "kuttl-test-growing-killdeer/demand-backup-instance1-r2k8-pgdata" pd.csi.storage.gke.io_gke-5df0b62e6b744c5ab008-1867-5d70-vm_44ec33d2-f0f8-4b5c-af89-26efd3f7b41b logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:44 +0000 UTC Normal StatefulSet.apps demand-backup-instance1-r2k8 SuccessfulCreate create Pod demand-backup-instance1-r2k8-0 in StatefulSet demand-backup-instance1-r2k8 successful statefulset-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:45 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-hhbk5 Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-pgbouncer-7cd7d68587-hhbk5 to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-dsht default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:45 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-qmqwt Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-pgbouncer-7cd7d68587-qmqwt to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-c949 default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:45 +0000 UTC Normal ReplicaSet.apps demand-backup-pgbouncer-7cd7d68587 SuccessfulCreate Created pod: demand-backup-pgbouncer-7cd7d68587-qmqwt replicaset-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:45 +0000 UTC Normal ReplicaSet.apps demand-backup-pgbouncer-7cd7d68587 SuccessfulCreate Created pod: demand-backup-pgbouncer-7cd7d68587-hhbk5 replicaset-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:45 +0000 UTC Normal ReplicaSet.apps demand-backup-pgbouncer-7cd7d68587 SuccessfulCreate Created pod: demand-backup-pgbouncer-7cd7d68587-2jlxw replicaset-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:45 +0000 UTC Normal Deployment.apps demand-backup-pgbouncer ScalingReplicaSet Scaled up replica set demand-backup-pgbouncer-7cd7d68587 from 0 to 3 deployment-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:45 +0000 UTC Normal PodDisruptionBudget.policy demand-backup-pgbouncer NoPods No matching pods found controllermanager logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:45 +0000 UTC Normal Pod demand-backup-repo-host-0 Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-repo-host-0 to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-5f1h default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:45 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:45 +0000 UTC Normal StatefulSet.apps demand-backup-repo-host SuccessfulCreate create Pod demand-backup-repo-host-0 in StatefulSet demand-backup-repo-host successful statefulset-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:45 +0000 UTC Normal PostgresCluster.postgres-operator.crunchydata.com demand-backup RepoHostCreated created pgBackRest repository host StatefulSet/demand-backup-repo-host postgrescluster-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:46 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-2jlxw Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-pgbouncer-7cd7d68587-2jlxw to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-5f1h default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:46 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-2jlxw.spec.containers{pgbouncer} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:46 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-2jlxw.spec.containers{pgbouncer} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18" in 195ms (195ms including waiting). Image size: 85546945 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:46 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-2jlxw.spec.containers{pgbouncer} Created Created container: pgbouncer kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:46 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-2jlxw.spec.containers{pgbouncer} Started Started container pgbouncer kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:46 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-2jlxw.spec.containers{pgbouncer-config} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:46 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-hhbk5.spec.containers{pgbouncer} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:46 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-hhbk5.spec.containers{pgbouncer} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18" in 259ms (259ms including waiting). Image size: 85546945 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:46 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-hhbk5.spec.containers{pgbouncer} Created Created container: pgbouncer kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:46 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-hhbk5.spec.containers{pgbouncer} Started Started container pgbouncer kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:46 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-hhbk5.spec.containers{pgbouncer-config} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:46 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-qmqwt.spec.containers{pgbouncer} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:46 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-qmqwt.spec.containers{pgbouncer} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18" in 242ms (242ms including waiting). Image size: 85546945 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:46 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-qmqwt.spec.containers{pgbouncer} Created Created container: pgbouncer kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:46 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-qmqwt.spec.containers{pgbouncer} Started Started container pgbouncer kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:46 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-qmqwt.spec.containers{pgbouncer-config} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:46 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-qmqwt.spec.containers{pgbouncer-config} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18" in 238ms (238ms including waiting). Image size: 85546945 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:46 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 223ms (223ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:46 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:46 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:47 +0000 UTC Normal PersistentVolumeClaim demand-backup-instance1-r2k8-pgdata ProvisioningSucceeded Successfully provisioned volume pvc-def14fb6-2481-42b9-9df4-8ed7abdf6da5 pd.csi.storage.gke.io_gke-5df0b62e6b744c5ab008-1867-5d70-vm_44ec33d2-f0f8-4b5c-af89-26efd3f7b41b logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:47 +0000 UTC Normal Pod demand-backup-instance1-tv29-0 Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-instance1-tv29-0 to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-5f1h default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:47 +0000 UTC Normal PersistentVolumeClaim demand-backup-instance1-tv29-pgdata ProvisioningSucceeded Successfully provisioned volume pvc-048c68bd-1efe-4949-be9d-35436d6c9253 pd.csi.storage.gke.io_gke-5df0b62e6b744c5ab008-1867-5d70-vm_44ec33d2-f0f8-4b5c-af89-26efd3f7b41b logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:47 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-2jlxw.spec.containers{pgbouncer-config} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18" in 226ms (226ms including waiting). Image size: 85546945 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:47 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-2jlxw.spec.containers{pgbouncer-config} Created Created container: pgbouncer-config kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:47 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-2jlxw.spec.containers{pgbouncer-config} Started Started container pgbouncer-config kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:47 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-hhbk5.spec.containers{pgbouncer-config} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18" in 221ms (221ms including waiting). Image size: 85546945 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:47 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-hhbk5.spec.containers{pgbouncer-config} Created Created container: pgbouncer-config kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:47 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-hhbk5.spec.containers{pgbouncer-config} Started Started container pgbouncer-config kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:47 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-qmqwt.spec.containers{pgbouncer-config} Created Created container: pgbouncer-config kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:47 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-qmqwt.spec.containers{pgbouncer-config} Started Started container pgbouncer-config kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:47 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:47 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 198ms (198ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:47 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:47 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:47 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.containers{pgbackrest-config} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:47 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.containers{pgbackrest-config} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 145ms (145ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:47 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.containers{pgbackrest-config} Created Created container: pgbackrest-config kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:48 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0 Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-instance1-5kbg-0 to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-dsht default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:48 +0000 UTC Normal PersistentVolumeClaim demand-backup-instance1-5kbg-pgdata ProvisioningSucceeded Successfully provisioned volume pvc-cab77847-a8e2-4fc9-ae6f-7295ae877ee8 pd.csi.storage.gke.io_gke-5df0b62e6b744c5ab008-1867-5d70-vm_44ec33d2-f0f8-4b5c-af89-26efd3f7b41b logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:48 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0 Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-instance1-r2k8-0 to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-c949 default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:48 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.containers{pgbackrest-config} Started Started container pgbackrest-config kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:52 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-def14fb6-2481-42b9-9df4-8ed7abdf6da5" attachdetach-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:53 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-cab77847-a8e2-4fc9-ae6f-7295ae877ee8" attachdetach-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:53 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{database-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:53 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{database-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" in 233ms (233ms including waiting). Image size: 88441640 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:53 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{database-init} Created Created container: database-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:54 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{database-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:54 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{database-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" in 244ms (244ms including waiting). Image size: 88441640 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:54 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{database-init} Created Created container: database-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:54 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{database-init} Started Started container database-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:54 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{database-init} Started Started container database-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:54 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{postgres-startup} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:55 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{postgres-startup} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:55 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{postgres-startup} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 229ms (229ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:55 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{postgres-startup} Created Created container: postgres-startup kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:55 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{postgres-startup} Started Started container postgres-startup kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:55 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{postgres-startup} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 198ms (199ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:55 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{postgres-startup} Created Created container: postgres-startup kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:55 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{postgres-startup} Started Started container postgres-startup kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:55 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:55 +0000 UTC Normal Pod demand-backup-instance1-tv29-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-048c68bd-1efe-4949-be9d-35436d6c9253" attachdetach-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:56 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:56 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 177ms (177ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:56 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:56 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:56 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 193ms (193ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:56 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:56 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:56 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{database} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:56 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{database-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:56 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{database-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" in 223ms (223ms including waiting). Image size: 88441640 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:56 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{database-init} Created Created container: database-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:56 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{database-init} Started Started container database-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:57 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{database} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:57 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{database} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 233ms (233ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:57 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{database} Created Created container: database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:57 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{database} Started Started container database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:57 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{replication-cert-copy} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:57 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{database} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 206ms (206ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:57 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{database} Created Created container: database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:57 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{database} Started Started container database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:57 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{replication-cert-copy} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:57 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{replication-cert-copy} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 273ms (273ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:57 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{replication-cert-copy} Created Created container: replication-cert-copy kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:57 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{replication-cert-copy} Started Started container replication-cert-copy kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:57 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:57 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 185ms (185ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:57 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:57 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{postgres-startup} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:57 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{postgres-startup} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 216ms (216ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:57 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{postgres-startup} Created Created container: postgres-startup kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:57 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{postgres-startup} Started Started container postgres-startup kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:58 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{replication-cert-copy} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 222ms (222ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:58 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{replication-cert-copy} Created Created container: replication-cert-copy kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:58 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{replication-cert-copy} Started Started container replication-cert-copy kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:58 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:58 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 187ms (187ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:58 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:58 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:58 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{pgbackrest-config} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:58 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:58 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{pgbackrest-config} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:58 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:58 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 213ms (213ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:58 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:58 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:59 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{database} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:59 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{database} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 243ms (243ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:59 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{database} Created Created container: database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:59 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{database} Started Started container database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:26:59 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{replication-cert-copy} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:00 +0000 UTC Warning Pod demand-backup-instance1-5kbg-0.spec.containers{database} Unhealthy Readiness probe failed: kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:00 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{replication-cert-copy} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 243ms (243ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:00 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{replication-cert-copy} Created Created container: replication-cert-copy kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:00 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{replication-cert-copy} Started Started container replication-cert-copy kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:00 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:00 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 182ms (182ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:00 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:00 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:00 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{pgbackrest-config} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:02 +0000 UTC Warning Pod demand-backup-instance1-tv29-0.spec.containers{database} Unhealthy Readiness probe failed: kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:07 +0000 UTC Normal Pod demand-backup-backup-dn8b-d7bg9 Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-backup-dn8b-d7bg9 to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-c949 default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:07 +0000 UTC Normal Job.batch demand-backup-backup-dn8b SuccessfulCreate Created pod: demand-backup-backup-dn8b-d7bg9 job-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:07 +0000 UTC Normal PostgresCluster.postgres-operator.crunchydata.com demand-backup StanzasCreated pgBackRest stanza creation completed successfully postgrescluster-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:08 +0000 UTC Normal Pod demand-backup-backup-dn8b-d7bg9.spec.initContainers{pgbackrest-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:08 +0000 UTC Normal Pod demand-backup-backup-dn8b-d7bg9.spec.initContainers{pgbackrest-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" in 199ms (199ms including waiting). Image size: 88441640 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:08 +0000 UTC Normal Pod demand-backup-backup-dn8b-d7bg9.spec.initContainers{pgbackrest-init} Created Created container: pgbackrest-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:08 +0000 UTC Normal Pod demand-backup-backup-dn8b-d7bg9.spec.initContainers{pgbackrest-init} Started Started container pgbackrest-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:09 +0000 UTC Normal Pod demand-backup-backup-dn8b-d7bg9.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:10 +0000 UTC Normal Pod demand-backup-backup-dn8b-d7bg9.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 228ms (228ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:10 +0000 UTC Normal Pod demand-backup-backup-dn8b-d7bg9.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:10 +0000 UTC Normal Pod demand-backup-backup-dn8b-d7bg9.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:12 +0000 UTC Warning PostgresCluster.postgres-operator.crunchydata.com demand-backup UnableToCreateStanzas 2026-03-19 03:27:12.060 P00 DEBUG: common/io/socket/common::sckInit: (block: false, keepAlive: true, tcpKeepAliveCount: 0, tcpKeepAliveIdle: 0, tcpKeepAliveInterval: 0) 2026-03-19 03:27:12.060 P00 DEBUG: common/io/socket/common::sckInit: => void 2026-03-19 03:27:12.063 P00 INFO: stanza-create command begin 2.58.0: --exec-id=361-1fe8aa67 --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db 2026-03-19 03:27:12.063 P00 DEBUG: common/lock::lockInit: (path: {"/tmp/pgbackrest"}, execId: {"361-1fe8aa67"}) 2026-03-19 03:27:12.063 P00 DEBUG: storage/posix/storage::storagePosixNew: (path: {"/tmp/pgbackrest"}, param.modeFile: 0000, param.modePath: 0000, param.write: true, param.noSymLink: false, param.pathExpressionFunction: null) 2026-03-19 03:27:12.063 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: (type: posix, path: {"/tmp/pgbackrest"}, modeFile: 0640, modePath: 0750, write: true, pathExpressionFunction: null, pathSync: true, symLink: true) 2026-03-19 03:27:12.063 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:27:12.063 P00 DEBUG: storage/posix/storage::storagePosixNew: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:27:12.063 P00 DEBUG: common/lock::lockInit: => void 2026-03-19 03:27:12.063 P00 DEBUG: command/lock::cmdLockAcquire: (param.returnOnNoLock: false) 2026-03-19 03:27:12.063 P00 DEBUG: command/exit::exitSafe: (result: 0, error: true, signalType: 0) 2026-03-19 03:27:12.063 P00 ERROR: [050]: unable to acquire lock on file '/tmp/pgbackrest/db-archive-1.lock': Resource temporarily unavailable HINT: is another pgBackRest process running? -------------------------------------------------------------------- If SUBMITTING AN ISSUE please provide the following information: version: 2.58.0 command: stanza-create options: --exec-id=361-1fe8aa67 --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db stack trace: common/lock.c:lockAcquire:254:(trace log level required for parameters) command/lock.c:cmdLockAcquire:(param.returnOnNoLock: false) config/load.c:cfgLoad:(debug log level required for parameters) main.c:main:(debug log level required for parameters) -------------------------------------------------------------------- 2026-03-19 03:27:12.064 P00 INFO: stanza-create command end: aborted with exception [050] 2026-03-19 03:27:12.064 P00 DEBUG: command/exit::exitSafe: => 50 2026-03-19 03:27:12.064 P00 DEBUG: main::main: => 50 2026-03-19 03:27:12.071 P00 DEBUG: common/io/socket/common::sckInit: (block: false, keepAlive: true, tcpKeepAliveCount: 0, tcpKeepAliveIdle: 0, tcpKeepAliveInterval: 0) 2026-03-19 03:27:12.071 P00 DEBUG: common/io/socket/common::sckInit: => void 2026-03-19 03:27:12.074 P00 INFO: stanza-upgrade command begin 2.58.0: --exec-id=362-1f4e53ee --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db 2026-03-19 03:27:12.074 P00 DEBUG: common/lock::lockInit: (path: {"/tmp/pgbackrest"}, execId: {"362-1f4e53ee"}) 2026-03-19 03:27:12.074 P00 DEBUG: storage/posix/storage::storagePosixNew: (path: {"/tmp/pgbackrest"}, param.modeFile: 0000, param.modePath: 0000, param.write: true, param.noSymLink: false, param.pathExpressionFunction: null) 2026-03-19 03:27:12.074 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: (type: posix, path: {"/tmp/pgbackrest"}, modeFile: 0640, modePath: 0750, write: true, pathExpressionFunction: null, pathSync: true, symLink: true) 2026-03-19 03:27:12.074 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:27:12.074 P00 DEBUG: storage/posix/storage::storagePosixNew: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:27:12.074 P00 DEBUG: common/lock::lockInit: => void 2026-03-19 03:27:12.074 P00 DEBUG: command/lock::cmdLockAcquire: (param.returnOnNoLock: false) 2026-03-19 03:27:12.074 P00 DEBUG: command/exit::exitSafe: (result: 0, error: true, signalType: 0) 2026-03-19 03:27:12.074 P00 ERROR: [050]: unable to acquire lock on file '/tmp/pgbackrest/db-archive-1.lock': Resource temporarily unavailable HINT: is another pgBackRest process running? -------------------------------------------------------------------- If SUBMITTING AN ISSUE please provide the following information: version: 2.58.0 command: stanza-upgrade options: --exec-id=362-1f4e53ee --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db stack trace: common/lock.c:lockAcquire:254:(trace log level required for parameters) command/lock.c:cmdLockAcquire:(param.returnOnNoLock: false) config/load.c:cfgLoad:(debug log level required for parameters) main.c:main:(debug log level required for parameters) -------------------------------------------------------------------- 2026-03-19 03:27:12.074 P00 INFO: stanza-upgrade command end: aborted with exception [050] 2026-03-19 03:27:12.074 P00 DEBUG: command/exit::exitSafe: => 50 2026-03-19 03:27:12.074 P00 DEBUG: main::main: => 50 : command terminated with exit code 50 postgrescluster-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:16 +0000 UTC Warning PostgresCluster.postgres-operator.crunchydata.com demand-backup UnableToCreateStanzas 2026-03-19 03:27:16.511 P00 DEBUG: common/io/socket/common::sckInit: (block: false, keepAlive: true, tcpKeepAliveCount: 0, tcpKeepAliveIdle: 0, tcpKeepAliveInterval: 0) 2026-03-19 03:27:16.511 P00 DEBUG: common/io/socket/common::sckInit: => void 2026-03-19 03:27:16.514 P00 INFO: stanza-create command begin 2.58.0: --exec-id=428-69cd91f9 --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db 2026-03-19 03:27:16.514 P00 DEBUG: common/lock::lockInit: (path: {"/tmp/pgbackrest"}, execId: {"428-69cd91f9"}) 2026-03-19 03:27:16.514 P00 DEBUG: storage/posix/storage::storagePosixNew: (path: {"/tmp/pgbackrest"}, param.modeFile: 0000, param.modePath: 0000, param.write: true, param.noSymLink: false, param.pathExpressionFunction: null) 2026-03-19 03:27:16.514 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: (type: posix, path: {"/tmp/pgbackrest"}, modeFile: 0640, modePath: 0750, write: true, pathExpressionFunction: null, pathSync: true, symLink: true) 2026-03-19 03:27:16.514 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:27:16.514 P00 DEBUG: storage/posix/storage::storagePosixNew: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:27:16.514 P00 DEBUG: common/lock::lockInit: => void 2026-03-19 03:27:16.514 P00 DEBUG: command/lock::cmdLockAcquire: (param.returnOnNoLock: false) 2026-03-19 03:27:16.514 P00 DEBUG: command/exit::exitSafe: (result: 0, error: true, signalType: 0) 2026-03-19 03:27:16.514 P00 ERROR: [050]: unable to acquire lock on file '/tmp/pgbackrest/db-backup-1.lock': Resource temporarily unavailable HINT: is another pgBackRest process running? -------------------------------------------------------------------- If SUBMITTING AN ISSUE please provide the following information: version: 2.58.0 command: stanza-create options: --exec-id=428-69cd91f9 --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db stack trace: common/lock.c:lockAcquire:254:(trace log level required for parameters) command/lock.c:cmdLockAcquire:(param.returnOnNoLock: false) config/load.c:cfgLoad:(debug log level required for parameters) main.c:main:(debug log level required for parameters) -------------------------------------------------------------------- 2026-03-19 03:27:16.514 P00 INFO: stanza-create command end: aborted with exception [050] 2026-03-19 03:27:16.514 P00 DEBUG: command/exit::exitSafe: => 50 2026-03-19 03:27:16.514 P00 DEBUG: main::main: => 50 2026-03-19 03:27:16.522 P00 DEBUG: common/io/socket/common::sckInit: (block: false, keepAlive: true, tcpKeepAliveCount: 0, tcpKeepAliveIdle: 0, tcpKeepAliveInterval: 0) 2026-03-19 03:27:16.522 P00 DEBUG: common/io/socket/common::sckInit: => void 2026-03-19 03:27:16.525 P00 INFO: stanza-upgrade command begin 2.58.0: --exec-id=429-efaf4209 --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db 2026-03-19 03:27:16.525 P00 DEBUG: common/lock::lockInit: (path: {"/tmp/pgbackrest"}, execId: {"429-efaf4209"}) 2026-03-19 03:27:16.525 P00 DEBUG: storage/posix/storage::storagePosixNew: (path: {"/tmp/pgbackrest"}, param.modeFile: 0000, param.modePath: 0000, param.write: true, param.noSymLink: false, param.pathExpressionFunction: null) 2026-03-19 03:27:16.525 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: (type: posix, path: {"/tmp/pgbackrest"}, modeFile: 0640, modePath: 0750, write: true, pathExpressionFunction: null, pathSync: true, symLink: true) 2026-03-19 03:27:16.525 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:27:16.525 P00 DEBUG: storage/posix/storage::storagePosixNew: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:27:16.525 P00 DEBUG: common/lock::lockInit: => void 2026-03-19 03:27:16.525 P00 DEBUG: command/lock::cmdLockAcquire: (param.returnOnNoLock: false) 2026-03-19 03:27:16.525 P00 DEBUG: command/exit::exitSafe: (result: 0, error: true, signalType: 0) 2026-03-19 03:27:16.525 P00 ERROR: [050]: unable to acquire lock on file '/tmp/pgbackrest/db-backup-1.lock': Resource temporarily unavailable HINT: is another pgBackRest process running? -------------------------------------------------------------------- If SUBMITTING AN ISSUE please provide the following information: version: 2.58.0 command: stanza-upgrade options: --exec-id=429-efaf4209 --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db stack trace: common/lock.c:lockAcquire:254:(trace log level required for parameters) command/lock.c:cmdLockAcquire:(param.returnOnNoLock: false) config/load.c:cfgLoad:(debug log level required for parameters) main.c:main:(debug log level required for parameters) -------------------------------------------------------------------- 2026-03-19 03:27:16.525 P00 INFO: stanza-upgrade command end: aborted with exception [050] 2026-03-19 03:27:16.525 P00 DEBUG: command/exit::exitSafe: => 50 2026-03-19 03:27:16.525 P00 DEBUG: main::main: => 50 : command terminated with exit code 50 postgrescluster-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:19 +0000 UTC Warning PostgresCluster.postgres-operator.crunchydata.com demand-backup UnableToCreateStanzas 2026-03-19 03:27:19.699 P00 DEBUG: common/io/socket/common::sckInit: (block: false, keepAlive: true, tcpKeepAliveCount: 0, tcpKeepAliveIdle: 0, tcpKeepAliveInterval: 0) 2026-03-19 03:27:19.699 P00 DEBUG: common/io/socket/common::sckInit: => void 2026-03-19 03:27:19.701 P00 INFO: stanza-create command begin 2.58.0: --exec-id=467-93e6e92e --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db 2026-03-19 03:27:19.701 P00 DEBUG: common/lock::lockInit: (path: {"/tmp/pgbackrest"}, execId: {"467-93e6e92e"}) 2026-03-19 03:27:19.701 P00 DEBUG: storage/posix/storage::storagePosixNew: (path: {"/tmp/pgbackrest"}, param.modeFile: 0000, param.modePath: 0000, param.write: true, param.noSymLink: false, param.pathExpressionFunction: null) 2026-03-19 03:27:19.701 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: (type: posix, path: {"/tmp/pgbackrest"}, modeFile: 0640, modePath: 0750, write: true, pathExpressionFunction: null, pathSync: true, symLink: true) 2026-03-19 03:27:19.701 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:27:19.701 P00 DEBUG: storage/posix/storage::storagePosixNew: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:27:19.701 P00 DEBUG: common/lock::lockInit: => void 2026-03-19 03:27:19.701 P00 DEBUG: command/lock::cmdLockAcquire: (param.returnOnNoLock: false) 2026-03-19 03:27:19.702 P00 DEBUG: command/exit::exitSafe: (result: 0, error: true, signalType: 0) 2026-03-19 03:27:19.702 P00 ERROR: [050]: unable to acquire lock on file '/tmp/pgbackrest/db-backup-1.lock': Resource temporarily unavailable HINT: is another pgBackRest process running? -------------------------------------------------------------------- If SUBMITTING AN ISSUE please provide the following information: version: 2.58.0 command: stanza-create options: --exec-id=467-93e6e92e --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db stack trace: common/lock.c:lockAcquire:254:(trace log level required for parameters) command/lock.c:cmdLockAcquire:(param.returnOnNoLock: false) config/load.c:cfgLoad:(debug log level required for parameters) main.c:main:(debug log level required for parameters) -------------------------------------------------------------------- 2026-03-19 03:27:19.702 P00 INFO: stanza-create command end: aborted with exception [050] 2026-03-19 03:27:19.702 P00 DEBUG: command/exit::exitSafe: => 50 2026-03-19 03:27:19.702 P00 DEBUG: main::main: => 50 2026-03-19 03:27:19.709 P00 DEBUG: common/io/socket/common::sckInit: (block: false, keepAlive: true, tcpKeepAliveCount: 0, tcpKeepAliveIdle: 0, tcpKeepAliveInterval: 0) 2026-03-19 03:27:19.709 P00 DEBUG: common/io/socket/common::sckInit: => void 2026-03-19 03:27:19.712 P00 INFO: stanza-upgrade command begin 2.58.0: --exec-id=468-8d5b02ec --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db 2026-03-19 03:27:19.712 P00 DEBUG: common/lock::lockInit: (path: {"/tmp/pgbackrest"}, execId: {"468-8d5b02ec"}) 2026-03-19 03:27:19.712 P00 DEBUG: storage/posix/storage::storagePosixNew: (path: {"/tmp/pgbackrest"}, param.modeFile: 0000, param.modePath: 0000, param.write: true, param.noSymLink: false, param.pathExpressionFunction: null) 2026-03-19 03:27:19.712 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: (type: posix, path: {"/tmp/pgbackrest"}, modeFile: 0640, modePath: 0750, write: true, pathExpressionFunction: null, pathSync: true, symLink: true) 2026-03-19 03:27:19.712 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:27:19.712 P00 DEBUG: storage/posix/storage::storagePosixNew: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:27:19.712 P00 DEBUG: common/lock::lockInit: => void 2026-03-19 03:27:19.712 P00 DEBUG: command/lock::cmdLockAcquire: (param.returnOnNoLock: false) 2026-03-19 03:27:19.712 P00 DEBUG: command/exit::exitSafe: (result: 0, error: true, signalType: 0) 2026-03-19 03:27:19.712 P00 ERROR: [050]: unable to acquire lock on file '/tmp/pgbackrest/db-backup-1.lock': Resource temporarily unavailable HINT: is another pgBackRest process running? -------------------------------------------------------------------- If SUBMITTING AN ISSUE please provide the following information: version: 2.58.0 command: stanza-upgrade options: --exec-id=468-8d5b02ec --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db stack trace: common/lock.c:lockAcquire:254:(trace log level required for parameters) command/lock.c:cmdLockAcquire:(param.returnOnNoLock: false) config/load.c:cfgLoad:(debug log level required for parameters) main.c:main:(debug log level required for parameters) -------------------------------------------------------------------- 2026-03-19 03:27:19.712 P00 INFO: stanza-upgrade command end: aborted with exception [050] 2026-03-19 03:27:19.712 P00 DEBUG: command/exit::exitSafe: => 50 2026-03-19 03:27:19.712 P00 DEBUG: main::main: => 50 : command terminated with exit code 50 postgrescluster-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:32 +0000 UTC Warning PostgresCluster.postgres-operator.crunchydata.com demand-backup UnableToCreateStanzas 2026-03-19 03:27:32.271 P00 DEBUG: common/io/socket/common::sckInit: (block: false, keepAlive: true, tcpKeepAliveCount: 0, tcpKeepAliveIdle: 0, tcpKeepAliveInterval: 0) 2026-03-19 03:27:32.271 P00 DEBUG: common/io/socket/common::sckInit: => void 2026-03-19 03:27:32.273 P00 INFO: stanza-create command begin 2.58.0: --exec-id=511-18490c8e --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db 2026-03-19 03:27:32.273 P00 DEBUG: common/lock::lockInit: (path: {"/tmp/pgbackrest"}, execId: {"511-18490c8e"}) 2026-03-19 03:27:32.273 P00 DEBUG: storage/posix/storage::storagePosixNew: (path: {"/tmp/pgbackrest"}, param.modeFile: 0000, param.modePath: 0000, param.write: true, param.noSymLink: false, param.pathExpressionFunction: null) 2026-03-19 03:27:32.273 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: (type: posix, path: {"/tmp/pgbackrest"}, modeFile: 0640, modePath: 0750, write: true, pathExpressionFunction: null, pathSync: true, symLink: true) 2026-03-19 03:27:32.273 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:27:32.273 P00 DEBUG: storage/posix/storage::storagePosixNew: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:27:32.273 P00 DEBUG: common/lock::lockInit: => void 2026-03-19 03:27:32.273 P00 DEBUG: command/lock::cmdLockAcquire: (param.returnOnNoLock: false) 2026-03-19 03:27:32.274 P00 DEBUG: command/exit::exitSafe: (result: 0, error: true, signalType: 0) 2026-03-19 03:27:32.274 P00 ERROR: [050]: unable to acquire lock on file '/tmp/pgbackrest/db-backup-1.lock': Resource temporarily unavailable HINT: is another pgBackRest process running? -------------------------------------------------------------------- If SUBMITTING AN ISSUE please provide the following information: version: 2.58.0 command: stanza-create options: --exec-id=511-18490c8e --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db stack trace: common/lock.c:lockAcquire:254:(trace log level required for parameters) command/lock.c:cmdLockAcquire:(param.returnOnNoLock: false) config/load.c:cfgLoad:(debug log level required for parameters) main.c:main:(debug log level required for parameters) -------------------------------------------------------------------- 2026-03-19 03:27:32.274 P00 INFO: stanza-create command end: aborted with exception [050] 2026-03-19 03:27:32.274 P00 DEBUG: command/exit::exitSafe: => 50 2026-03-19 03:27:32.274 P00 DEBUG: main::main: => 50 2026-03-19 03:27:32.282 P00 DEBUG: common/io/socket/common::sckInit: (block: false, keepAlive: true, tcpKeepAliveCount: 0, tcpKeepAliveIdle: 0, tcpKeepAliveInterval: 0) 2026-03-19 03:27:32.282 P00 DEBUG: common/io/socket/common::sckInit: => void 2026-03-19 03:27:32.284 P00 INFO: stanza-upgrade command begin 2.58.0: --exec-id=512-58d54b1e --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db 2026-03-19 03:27:32.284 P00 DEBUG: common/lock::lockInit: (path: {"/tmp/pgbackrest"}, execId: {"512-58d54b1e"}) 2026-03-19 03:27:32.284 P00 DEBUG: storage/posix/storage::storagePosixNew: (path: {"/tmp/pgbackrest"}, param.modeFile: 0000, param.modePath: 0000, param.write: true, param.noSymLink: false, param.pathExpressionFunction: null) 2026-03-19 03:27:32.284 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: (type: posix, path: {"/tmp/pgbackrest"}, modeFile: 0640, modePath: 0750, write: true, pathExpressionFunction: null, pathSync: true, symLink: true) 2026-03-19 03:27:32.284 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:27:32.284 P00 DEBUG: storage/posix/storage::storagePosixNew: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:27:32.284 P00 DEBUG: common/lock::lockInit: => void 2026-03-19 03:27:32.284 P00 DEBUG: command/lock::cmdLockAcquire: (param.returnOnNoLock: false) 2026-03-19 03:27:32.284 P00 DEBUG: command/exit::exitSafe: (result: 0, error: true, signalType: 0) 2026-03-19 03:27:32.285 P00 ERROR: [050]: unable to acquire lock on file '/tmp/pgbackrest/db-backup-1.lock': Resource temporarily unavailable HINT: is another pgBackRest process running? -------------------------------------------------------------------- If SUBMITTING AN ISSUE please provide the following information: version: 2.58.0 command: stanza-upgrade options: --exec-id=512-58d54b1e --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db stack trace: common/lock.c:lockAcquire:254:(trace log level required for parameters) command/lock.c:cmdLockAcquire:(param.returnOnNoLock: false) config/load.c:cfgLoad:(debug log level required for parameters) main.c:main:(debug log level required for parameters) -------------------------------------------------------------------- 2026-03-19 03:27:32.285 P00 INFO: stanza-upgrade command end: aborted with exception [050] 2026-03-19 03:27:32.285 P00 DEBUG: command/exit::exitSafe: => 50 2026-03-19 03:27:32.285 P00 DEBUG: main::main: => 50 : command terminated with exit code 50 postgrescluster-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:45 +0000 UTC Warning PostgresCluster.postgres-operator.crunchydata.com demand-backup UnableToCreateStanzas 2026-03-19 03:27:45.090 P00 DEBUG: common/io/socket/common::sckInit: (block: false, keepAlive: true, tcpKeepAliveCount: 0, tcpKeepAliveIdle: 0, tcpKeepAliveInterval: 0) 2026-03-19 03:27:45.090 P00 DEBUG: common/io/socket/common::sckInit: => void 2026-03-19 03:27:45.092 P00 INFO: stanza-create command begin 2.58.0: --exec-id=563-40bc27c2 --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db 2026-03-19 03:27:45.092 P00 DEBUG: common/lock::lockInit: (path: {"/tmp/pgbackrest"}, execId: {"563-40bc27c2"}) 2026-03-19 03:27:45.092 P00 DEBUG: storage/posix/storage::storagePosixNew: (path: {"/tmp/pgbackrest"}, param.modeFile: 0000, param.modePath: 0000, param.write: true, param.noSymLink: false, param.pathExpressionFunction: null) 2026-03-19 03:27:45.093 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: (type: posix, path: {"/tmp/pgbackrest"}, modeFile: 0640, modePath: 0750, write: true, pathExpressionFunction: null, pathSync: true, symLink: true) 2026-03-19 03:27:45.093 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:27:45.093 P00 DEBUG: storage/posix/storage::storagePosixNew: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:27:45.093 P00 DEBUG: common/lock::lockInit: => void 2026-03-19 03:27:45.093 P00 DEBUG: command/lock::cmdLockAcquire: (param.returnOnNoLock: false) 2026-03-19 03:27:45.093 P00 DEBUG: command/exit::exitSafe: (result: 0, error: true, signalType: 0) 2026-03-19 03:27:45.093 P00 ERROR: [050]: unable to acquire lock on file '/tmp/pgbackrest/db-backup-1.lock': Resource temporarily unavailable HINT: is another pgBackRest process running? -------------------------------------------------------------------- If SUBMITTING AN ISSUE please provide the following information: version: 2.58.0 command: stanza-create options: --exec-id=563-40bc27c2 --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db stack trace: common/lock.c:lockAcquire:254:(trace log level required for parameters) command/lock.c:cmdLockAcquire:(param.returnOnNoLock: false) config/load.c:cfgLoad:(debug log level required for parameters) main.c:main:(debug log level required for parameters) -------------------------------------------------------------------- 2026-03-19 03:27:45.093 P00 INFO: stanza-create command end: aborted with exception [050] 2026-03-19 03:27:45.093 P00 DEBUG: command/exit::exitSafe: => 50 2026-03-19 03:27:45.093 P00 DEBUG: main::main: => 50 2026-03-19 03:27:45.101 P00 DEBUG: common/io/socket/common::sckInit: (block: false, keepAlive: true, tcpKeepAliveCount: 0, tcpKeepAliveIdle: 0, tcpKeepAliveInterval: 0) 2026-03-19 03:27:45.101 P00 DEBUG: common/io/socket/common::sckInit: => void 2026-03-19 03:27:45.104 P00 INFO: stanza-upgrade command begin 2.58.0: --exec-id=564-1add3760 --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db 2026-03-19 03:27:45.104 P00 DEBUG: common/lock::lockInit: (path: {"/tmp/pgbackrest"}, execId: {"564-1add3760"}) 2026-03-19 03:27:45.104 P00 DEBUG: storage/posix/storage::storagePosixNew: (path: {"/tmp/pgbackrest"}, param.modeFile: 0000, param.modePath: 0000, param.write: true, param.noSymLink: false, param.pathExpressionFunction: null) 2026-03-19 03:27:45.104 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: (type: posix, path: {"/tmp/pgbackrest"}, modeFile: 0640, modePath: 0750, write: true, pathExpressionFunction: null, pathSync: true, symLink: true) 2026-03-19 03:27:45.104 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:27:45.104 P00 DEBUG: storage/posix/storage::storagePosixNew: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:27:45.104 P00 DEBUG: common/lock::lockInit: => void 2026-03-19 03:27:45.104 P00 DEBUG: command/lock::cmdLockAcquire: (param.returnOnNoLock: false) 2026-03-19 03:27:45.105 P00 DEBUG: command/exit::exitSafe: (result: 0, error: true, signalType: 0) 2026-03-19 03:27:45.105 P00 ERROR: [050]: unable to acquire lock on file '/tmp/pgbackrest/db-backup-1.lock': Resource temporarily unavailable HINT: is another pgBackRest process running? -------------------------------------------------------------------- If SUBMITTING AN ISSUE please provide the following information: version: 2.58.0 command: stanza-upgrade options: --exec-id=564-1add3760 --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db stack trace: common/lock.c:lockAcquire:254:(trace log level required for parameters) command/lock.c:cmdLockAcquire:(param.returnOnNoLock: false) config/load.c:cfgLoad:(debug log level required for parameters) main.c:main:(debug log level required for parameters) -------------------------------------------------------------------- 2026-03-19 03:27:45.105 P00 INFO: stanza-upgrade command end: aborted with exception [050] 2026-03-19 03:27:45.105 P00 DEBUG: command/exit::exitSafe: => 50 2026-03-19 03:27:45.105 P00 DEBUG: main::main: => 50 : command terminated with exit code 50 postgrescluster-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:27:57 +0000 UTC Warning PostgresCluster.postgres-operator.crunchydata.com demand-backup UnableToCreateStanzas 2026-03-19 03:27:57.745 P00 DEBUG: common/io/socket/common::sckInit: (block: false, keepAlive: true, tcpKeepAliveCount: 0, tcpKeepAliveIdle: 0, tcpKeepAliveInterval: 0) 2026-03-19 03:27:57.745 P00 DEBUG: common/io/socket/common::sckInit: => void 2026-03-19 03:27:57.748 P00 INFO: stanza-create command begin 2.58.0: --exec-id=598-721e9d15 --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db 2026-03-19 03:27:57.748 P00 DEBUG: common/lock::lockInit: (path: {"/tmp/pgbackrest"}, execId: {"598-721e9d15"}) 2026-03-19 03:27:57.748 P00 DEBUG: storage/posix/storage::storagePosixNew: (path: {"/tmp/pgbackrest"}, param.modeFile: 0000, param.modePath: 0000, param.write: true, param.noSymLink: false, param.pathExpressionFunction: null) 2026-03-19 03:27:57.748 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: (type: posix, path: {"/tmp/pgbackrest"}, modeFile: 0640, modePath: 0750, write: true, pathExpressionFunction: null, pathSync: true, symLink: true) 2026-03-19 03:27:57.748 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:27:57.748 P00 DEBUG: storage/posix/storage::storagePosixNew: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:27:57.748 P00 DEBUG: common/lock::lockInit: => void 2026-03-19 03:27:57.748 P00 DEBUG: command/lock::cmdLockAcquire: (param.returnOnNoLock: false) 2026-03-19 03:27:57.748 P00 DEBUG: command/exit::exitSafe: (result: 0, error: true, signalType: 0) 2026-03-19 03:27:57.748 P00 ERROR: [050]: unable to acquire lock on file '/tmp/pgbackrest/db-backup-1.lock': Resource temporarily unavailable HINT: is another pgBackRest process running? -------------------------------------------------------------------- If SUBMITTING AN ISSUE please provide the following information: version: 2.58.0 command: stanza-create options: --exec-id=598-721e9d15 --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db stack trace: common/lock.c:lockAcquire:254:(trace log level required for parameters) command/lock.c:cmdLockAcquire:(param.returnOnNoLock: false) config/load.c:cfgLoad:(debug log level required for parameters) main.c:main:(debug log level required for parameters) -------------------------------------------------------------------- 2026-03-19 03:27:57.749 P00 INFO: stanza-create command end: aborted with exception [050] 2026-03-19 03:27:57.749 P00 DEBUG: command/exit::exitSafe: => 50 2026-03-19 03:27:57.749 P00 DEBUG: main::main: => 50 2026-03-19 03:27:57.757 P00 DEBUG: common/io/socket/common::sckInit: (block: false, keepAlive: true, tcpKeepAliveCount: 0, tcpKeepAliveIdle: 0, tcpKeepAliveInterval: 0) 2026-03-19 03:27:57.757 P00 DEBUG: common/io/socket/common::sckInit: => void 2026-03-19 03:27:57.760 P00 INFO: stanza-upgrade command begin 2.58.0: --exec-id=599-a6d6b2ee --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db 2026-03-19 03:27:57.760 P00 DEBUG: common/lock::lockInit: (path: {"/tmp/pgbackrest"}, execId: {"599-a6d6b2ee"}) 2026-03-19 03:27:57.760 P00 DEBUG: storage/posix/storage::storagePosixNew: (path: {"/tmp/pgbackrest"}, param.modeFile: 0000, param.modePath: 0000, param.write: true, param.noSymLink: false, param.pathExpressionFunction: null) 2026-03-19 03:27:57.760 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: (type: posix, path: {"/tmp/pgbackrest"}, modeFile: 0640, modePath: 0750, write: true, pathExpressionFunction: null, pathSync: true, symLink: true) 2026-03-19 03:27:57.760 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:27:57.760 P00 DEBUG: storage/posix/storage::storagePosixNew: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:27:57.760 P00 DEBUG: common/lock::lockInit: => void 2026-03-19 03:27:57.760 P00 DEBUG: command/lock::cmdLockAcquire: (param.returnOnNoLock: false) 2026-03-19 03:27:57.760 P00 DEBUG: command/exit::exitSafe: (result: 0, error: true, signalType: 0) 2026-03-19 03:27:57.760 P00 ERROR: [050]: unable to acquire lock on file '/tmp/pgbackrest/db-backup-1.lock': Resource temporarily unavailable HINT: is another pgBackRest process running? -------------------------------------------------------------------- If SUBMITTING AN ISSUE please provide the following information: version: 2.58.0 command: stanza-upgrade options: --exec-id=599-a6d6b2ee --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db stack trace: common/lock.c:lockAcquire:254:(trace log level required for parameters) command/lock.c:cmdLockAcquire:(param.returnOnNoLock: false) config/load.c:cfgLoad:(debug log level required for parameters) main.c:main:(debug log level required for parameters) -------------------------------------------------------------------- 2026-03-19 03:27:57.760 P00 INFO: stanza-upgrade command end: aborted with exception [050] 2026-03-19 03:27:57.760 P00 DEBUG: command/exit::exitSafe: => 50 2026-03-19 03:27:57.760 P00 DEBUG: main::main: => 50 : command terminated with exit code 50 postgrescluster-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:28:10 +0000 UTC Warning PostgresCluster.postgres-operator.crunchydata.com demand-backup UnableToCreateStanzas 2026-03-19 03:28:10.334 P00 DEBUG: common/io/socket/common::sckInit: (block: false, keepAlive: true, tcpKeepAliveCount: 0, tcpKeepAliveIdle: 0, tcpKeepAliveInterval: 0) 2026-03-19 03:28:10.334 P00 DEBUG: common/io/socket/common::sckInit: => void 2026-03-19 03:28:10.337 P00 INFO: stanza-create command begin 2.58.0: --exec-id=633-ec20fb35 --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db 2026-03-19 03:28:10.337 P00 DEBUG: common/lock::lockInit: (path: {"/tmp/pgbackrest"}, execId: {"633-ec20fb35"}) 2026-03-19 03:28:10.337 P00 DEBUG: storage/posix/storage::storagePosixNew: (path: {"/tmp/pgbackrest"}, param.modeFile: 0000, param.modePath: 0000, param.write: true, param.noSymLink: false, param.pathExpressionFunction: null) 2026-03-19 03:28:10.337 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: (type: posix, path: {"/tmp/pgbackrest"}, modeFile: 0640, modePath: 0750, write: true, pathExpressionFunction: null, pathSync: true, symLink: true) 2026-03-19 03:28:10.337 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:28:10.337 P00 DEBUG: storage/posix/storage::storagePosixNew: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:28:10.337 P00 DEBUG: common/lock::lockInit: => void 2026-03-19 03:28:10.337 P00 DEBUG: command/lock::cmdLockAcquire: (param.returnOnNoLock: false) 2026-03-19 03:28:10.338 P00 DEBUG: command/exit::exitSafe: (result: 0, error: true, signalType: 0) 2026-03-19 03:28:10.338 P00 ERROR: [050]: unable to acquire lock on file '/tmp/pgbackrest/db-backup-1.lock': Resource temporarily unavailable HINT: is another pgBackRest process running? -------------------------------------------------------------------- If SUBMITTING AN ISSUE please provide the following information: version: 2.58.0 command: stanza-create options: --exec-id=633-ec20fb35 --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db stack trace: common/lock.c:lockAcquire:254:(trace log level required for parameters) command/lock.c:cmdLockAcquire:(param.returnOnNoLock: false) config/load.c:cfgLoad:(debug log level required for parameters) main.c:main:(debug log level required for parameters) -------------------------------------------------------------------- 2026-03-19 03:28:10.338 P00 INFO: stanza-create command end: aborted with exception [050] 2026-03-19 03:28:10.338 P00 DEBUG: command/exit::exitSafe: => 50 2026-03-19 03:28:10.338 P00 DEBUG: main::main: => 50 2026-03-19 03:28:10.346 P00 DEBUG: common/io/socket/common::sckInit: (block: false, keepAlive: true, tcpKeepAliveCount: 0, tcpKeepAliveIdle: 0, tcpKeepAliveInterval: 0) 2026-03-19 03:28:10.346 P00 DEBUG: common/io/socket/common::sckInit: => void 2026-03-19 03:28:10.348 P00 INFO: stanza-upgrade command begin 2.58.0: --exec-id=634-5bdcf2b0 --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db 2026-03-19 03:28:10.348 P00 DEBUG: common/lock::lockInit: (path: {"/tmp/pgbackrest"}, execId: {"634-5bdcf2b0"}) 2026-03-19 03:28:10.348 P00 DEBUG: storage/posix/storage::storagePosixNew: (path: {"/tmp/pgbackrest"}, param.modeFile: 0000, param.modePath: 0000, param.write: true, param.noSymLink: false, param.pathExpressionFunction: null) 2026-03-19 03:28:10.348 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: (type: posix, path: {"/tmp/pgbackrest"}, modeFile: 0640, modePath: 0750, write: true, pathExpressionFunction: null, pathSync: true, symLink: true) 2026-03-19 03:28:10.348 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:28:10.348 P00 DEBUG: storage/posix/storage::storagePosixNew: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:28:10.348 P00 DEBUG: common/lock::lockInit: => void 2026-03-19 03:28:10.348 P00 DEBUG: command/lock::cmdLockAcquire: (param.returnOnNoLock: false) 2026-03-19 03:28:10.349 P00 DEBUG: command/exit::exitSafe: (result: 0, error: true, signalType: 0) 2026-03-19 03:28:10.349 P00 ERROR: [050]: unable to acquire lock on file '/tmp/pgbackrest/db-backup-1.lock': Resource temporarily unavailable HINT: is another pgBackRest process running? -------------------------------------------------------------------- If SUBMITTING AN ISSUE please provide the following information: version: 2.58.0 command: stanza-upgrade options: --exec-id=634-5bdcf2b0 --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db stack trace: common/lock.c:lockAcquire:254:(trace log level required for parameters) command/lock.c:cmdLockAcquire:(param.returnOnNoLock: false) config/load.c:cfgLoad:(debug log level required for parameters) main.c:main:(debug log level required for parameters) -------------------------------------------------------------------- 2026-03-19 03:28:10.349 P00 INFO: stanza-upgrade command end: aborted with exception [050] 2026-03-19 03:28:10.349 P00 DEBUG: command/exit::exitSafe: => 50 2026-03-19 03:28:10.349 P00 DEBUG: main::main: => 50 : command terminated with exit code 50 postgrescluster-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:28:22 +0000 UTC Warning PostgresCluster.postgres-operator.crunchydata.com demand-backup UnableToCreateStanzas 2026-03-19 03:28:22.927 P00 DEBUG: common/io/socket/common::sckInit: (block: false, keepAlive: true, tcpKeepAliveCount: 0, tcpKeepAliveIdle: 0, tcpKeepAliveInterval: 0) 2026-03-19 03:28:22.927 P00 DEBUG: common/io/socket/common::sckInit: => void 2026-03-19 03:28:22.929 P00 INFO: stanza-create command begin 2.58.0: --exec-id=689-fd7f49ea --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db 2026-03-19 03:28:22.930 P00 DEBUG: common/lock::lockInit: (path: {"/tmp/pgbackrest"}, execId: {"689-fd7f49ea"}) 2026-03-19 03:28:22.930 P00 DEBUG: storage/posix/storage::storagePosixNew: (path: {"/tmp/pgbackrest"}, param.modeFile: 0000, param.modePath: 0000, param.write: true, param.noSymLink: false, param.pathExpressionFunction: null) 2026-03-19 03:28:22.930 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: (type: posix, path: {"/tmp/pgbackrest"}, modeFile: 0640, modePath: 0750, write: true, pathExpressionFunction: null, pathSync: true, symLink: true) 2026-03-19 03:28:22.930 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:28:22.930 P00 DEBUG: storage/posix/storage::storagePosixNew: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:28:22.930 P00 DEBUG: common/lock::lockInit: => void 2026-03-19 03:28:22.930 P00 DEBUG: command/lock::cmdLockAcquire: (param.returnOnNoLock: false) 2026-03-19 03:28:22.930 P00 DEBUG: command/exit::exitSafe: (result: 0, error: true, signalType: 0) 2026-03-19 03:28:22.930 P00 ERROR: [050]: unable to acquire lock on file '/tmp/pgbackrest/db-backup-1.lock': Resource temporarily unavailable HINT: is another pgBackRest process running? -------------------------------------------------------------------- If SUBMITTING AN ISSUE please provide the following information: version: 2.58.0 command: stanza-create options: --exec-id=689-fd7f49ea --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db stack trace: common/lock.c:lockAcquire:254:(trace log level required for parameters) command/lock.c:cmdLockAcquire:(param.returnOnNoLock: false) config/load.c:cfgLoad:(debug log level required for parameters) main.c:main:(debug log level required for parameters) -------------------------------------------------------------------- 2026-03-19 03:28:22.930 P00 INFO: stanza-create command end: aborted with exception [050] 2026-03-19 03:28:22.930 P00 DEBUG: command/exit::exitSafe: => 50 2026-03-19 03:28:22.930 P00 DEBUG: main::main: => 50 2026-03-19 03:28:22.938 P00 DEBUG: common/io/socket/common::sckInit: (block: false, keepAlive: true, tcpKeepAliveCount: 0, tcpKeepAliveIdle: 0, tcpKeepAliveInterval: 0) 2026-03-19 03:28:22.938 P00 DEBUG: common/io/socket/common::sckInit: => void 2026-03-19 03:28:22.940 P00 INFO: stanza-upgrade command begin 2.58.0: --exec-id=690-f5b4041b --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db 2026-03-19 03:28:22.940 P00 DEBUG: common/lock::lockInit: (path: {"/tmp/pgbackrest"}, execId: {"690-f5b4041b"}) 2026-03-19 03:28:22.940 P00 DEBUG: storage/posix/storage::storagePosixNew: (path: {"/tmp/pgbackrest"}, param.modeFile: 0000, param.modePath: 0000, param.write: true, param.noSymLink: false, param.pathExpressionFunction: null) 2026-03-19 03:28:22.940 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: (type: posix, path: {"/tmp/pgbackrest"}, modeFile: 0640, modePath: 0750, write: true, pathExpressionFunction: null, pathSync: true, symLink: true) 2026-03-19 03:28:22.940 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:28:22.940 P00 DEBUG: storage/posix/storage::storagePosixNew: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:28:22.940 P00 DEBUG: common/lock::lockInit: => void 2026-03-19 03:28:22.940 P00 DEBUG: command/lock::cmdLockAcquire: (param.returnOnNoLock: false) 2026-03-19 03:28:22.940 P00 DEBUG: command/exit::exitSafe: (result: 0, error: true, signalType: 0) 2026-03-19 03:28:22.940 P00 ERROR: [050]: unable to acquire lock on file '/tmp/pgbackrest/db-backup-1.lock': Resource temporarily unavailable HINT: is another pgBackRest process running? -------------------------------------------------------------------- If SUBMITTING AN ISSUE please provide the following information: version: 2.58.0 command: stanza-upgrade options: --exec-id=690-f5b4041b --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db stack trace: common/lock.c:lockAcquire:254:(trace log level required for parameters) command/lock.c:cmdLockAcquire:(param.returnOnNoLock: false) config/load.c:cfgLoad:(debug log level required for parameters) main.c:main:(debug log level required for parameters) -------------------------------------------------------------------- 2026-03-19 03:28:22.940 P00 INFO: stanza-upgrade command end: aborted with exception [050] 2026-03-19 03:28:22.940 P00 DEBUG: command/exit::exitSafe: => 50 2026-03-19 03:28:22.940 P00 DEBUG: main::main: => 50 : command terminated with exit code 50 postgrescluster-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:28:35 +0000 UTC Warning PostgresCluster.postgres-operator.crunchydata.com demand-backup UnableToCreateStanzas 2026-03-19 03:28:35.069 P00 DEBUG: common/io/socket/common::sckInit: (block: false, keepAlive: true, tcpKeepAliveCount: 0, tcpKeepAliveIdle: 0, tcpKeepAliveInterval: 0) 2026-03-19 03:28:35.069 P00 DEBUG: common/io/socket/common::sckInit: => void 2026-03-19 03:28:35.072 P00 INFO: stanza-create command begin 2.58.0: --exec-id=740-aafa49be --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db 2026-03-19 03:28:35.072 P00 DEBUG: common/lock::lockInit: (path: {"/tmp/pgbackrest"}, execId: {"740-aafa49be"}) 2026-03-19 03:28:35.072 P00 DEBUG: storage/posix/storage::storagePosixNew: (path: {"/tmp/pgbackrest"}, param.modeFile: 0000, param.modePath: 0000, param.write: true, param.noSymLink: false, param.pathExpressionFunction: null) 2026-03-19 03:28:35.072 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: (type: posix, path: {"/tmp/pgbackrest"}, modeFile: 0640, modePath: 0750, write: true, pathExpressionFunction: null, pathSync: true, symLink: true) 2026-03-19 03:28:35.072 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:28:35.072 P00 DEBUG: storage/posix/storage::storagePosixNew: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:28:35.072 P00 DEBUG: common/lock::lockInit: => void 2026-03-19 03:28:35.072 P00 DEBUG: command/lock::cmdLockAcquire: (param.returnOnNoLock: false) 2026-03-19 03:28:35.073 P00 DEBUG: command/exit::exitSafe: (result: 0, error: true, signalType: 0) 2026-03-19 03:28:35.073 P00 ERROR: [050]: unable to acquire lock on file '/tmp/pgbackrest/db-backup-1.lock': Resource temporarily unavailable HINT: is another pgBackRest process running? -------------------------------------------------------------------- If SUBMITTING AN ISSUE please provide the following information: version: 2.58.0 command: stanza-create options: --exec-id=740-aafa49be --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db stack trace: common/lock.c:lockAcquire:254:(trace log level required for parameters) command/lock.c:cmdLockAcquire:(param.returnOnNoLock: false) config/load.c:cfgLoad:(debug log level required for parameters) main.c:main:(debug log level required for parameters) -------------------------------------------------------------------- 2026-03-19 03:28:35.073 P00 INFO: stanza-create command end: aborted with exception [050] 2026-03-19 03:28:35.073 P00 DEBUG: command/exit::exitSafe: => 50 2026-03-19 03:28:35.073 P00 DEBUG: main::main: => 50 2026-03-19 03:28:35.081 P00 DEBUG: common/io/socket/common::sckInit: (block: false, keepAlive: true, tcpKeepAliveCount: 0, tcpKeepAliveIdle: 0, tcpKeepAliveInterval: 0) 2026-03-19 03:28:35.081 P00 DEBUG: common/io/socket/common::sckInit: => void 2026-03-19 03:28:35.083 P00 INFO: stanza-upgrade command begin 2.58.0: --exec-id=741-87ad913e --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db 2026-03-19 03:28:35.084 P00 DEBUG: common/lock::lockInit: (path: {"/tmp/pgbackrest"}, execId: {"741-87ad913e"}) 2026-03-19 03:28:35.084 P00 DEBUG: storage/posix/storage::storagePosixNew: (path: {"/tmp/pgbackrest"}, param.modeFile: 0000, param.modePath: 0000, param.write: true, param.noSymLink: false, param.pathExpressionFunction: null) 2026-03-19 03:28:35.084 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: (type: posix, path: {"/tmp/pgbackrest"}, modeFile: 0640, modePath: 0750, write: true, pathExpressionFunction: null, pathSync: true, symLink: true) 2026-03-19 03:28:35.084 P00 DEBUG: storage/posix/storage::storagePosixNewInternal: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:28:35.084 P00 DEBUG: storage/posix/storage::storagePosixNew: => {type: posix, path: /tmp/pgbackrest, write: true} 2026-03-19 03:28:35.084 P00 DEBUG: common/lock::lockInit: => void 2026-03-19 03:28:35.084 P00 DEBUG: command/lock::cmdLockAcquire: (param.returnOnNoLock: false) 2026-03-19 03:28:35.084 P00 DEBUG: command/exit::exitSafe: (result: 0, error: true, signalType: 0) 2026-03-19 03:28:35.084 P00 ERROR: [050]: unable to acquire lock on file '/tmp/pgbackrest/db-backup-1.lock': Resource temporarily unavailable HINT: is another pgBackRest process running? -------------------------------------------------------------------- If SUBMITTING AN ISSUE please provide the following information: version: 2.58.0 command: stanza-upgrade options: --exec-id=741-87ad913e --log-level-console=debug --log-path=/pgdata/pgbackrest/log --pg1-path=/pgdata/pg18 --pg1-port=5432 --pg1-socket-path=/tmp/postgres --repo3-azure-account= --repo3-azure-container=pg-operator-testing --repo3-azure-key= --repo1-path=/backrestrepo/postgres-operator/22413/repo1 --repo3-path=/backrestrepo/postgres-operator/22413/repo3 --repo1-s3-bucket=pg-operator-testing --repo1-s3-endpoint=s3.amazonaws.com --repo1-s3-key= --repo1-s3-key-secret= --repo1-s3-region=us-east-1 --repo1-type=s3 --repo3-type=azure --stanza=db stack trace: common/lock.c:lockAcquire:254:(trace log level required for parameters) command/lock.c:cmdLockAcquire:(param.returnOnNoLock: false) config/load.c:cfgLoad:(debug log level required for parameters) main.c:main:(debug log level required for parameters) -------------------------------------------------------------------- 2026-03-19 03:28:35.084 P00 INFO: stanza-upgrade command end: aborted with exception [050] 2026-03-19 03:28:35.084 P00 DEBUG: command/exit::exitSafe: => 50 2026-03-19 03:28:35.084 P00 DEBUG: main::main: => 50 : command terminated with exit code 50 postgrescluster-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:28:50 +0000 UTC Normal Job.batch demand-backup-backup-dn8b Completed Job completed job-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:29:03 +0000 UTC Normal Pod demand-backup-backup-877f-qqcsq Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-backup-877f-qqcsq to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-c949 default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:29:03 +0000 UTC Normal Job.batch demand-backup-backup-877f SuccessfulCreate Created pod: demand-backup-backup-877f-qqcsq job-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:29:04 +0000 UTC Normal Pod demand-backup-backup-877f-qqcsq.spec.initContainers{pgbackrest-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:29:04 +0000 UTC Normal Pod demand-backup-backup-877f-qqcsq.spec.initContainers{pgbackrest-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" in 230ms (230ms including waiting). Image size: 88441640 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:29:04 +0000 UTC Normal Pod demand-backup-backup-877f-qqcsq.spec.initContainers{pgbackrest-init} Created Created container: pgbackrest-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:29:04 +0000 UTC Normal Pod demand-backup-backup-877f-qqcsq.spec.initContainers{pgbackrest-init} Started Started container pgbackrest-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:29:05 +0000 UTC Normal Pod demand-backup-backup-877f-qqcsq.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:29:05 +0000 UTC Normal Pod demand-backup-backup-877f-qqcsq.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 179ms (179ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:29:05 +0000 UTC Normal Pod demand-backup-backup-877f-qqcsq.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:29:05 +0000 UTC Normal Pod demand-backup-backup-877f-qqcsq.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:32:37 +0000 UTC Normal Job.batch demand-backup-backup-877f Completed Job completed job-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:33:04 +0000 UTC Normal Pod demand-backup-backup-7z25-wjfxh Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-backup-7z25-wjfxh to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-c949 default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:33:04 +0000 UTC Normal Pod demand-backup-backup-7z25-wjfxh.spec.initContainers{pgbackrest-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:33:04 +0000 UTC Normal Pod demand-backup-backup-7z25-wjfxh.spec.initContainers{pgbackrest-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" in 233ms (233ms including waiting). Image size: 88441640 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:33:04 +0000 UTC Normal Pod demand-backup-backup-7z25-wjfxh.spec.initContainers{pgbackrest-init} Created Created container: pgbackrest-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:33:04 +0000 UTC Normal Job.batch demand-backup-backup-7z25 SuccessfulCreate Created pod: demand-backup-backup-7z25-wjfxh job-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:33:05 +0000 UTC Normal Pod demand-backup-backup-7z25-wjfxh.spec.initContainers{pgbackrest-init} Started Started container pgbackrest-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:33:05 +0000 UTC Normal Pod demand-backup-backup-7z25-wjfxh.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:33:06 +0000 UTC Normal Pod demand-backup-backup-7z25-wjfxh.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 200ms (200ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:33:06 +0000 UTC Normal Pod demand-backup-backup-7z25-wjfxh.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:33:06 +0000 UTC Normal Pod demand-backup-backup-7z25-wjfxh.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:34:08 +0000 UTC Normal Job.batch demand-backup-backup-7z25 Completed Job completed job-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:34:38 +0000 UTC Normal Pod demand-backup-backup-bvnk-wwfcq Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-backup-bvnk-wwfcq to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-c949 default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:34:38 +0000 UTC Normal Pod demand-backup-backup-bvnk-wwfcq.spec.initContainers{pgbackrest-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:34:38 +0000 UTC Normal Job.batch demand-backup-backup-bvnk SuccessfulCreate Created pod: demand-backup-backup-bvnk-wwfcq job-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:34:39 +0000 UTC Normal Pod demand-backup-backup-bvnk-wwfcq.spec.initContainers{pgbackrest-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" in 223ms (223ms including waiting). Image size: 88441640 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:34:39 +0000 UTC Normal Pod demand-backup-backup-bvnk-wwfcq.spec.initContainers{pgbackrest-init} Created Created container: pgbackrest-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:34:39 +0000 UTC Normal Pod demand-backup-backup-bvnk-wwfcq.spec.initContainers{pgbackrest-init} Started Started container pgbackrest-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:34:40 +0000 UTC Normal Pod demand-backup-backup-bvnk-wwfcq.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:34:40 +0000 UTC Normal Pod demand-backup-backup-bvnk-wwfcq.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 221ms (221ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:34:40 +0000 UTC Normal Pod demand-backup-backup-bvnk-wwfcq.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:34:40 +0000 UTC Normal Pod demand-backup-backup-bvnk-wwfcq.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:36:10 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{database} Killing Stopping container database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:36:10 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-hhbk5.spec.containers{pgbouncer} Killing Stopping container pgbouncer kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:36:10 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-hhbk5.spec.containers{pgbouncer-config} Killing Stopping container pgbouncer-config kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:36:10 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-jcd5d Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-pgbouncer-7cd7d68587-jcd5d to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-c949 default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:36:10 +0000 UTC Normal ReplicaSet.apps demand-backup-pgbouncer-7cd7d68587 SuccessfulCreate Created pod: demand-backup-pgbouncer-7cd7d68587-jcd5d replicaset-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:36:11 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0 Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-instance1-5kbg-0 to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-c949 default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:36:11 +0000 UTC Warning Pod demand-backup-instance1-5kbg-0 FailedAttachVolume Multi-Attach error for volume "pvc-cab77847-a8e2-4fc9-ae6f-7295ae877ee8" Volume is already exclusively attached to one node and can't be attached to another attachdetach-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:36:11 +0000 UTC Normal StatefulSet.apps demand-backup-instance1-5kbg SuccessfulDelete delete Pod demand-backup-instance1-5kbg-0 in StatefulSet demand-backup-instance1-5kbg successful statefulset-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:36:11 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-jcd5d.spec.containers{pgbouncer} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:36:11 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-jcd5d.spec.containers{pgbouncer} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18" in 218ms (218ms including waiting). Image size: 85546945 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:36:11 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-jcd5d.spec.containers{pgbouncer} Created Created container: pgbouncer kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:36:11 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-jcd5d.spec.containers{pgbouncer} Started Started container pgbouncer kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:36:11 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-jcd5d.spec.containers{pgbouncer-config} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:36:11 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-jcd5d.spec.containers{pgbouncer-config} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18" in 208ms (208ms including waiting). Image size: 85546945 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:36:11 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-jcd5d.spec.containers{pgbouncer-config} Created Created container: pgbouncer-config kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:36:12 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-jcd5d.spec.containers{pgbouncer-config} Started Started container pgbouncer-config kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:36:33 +0000 UTC Normal Pod demand-backup-backup-bvnk-wwfcq.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 209ms (209ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:02 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-cab77847-a8e2-4fc9-ae6f-7295ae877ee8" attachdetach-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:03 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{database-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:03 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{database-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" in 269ms (269ms including waiting). Image size: 88441640 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:03 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{database-init} Created Created container: database-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:03 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{database-init} Started Started container database-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:04 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{postgres-startup} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:04 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{postgres-startup} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 201ms (201ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:04 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{postgres-startup} Created Created container: postgres-startup kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:04 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{postgres-startup} Started Started container postgres-startup kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:05 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:05 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 204ms (205ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:05 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:05 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:06 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{database} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:06 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{database} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 252ms (252ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:06 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{database} Created Created container: database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:06 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{database} Started Started container database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:07 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{replication-cert-copy} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:07 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{replication-cert-copy} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 234ms (234ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:07 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{replication-cert-copy} Created Created container: replication-cert-copy kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:07 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{replication-cert-copy} Started Started container replication-cert-copy kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:07 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:07 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 247ms (247ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:07 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:07 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:07 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{pgbackrest-config} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:09 +0000 UTC Warning Pod demand-backup-instance1-5kbg-0.spec.containers{database} Unhealthy Readiness probe failed: kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:14 +0000 UTC Warning Pod demand-backup-backup-bvnk-wwfcq.spec.containers{pgbackrest} BackOff Back-off restarting failed container pgbackrest in pod demand-backup-backup-bvnk-wwfcq_kuttl-test-growing-killdeer(0553e024-1085-4c4f-8d1d-25d47c838269) kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:25 +0000 UTC Normal Pod demand-backup-backup-bvnk-wwfcq.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 217ms (217ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:37:30 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-hhbk5 TaintManagerEviction Cancelling deletion of Pod kuttl-test-growing-killdeer/demand-backup-pgbouncer-7cd7d68587-hhbk5 taint-eviction-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:40:37 +0000 UTC Normal Job.batch demand-backup-backup-bvnk Completed Job completed job-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:41:03 +0000 UTC Normal Pod demand-backup-backup-2t7s-7zlp7 Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-backup-2t7s-7zlp7 to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-dsht default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:41:03 +0000 UTC Normal Pod demand-backup-backup-2t7s-7zlp7.spec.initContainers{pgbackrest-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:41:03 +0000 UTC Normal Job.batch demand-backup-backup-2t7s SuccessfulCreate Created pod: demand-backup-backup-2t7s-7zlp7 job-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:41:07 +0000 UTC Normal Pod demand-backup-backup-2t7s-7zlp7.spec.initContainers{pgbackrest-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" in 3.97s (3.97s including waiting). Image size: 88441640 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:41:07 +0000 UTC Normal Pod demand-backup-backup-2t7s-7zlp7.spec.initContainers{pgbackrest-init} Created Created container: pgbackrest-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:41:07 +0000 UTC Normal Pod demand-backup-backup-2t7s-7zlp7.spec.initContainers{pgbackrest-init} Started Started container pgbackrest-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:41:10 +0000 UTC Normal Pod demand-backup-backup-2t7s-7zlp7.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:41:18 +0000 UTC Normal Pod demand-backup-backup-2t7s-7zlp7.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 8.044s (8.044s including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:41:18 +0000 UTC Normal Pod demand-backup-backup-2t7s-7zlp7.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:41:18 +0000 UTC Normal Pod demand-backup-backup-2t7s-7zlp7.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:42:22 +0000 UTC Normal Job.batch demand-backup-backup-2t7s Completed Job completed job-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:42:48 +0000 UTC Normal Pod demand-backup-backup-7jpq-5w2b6 Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-backup-7jpq-5w2b6 to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-dsht default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:42:48 +0000 UTC Normal Pod demand-backup-backup-7jpq-5w2b6.spec.initContainers{pgbackrest-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:42:48 +0000 UTC Normal Pod demand-backup-backup-7jpq-5w2b6.spec.initContainers{pgbackrest-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" in 212ms (212ms including waiting). Image size: 88441640 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:42:48 +0000 UTC Normal Pod demand-backup-backup-7jpq-5w2b6.spec.initContainers{pgbackrest-init} Created Created container: pgbackrest-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:42:48 +0000 UTC Normal Pod demand-backup-backup-7jpq-5w2b6.spec.initContainers{pgbackrest-init} Started Started container pgbackrest-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:42:48 +0000 UTC Normal Job.batch demand-backup-backup-7jpq SuccessfulCreate Created pod: demand-backup-backup-7jpq-5w2b6 job-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:42:50 +0000 UTC Normal Pod demand-backup-backup-7jpq-5w2b6.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:42:50 +0000 UTC Normal Pod demand-backup-backup-7jpq-5w2b6.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 197ms (197ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:42:50 +0000 UTC Normal Pod demand-backup-backup-7jpq-5w2b6.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:42:50 +0000 UTC Normal Pod demand-backup-backup-7jpq-5w2b6.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:46 +0000 UTC Normal Job.batch demand-backup-backup-7jpq Completed Job completed job-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:48 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{database} Killing Stopping container database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:48 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{pgbackrest-config} Killing Stopping container pgbackrest-config kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:48 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{pgbackrest} Killing Stopping container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:48 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-2jlxw.spec.containers{pgbouncer} Killing Stopping container pgbouncer kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:48 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-2jlxw.spec.containers{pgbouncer-config} Killing Stopping container pgbouncer-config kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:48 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.containers{pgbackrest} Killing Stopping container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:48 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.containers{pgbackrest-config} Killing Stopping container pgbackrest-config kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:48 +0000 UTC Normal Pod pg-client-65d98588cc-8jjx9.spec.containers{pg-client} Killing Stopping container pg-client kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:49 +0000 UTC Normal StatefulSet.apps demand-backup-instance1-tv29 SuccessfulDelete delete Pod demand-backup-instance1-tv29-0 in StatefulSet demand-backup-instance1-tv29 successful statefulset-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:49 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-nt96m Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-pgbouncer-7cd7d68587-nt96m to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-dsht default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:49 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-nt96m.spec.containers{pgbouncer} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:49 +0000 UTC Normal ReplicaSet.apps demand-backup-pgbouncer-7cd7d68587 SuccessfulCreate Created pod: demand-backup-pgbouncer-7cd7d68587-nt96m replicaset-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:49 +0000 UTC Normal Pod demand-backup-repo-host-0 Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-repo-host-0 to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-dsht default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:49 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:49 +0000 UTC Normal StatefulSet.apps demand-backup-repo-host SuccessfulDelete delete Pod demand-backup-repo-host-0 in StatefulSet demand-backup-repo-host successful statefulset-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:50 +0000 UTC Normal Pod demand-backup-instance1-tv29-0 Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-instance1-tv29-0 to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-dsht default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:50 +0000 UTC Warning Pod demand-backup-instance1-tv29-0 FailedAttachVolume Multi-Attach error for volume "pvc-048c68bd-1efe-4949-be9d-35436d6c9253" Volume is already exclusively attached to one node and can't be attached to another attachdetach-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:50 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 190ms (190ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:50 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:50 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:50 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:50 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 187ms (187ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:50 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:50 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:50 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.containers{pgbackrest-config} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:51 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.containers{pgbackrest-config} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 210ms (210ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:51 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.containers{pgbackrest-config} Created Created container: pgbackrest-config kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:51 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.containers{pgbackrest-config} Started Started container pgbackrest-config kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:53 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-nt96m.spec.containers{pgbouncer} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18" in 3.512s (3.513s including waiting). Image size: 85546945 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:53 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-nt96m.spec.containers{pgbouncer} Created Created container: pgbouncer kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:53 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-nt96m.spec.containers{pgbouncer} Started Started container pgbouncer kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:53 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-nt96m.spec.containers{pgbouncer-config} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:53 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-nt96m.spec.containers{pgbouncer-config} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbouncer18" in 198ms (198ms including waiting). Image size: 85546945 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:53 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-nt96m.spec.containers{pgbouncer-config} Created Created container: pgbouncer-config kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:44:53 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-nt96m.spec.containers{pgbouncer-config} Started Started container pgbouncer-config kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:45:04 +0000 UTC Normal Pod pg-client-65d98588cc-qfkgp Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/pg-client-65d98588cc-qfkgp to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-dsht default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:45:04 +0000 UTC Normal ReplicaSet.apps pg-client-65d98588cc SuccessfulCreate Created pod: pg-client-65d98588cc-qfkgp replicaset-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:45:05 +0000 UTC Normal Pod pg-client-65d98588cc-qfkgp.spec.containers{pg-client} Pulling Pulling image "perconalab/percona-distribution-postgresql:16" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:45:32 +0000 UTC Normal Pod pg-client-65d98588cc-qfkgp.spec.containers{pg-client} Pulled Successfully pulled image "perconalab/percona-distribution-postgresql:16" in 26.082s (27.096s including waiting). Image size: 564259787 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:45:32 +0000 UTC Normal Pod pg-client-65d98588cc-qfkgp.spec.containers{pg-client} Created Created container: pg-client kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:45:32 +0000 UTC Normal Pod pg-client-65d98588cc-qfkgp.spec.containers{pg-client} Started Started container pg-client kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:45:44 +0000 UTC Normal Pod demand-backup-instance1-tv29-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-048c68bd-1efe-4949-be9d-35436d6c9253" attachdetach-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:45:46 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{database-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:45:46 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{database-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" in 221ms (221ms including waiting). Image size: 88441640 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:45:46 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{database-init} Created Created container: database-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:45:46 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{database-init} Started Started container database-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:45:50 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{postgres-startup} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:14 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{postgres-startup} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 23.532s (23.532s including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:14 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{postgres-startup} Created Created container: postgres-startup kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:14 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{postgres-startup} Started Started container postgres-startup kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:19 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-2jlxw TaintManagerEviction Cancelling deletion of Pod kuttl-test-growing-killdeer/demand-backup-pgbouncer-7cd7d68587-2jlxw taint-eviction-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:19 +0000 UTC Normal Pod pg-client-65d98588cc-8jjx9 TaintManagerEviction Cancelling deletion of Pod kuttl-test-growing-killdeer/pg-client-65d98588cc-8jjx9 taint-eviction-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:20 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:22 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 252ms (1.747s including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:22 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:22 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:23 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{database} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:25 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{database} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 244ms (1.207s including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:25 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{database} Created Created container: database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:25 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{database} Started Started container database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:25 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{replication-cert-copy} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:25 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{replication-cert-copy} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 199ms (199ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:25 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{replication-cert-copy} Created Created container: replication-cert-copy kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:25 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{replication-cert-copy} Started Started container replication-cert-copy kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:25 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:25 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 212ms (212ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:25 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:26 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:26 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{pgbackrest-config} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:47 +0000 UTC Normal Pod demand-backup-backup-mvzb-qs7gs Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-backup-mvzb-qs7gs to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-5f1h default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:47 +0000 UTC Normal Job.batch demand-backup-backup-mvzb SuccessfulCreate Created pod: demand-backup-backup-mvzb-qs7gs job-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:48 +0000 UTC Normal Pod demand-backup-backup-mvzb-qs7gs.spec.initContainers{pgbackrest-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:52 +0000 UTC Normal Pod demand-backup-backup-mvzb-qs7gs.spec.initContainers{pgbackrest-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" in 4.051s (4.052s including waiting). Image size: 88441640 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:52 +0000 UTC Normal Pod demand-backup-backup-mvzb-qs7gs.spec.initContainers{pgbackrest-init} Created Created container: pgbackrest-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:52 +0000 UTC Normal Pod demand-backup-backup-mvzb-qs7gs.spec.initContainers{pgbackrest-init} Started Started container pgbackrest-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:46:58 +0000 UTC Normal Pod demand-backup-backup-mvzb-qs7gs.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:47:06 +0000 UTC Normal Pod demand-backup-backup-mvzb-qs7gs.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 8.034s (8.034s including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:47:06 +0000 UTC Normal Pod demand-backup-backup-mvzb-qs7gs.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:47:06 +0000 UTC Normal Pod demand-backup-backup-mvzb-qs7gs.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:48:51 +0000 UTC Normal Job.batch demand-backup-backup-mvzb Completed Job completed job-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:49:19 +0000 UTC Normal Pod demand-backup-backup-xvz6-jphpb Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-backup-xvz6-jphpb to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-5f1h default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:49:19 +0000 UTC Normal Pod demand-backup-backup-xvz6-jphpb.spec.initContainers{pgbackrest-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:49:19 +0000 UTC Normal Job.batch demand-backup-backup-xvz6 SuccessfulCreate Created pod: demand-backup-backup-xvz6-jphpb job-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:49:20 +0000 UTC Normal Pod demand-backup-backup-xvz6-jphpb.spec.initContainers{pgbackrest-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" in 267ms (267ms including waiting). Image size: 88441640 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:49:20 +0000 UTC Normal Pod demand-backup-backup-xvz6-jphpb.spec.initContainers{pgbackrest-init} Created Created container: pgbackrest-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:49:20 +0000 UTC Normal Pod demand-backup-backup-xvz6-jphpb.spec.initContainers{pgbackrest-init} Started Started container pgbackrest-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:49:21 +0000 UTC Normal Pod demand-backup-backup-xvz6-jphpb.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:49:21 +0000 UTC Normal Pod demand-backup-backup-xvz6-jphpb.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 211ms (211ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:49:21 +0000 UTC Normal Pod demand-backup-backup-xvz6-jphpb.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:49:21 +0000 UTC Normal Pod demand-backup-backup-xvz6-jphpb.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:51:21 +0000 UTC Normal Job.batch demand-backup-backup-xvz6 Completed Job completed job-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:51:54 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{database} Killing Stopping container database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:51:54 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{pgbackrest-config} Killing Stopping container pgbackrest-config kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:51:54 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{database} Killing Stopping container database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:51:54 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{pgbackrest-config} Killing Stopping container pgbackrest-config kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:51:54 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{replication-cert-copy} Killing Stopping container replication-cert-copy kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:51:54 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{pgbackrest} Killing Stopping container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:51:54 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{database} Killing Stopping container database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:51:58 +0000 UTC Normal PodDisruptionBudget.policy demand-backup-set-instance1 NoPods No matching pods found controllermanager logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:51:59 +0000 UTC Normal Pod demand-backup-pgbackrest-restore-qxsnr Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-pgbackrest-restore-qxsnr to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-5f1h default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:51:59 +0000 UTC Normal Job.batch demand-backup-pgbackrest-restore SuccessfulCreate Created pod: demand-backup-pgbackrest-restore-qxsnr job-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:52:08 +0000 UTC Normal Pod demand-backup-pgbackrest-restore-qxsnr SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-def14fb6-2481-42b9-9df4-8ed7abdf6da5" attachdetach-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:52:09 +0000 UTC Normal Pod demand-backup-pgbackrest-restore-qxsnr.spec.initContainers{nss-wrapper-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:52:10 +0000 UTC Normal Pod demand-backup-pgbackrest-restore-qxsnr.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 240ms (240ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:52:10 +0000 UTC Normal Pod demand-backup-pgbackrest-restore-qxsnr.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:52:10 +0000 UTC Normal Pod demand-backup-pgbackrest-restore-qxsnr.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:52:10 +0000 UTC Normal Pod demand-backup-pgbackrest-restore-qxsnr.spec.containers{pgbackrest-restore} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:52:33 +0000 UTC Normal Pod demand-backup-pgbackrest-restore-qxsnr.spec.containers{pgbackrest-restore} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 22.388s (22.388s including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:52:33 +0000 UTC Normal Pod demand-backup-pgbackrest-restore-qxsnr.spec.containers{pgbackrest-restore} Created Created container: pgbackrest-restore kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:52:33 +0000 UTC Normal Pod demand-backup-pgbackrest-restore-qxsnr.spec.containers{pgbackrest-restore} Started Started container pgbackrest-restore kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:52:52 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0 Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-instance1-r2k8-0 to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-5f1h default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:52:52 +0000 UTC Normal StatefulSet.apps demand-backup-instance1-r2k8 SuccessfulCreate create Pod demand-backup-instance1-r2k8-0 in StatefulSet demand-backup-instance1-r2k8 successful statefulset-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:52:52 +0000 UTC Normal Job.batch demand-backup-pgbackrest-restore Completed Job completed job-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:52:59 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{database-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:52:59 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{database-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" in 221ms (221ms including waiting). Image size: 88441640 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:52:59 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{database-init} Created Created container: database-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:52:59 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{database-init} Started Started container database-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:00 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{postgres-startup} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:01 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{postgres-startup} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 223ms (223ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:01 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{postgres-startup} Created Created container: postgres-startup kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:01 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{postgres-startup} Started Started container postgres-startup kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:01 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:02 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 235ms (235ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:02 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:02 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:02 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{database} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:03 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{database} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 207ms (207ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:03 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{database} Created Created container: database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:03 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{database} Started Started container database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:03 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{replication-cert-copy} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:03 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{replication-cert-copy} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 221ms (221ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:03 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{replication-cert-copy} Created Created container: replication-cert-copy kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:03 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{replication-cert-copy} Started Started container replication-cert-copy kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:03 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:03 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 206ms (206ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:03 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:04 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:04 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{pgbackrest-config} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:08 +0000 UTC Normal Pod demand-backup-backup-mhbb-7f6ht Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-backup-mhbb-7f6ht to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-5f1h default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:08 +0000 UTC Normal Job.batch demand-backup-backup-mhbb SuccessfulCreate Created pod: demand-backup-backup-mhbb-7f6ht job-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:08 +0000 UTC Normal PostgresCluster.postgres-operator.crunchydata.com demand-backup StanzasCreated pgBackRest stanza creation completed successfully postgrescluster-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:09 +0000 UTC Normal Pod demand-backup-backup-mhbb-7f6ht.spec.initContainers{pgbackrest-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:09 +0000 UTC Normal Pod demand-backup-backup-mhbb-7f6ht.spec.initContainers{pgbackrest-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" in 235ms (235ms including waiting). Image size: 88441640 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:09 +0000 UTC Normal Pod demand-backup-backup-mhbb-7f6ht.spec.initContainers{pgbackrest-init} Created Created container: pgbackrest-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:09 +0000 UTC Normal Pod demand-backup-backup-mhbb-7f6ht.spec.initContainers{pgbackrest-init} Started Started container pgbackrest-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:10 +0000 UTC Normal Pod demand-backup-backup-mhbb-7f6ht.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:11 +0000 UTC Normal Pod demand-backup-backup-mhbb-7f6ht.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 201ms (201ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:11 +0000 UTC Normal Pod demand-backup-backup-mhbb-7f6ht.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:11 +0000 UTC Normal Pod demand-backup-backup-mhbb-7f6ht.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:11 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0 Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-instance1-5kbg-0 to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-c949 default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:11 +0000 UTC Normal StatefulSet.apps demand-backup-instance1-5kbg SuccessfulCreate create Pod demand-backup-instance1-5kbg-0 in StatefulSet demand-backup-instance1-5kbg successful statefulset-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:11 +0000 UTC Normal Pod demand-backup-instance1-tv29-0 Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-instance1-tv29-0 to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-dsht default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:11 +0000 UTC Normal StatefulSet.apps demand-backup-instance1-tv29 SuccessfulCreate create Pod demand-backup-instance1-tv29-0 in StatefulSet demand-backup-instance1-tv29 successful statefulset-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:12 +0000 UTC Normal Pod demand-backup-backup-mhbb-7f6ht.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 218ms (218ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:14 +0000 UTC Warning Pod demand-backup-backup-mhbb-7f6ht.spec.containers{pgbackrest} BackOff Back-off restarting failed container pgbackrest in pod demand-backup-backup-mhbb-7f6ht_kuttl-test-growing-killdeer(58cacd80-df7a-4c23-8a44-0bf00889e341) kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:15 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-cab77847-a8e2-4fc9-ae6f-7295ae877ee8" attachdetach-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:16 +0000 UTC Normal Pod demand-backup-instance1-tv29-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-048c68bd-1efe-4949-be9d-35436d6c9253" attachdetach-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:21 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{database-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:21 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{database-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:22 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{database-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" in 210ms (210ms including waiting). Image size: 88441640 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:22 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{database-init} Created Created container: database-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:22 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{database-init} Started Started container database-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:22 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{database-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" in 214ms (214ms including waiting). Image size: 88441640 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:22 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{database-init} Created Created container: database-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:22 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{database-init} Started Started container database-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:23 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{postgres-startup} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:23 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{postgres-startup} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 222ms (222ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:23 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{postgres-startup} Created Created container: postgres-startup kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:23 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{postgres-startup} Started Started container postgres-startup kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:23 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{postgres-startup} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:23 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{postgres-startup} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 215ms (215ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:23 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{postgres-startup} Created Created container: postgres-startup kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:23 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{postgres-startup} Started Started container postgres-startup kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:24 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:24 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 239ms (239ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:24 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:24 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:24 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:24 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 230ms (230ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:24 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:24 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:25 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{database} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:25 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{database} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 208ms (208ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:25 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{database} Created Created container: database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:25 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{database} Started Started container database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:25 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{replication-cert-copy} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:25 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{database} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:25 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{database} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 212ms (212ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:25 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{database} Created Created container: database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:25 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{database} Started Started container database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:25 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{replication-cert-copy} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:25 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{replication-cert-copy} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 226ms (226ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:25 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{replication-cert-copy} Created Created container: replication-cert-copy kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:26 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{replication-cert-copy} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 212ms (212ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:26 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{replication-cert-copy} Created Created container: replication-cert-copy kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:26 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{replication-cert-copy} Started Started container replication-cert-copy kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:26 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:26 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 188ms (188ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:26 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:26 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:26 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{pgbackrest-config} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:26 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{replication-cert-copy} Started Started container replication-cert-copy kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:26 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:26 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 189ms (189ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:26 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:26 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:26 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{pgbackrest-config} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:27 +0000 UTC Normal Pod demand-backup-backup-mhbb-7f6ht.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 188ms (188ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:28 +0000 UTC Warning Pod demand-backup-instance1-5kbg-0.spec.containers{database} Unhealthy Readiness probe failed: kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:28 +0000 UTC Warning Pod demand-backup-instance1-tv29-0.spec.containers{database} Unhealthy Readiness probe failed: kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:53:53 +0000 UTC Normal Pod demand-backup-backup-mhbb-7f6ht.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 224ms (224ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:54:12 +0000 UTC Normal Job.batch demand-backup-backup-mhbb Completed Job completed job-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:54:34 +0000 UTC Normal Job.batch demand-backup-pgbackrest-restore SuccessfulCreate Created pod: demand-backup-pgbackrest-restore-x8bnx job-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:54:35 +0000 UTC Normal Pod demand-backup-pgbackrest-restore-x8bnx Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-pgbackrest-restore-x8bnx to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-5f1h default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:54:40 +0000 UTC Normal Pod demand-backup-pgbackrest-restore-x8bnx.spec.initContainers{nss-wrapper-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:54:41 +0000 UTC Normal Pod demand-backup-pgbackrest-restore-x8bnx.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 247ms (247ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:54:41 +0000 UTC Normal Pod demand-backup-pgbackrest-restore-x8bnx.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:54:41 +0000 UTC Normal Pod demand-backup-pgbackrest-restore-x8bnx.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:54:42 +0000 UTC Normal Pod demand-backup-pgbackrest-restore-x8bnx.spec.containers{pgbackrest-restore} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:54:42 +0000 UTC Normal Pod demand-backup-pgbackrest-restore-x8bnx.spec.containers{pgbackrest-restore} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 191ms (191ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:54:42 +0000 UTC Normal Pod demand-backup-pgbackrest-restore-x8bnx.spec.containers{pgbackrest-restore} Created Created container: pgbackrest-restore kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:54:42 +0000 UTC Normal Pod demand-backup-pgbackrest-restore-x8bnx.spec.containers{pgbackrest-restore} Started Started container pgbackrest-restore kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:54:56 +0000 UTC Normal Job.batch demand-backup-pgbackrest-restore Completed Job completed job-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:54:57 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0 Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-instance1-r2k8-0 to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-5f1h default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:54:57 +0000 UTC Normal StatefulSet.apps demand-backup-instance1-r2k8 SuccessfulCreate create Pod demand-backup-instance1-r2k8-0 in StatefulSet demand-backup-instance1-r2k8 successful statefulset-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:01 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{database-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:01 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{database-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" in 239ms (239ms including waiting). Image size: 88441640 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:01 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{database-init} Created Created container: database-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:02 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{database-init} Started Started container database-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:03 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{postgres-startup} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:03 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{postgres-startup} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 225ms (225ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:03 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{postgres-startup} Created Created container: postgres-startup kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:03 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{postgres-startup} Started Started container postgres-startup kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:04 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:04 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 177ms (177ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:04 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:04 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:05 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{database} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:05 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{database} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 237ms (237ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:05 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{database} Created Created container: database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:05 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{database} Started Started container database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:05 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{replication-cert-copy} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:05 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{replication-cert-copy} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 215ms (215ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:05 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{replication-cert-copy} Created Created container: replication-cert-copy kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:06 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{replication-cert-copy} Started Started container replication-cert-copy kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:06 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:06 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 191ms (191ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:06 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:06 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:06 +0000 UTC Normal Pod demand-backup-instance1-r2k8-0.spec.containers{pgbackrest-config} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:08 +0000 UTC Warning Pod demand-backup-instance1-r2k8-0.spec.containers{database} Unhealthy Readiness probe failed: kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:13 +0000 UTC Normal Pod demand-backup-instance1-tv29-0 Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-instance1-tv29-0 to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-c949 default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:13 +0000 UTC Normal StatefulSet.apps demand-backup-instance1-tv29 SuccessfulCreate create Pod demand-backup-instance1-tv29-0 in StatefulSet demand-backup-instance1-tv29 successful statefulset-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:14 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0 Binding Scheduled Successfully assigned kuttl-test-growing-killdeer/demand-backup-instance1-5kbg-0 to gke-jen-pg-1506-e8248869-default-pool-9fe0e76e-dsht default-scheduler logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:14 +0000 UTC Normal StatefulSet.apps demand-backup-instance1-5kbg SuccessfulCreate create Pod demand-backup-instance1-5kbg-0 in StatefulSet demand-backup-instance1-5kbg successful statefulset-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:18 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-cab77847-a8e2-4fc9-ae6f-7295ae877ee8" attachdetach-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:18 +0000 UTC Normal Pod demand-backup-instance1-tv29-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-048c68bd-1efe-4949-be9d-35436d6c9253" attachdetach-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:24 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{database-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:24 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{database-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" in 293ms (293ms including waiting). Image size: 88441640 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:24 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{database-init} Created Created container: database-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:24 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{database-init} Started Started container database-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:25 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{database-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:25 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{database-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:PR-1506-e8248869c" in 236ms (236ms including waiting). Image size: 88441640 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:25 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{database-init} Created Created container: database-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:25 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{database-init} Started Started container database-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:25 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{postgres-startup} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:26 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{postgres-startup} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:26 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{postgres-startup} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 195ms (195ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:26 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{postgres-startup} Created Created container: postgres-startup kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:26 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{postgres-startup} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 205ms (205ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:26 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{postgres-startup} Created Created container: postgres-startup kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:26 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{postgres-startup} Started Started container postgres-startup kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:26 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:27 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{postgres-startup} Started Started container postgres-startup kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:27 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:27 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 245ms (245ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:27 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:27 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 232ms (232ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:27 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:27 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:27 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{database} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:28 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:28 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{database} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:28 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{database} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 219ms (219ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:28 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{database} Created Created container: database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:28 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{database} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 227ms (227ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:28 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{database} Created Created container: database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:28 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{database} Started Started container database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:28 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{replication-cert-copy} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:28 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{replication-cert-copy} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 265ms (265ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:28 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{replication-cert-copy} Created Created container: replication-cert-copy kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:28 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{replication-cert-copy} Started Started container replication-cert-copy kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:28 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:28 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 211ms (211ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:28 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:29 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{database} Started Started container database kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:29 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{replication-cert-copy} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:29 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{replication-cert-copy} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-ppg18-postgres" in 228ms (228ms including waiting). Image size: 561946077 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:29 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{replication-cert-copy} Created Created container: replication-cert-copy kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:29 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{replication-cert-copy} Started Started container replication-cert-copy kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:29 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{pgbackrest} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:29 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{pgbackrest} Pulled Successfully pulled image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" in 239ms (239ms including waiting). Image size: 162507906 bytes. kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:29 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:29 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:29 +0000 UTC Normal Pod demand-backup-instance1-5kbg-0.spec.containers{pgbackrest-config} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:29 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:55:29 +0000 UTC Normal Pod demand-backup-instance1-tv29-0.spec.containers{pgbackrest-config} Pulling Pulling image "docker.io/perconalab/percona-postgresql-operator:main-pgbackrest18" kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:57:40 +0000 UTC Normal StatefulSet.apps demand-backup-instance1-5kbg SuccessfulDelete delete Pod demand-backup-instance1-5kbg-0 in StatefulSet demand-backup-instance1-5kbg successful statefulset-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:57:40 +0000 UTC Normal StatefulSet.apps demand-backup-instance1-tv29 SuccessfulDelete delete Pod demand-backup-instance1-tv29-0 in StatefulSet demand-backup-instance1-tv29 successful statefulset-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:57:43 +0000 UTC Normal StatefulSet.apps demand-backup-instance1-r2k8 SuccessfulDelete delete Pod demand-backup-instance1-r2k8-0 in StatefulSet demand-backup-instance1-r2k8 successful statefulset-controller logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:57:47 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-jcd5d.spec.containers{pgbouncer} Killing Stopping container pgbouncer kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:57:47 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-jcd5d.spec.containers{pgbouncer-config} Killing Stopping container pgbouncer-config kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:57:47 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-nt96m.spec.containers{pgbouncer} Killing Stopping container pgbouncer kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:57:47 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-nt96m.spec.containers{pgbouncer-config} Killing Stopping container pgbouncer-config kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:57:47 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-qmqwt.spec.containers{pgbouncer} Killing Stopping container pgbouncer kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:57:47 +0000 UTC Normal Pod demand-backup-pgbouncer-7cd7d68587-qmqwt.spec.containers{pgbouncer-config} Killing Stopping container pgbouncer-config kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:57:47 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.containers{pgbackrest} Killing Stopping container pgbackrest kubelet logger.go:42: 03:58:16 | demand-backup | 2026-03-19 03:57:47 +0000 UTC Normal Pod demand-backup-repo-host-0.spec.containers{pgbackrest-config} Killing Stopping container pgbackrest-config kubelet logger.go:42: 03:58:17 | demand-backup | Deleting namespace "kuttl-test-growing-killdeer" === NAME kuttl harness.go:404: run tests finished harness.go:511: cleaning up harness.go:568: removing temp folder: "" --- PASS: kuttl (1965.82s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/demand-backup (1965.03s) PASS