=== RUN   kuttl
    harness.go:464: starting setup
    harness.go:255: running tests using configured kubeconfig.
    harness.go:278: Successful connection to cluster at: https://34.170.147.61
    harness.go:363: running tests
    harness.go:75: going to run test suite with timeout of 180 seconds for each step
    harness.go:375: testsuite: e2e-tests/tests has 22 tests
=== RUN   kuttl/harness
=== RUN   kuttl/harness/start-from-backup
=== PAUSE kuttl/harness/start-from-backup
=== CONT  kuttl/harness/start-from-backup
    logger.go:42: 15:45:33 | start-from-backup | Creating namespace: kuttl-test-resolved-akita
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | starting test step 0-deploy-operator
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        init_temp_dir # do this only in the first TestStep
        
        deploy_operator
        deploy_client
        deploy_s3_secrets]
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | + source ../../functions
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ realpath ../../..
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | ++ CERT_MANAGER_VER=1.15.3
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | ++++ pwd
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1021/e2e-tests/tests/start-from-backup
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | ++ test_name=start-from-backup
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1021/e2e-tests/vars.sh
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/deploy
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/deploy
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/e2e-tests
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/e2e-tests
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/e2e-tests/conf
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/e2e-tests/conf
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ export TEMP_DIR=/tmp/kuttl/pg/start-from-backup
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ TEMP_DIR=/tmp/kuttl/pg/start-from-backup
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ export GIT_BRANCH=PR-1021
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ GIT_BRANCH=PR-1021
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ export VERSION=PR-1021-eed29e68b
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ VERSION=PR-1021-eed29e68b
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ IMAGE_BASE=perconalab/percona-postgresql-operator
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1021-eed29e68b
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1021-eed29e68b
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ export PG_VER=17
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ PG_VER=17
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-ppg17-pgbouncer
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-ppg17-pgbouncer
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ export IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg17-postgres
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg17-postgres
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-ppg17-pgbackrest
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-ppg17-pgbackrest
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ export BUCKET=pg-operator-testing
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ BUCKET=pg-operator-testing
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ export PMM_SERVER_VERSION=9.9.9
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ PMM_SERVER_VERSION=9.9.9
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ export PGOV1_TAG=1.4.0
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ PGOV1_TAG=1.4.0
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ export PGOV1_VER=14
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ PGOV1_VER=14
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | ++++ which gdate
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | ++++ which date
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ date=/usr/bin/date
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | ++++ which gsed
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | ++++ which sed
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ sed=/usr/bin/sed
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | +++ command -v oc
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | ++ oc get projects
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | + init_temp_dir
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | + rm -rf /tmp/kuttl/pg/start-from-backup
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | + mkdir -p /tmp/kuttl/pg/start-from-backup
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | + deploy_operator
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | + local cw_prefix=
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | + destroy_operator
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | + kubectl -n pg-operator delete deployment percona-postgresql-operator --force --grace-period=0
    logger.go:42: 15:45:33 | start-from-backup/0-deploy-operator | Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.
    logger.go:42: 15:45:34 | start-from-backup/0-deploy-operator | Error from server (NotFound): deployments.apps "percona-postgresql-operator" not found
    logger.go:42: 15:45:34 | start-from-backup/0-deploy-operator | + true
    logger.go:42: 15:45:34 | start-from-backup/0-deploy-operator | + [[ -n pg-operator ]]
    logger.go:42: 15:45:34 | start-from-backup/0-deploy-operator | + kubectl delete namespace pg-operator --force --grace-period=0
    logger.go:42: 15:45:34 | start-from-backup/0-deploy-operator | Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.
    logger.go:42: 15:45:34 | start-from-backup/0-deploy-operator | Error from server (NotFound): namespaces "pg-operator" not found
    logger.go:42: 15:45:34 | start-from-backup/0-deploy-operator | + true
    logger.go:42: 15:45:34 | start-from-backup/0-deploy-operator | + [[ -n pg-operator ]]
    logger.go:42: 15:45:34 | start-from-backup/0-deploy-operator | + create_namespace pg-operator
    logger.go:42: 15:45:34 | start-from-backup/0-deploy-operator | + local namespace=pg-operator
    logger.go:42: 15:45:34 | start-from-backup/0-deploy-operator | + [[ -n '' ]]
    logger.go:42: 15:45:34 | start-from-backup/0-deploy-operator | + kubectl delete namespace pg-operator --ignore-not-found
    logger.go:42: 15:45:35 | start-from-backup/0-deploy-operator | + kubectl wait --for=delete namespace pg-operator
    logger.go:42: 15:45:35 | start-from-backup/0-deploy-operator | + kubectl create namespace pg-operator
    logger.go:42: 15:45:35 | start-from-backup/0-deploy-operator | namespace/pg-operator created
    logger.go:42: 15:45:35 | start-from-backup/0-deploy-operator | + cw_prefix=cw-
    logger.go:42: 15:45:35 | start-from-backup/0-deploy-operator | + kubectl -n pg-operator apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pg-operator_PR-1021/deploy/crd.yaml
    logger.go:42: 15:45:36 | start-from-backup/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/crunchybridgeclusters.postgres-operator.crunchydata.com serverside-applied
    logger.go:42: 15:45:37 | start-from-backup/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconapgbackups.pgv2.percona.com serverside-applied
    logger.go:42: 15:45:39 | start-from-backup/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconapgclusters.pgv2.percona.com serverside-applied
    logger.go:42: 15:45:39 | start-from-backup/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconapgrestores.pgv2.percona.com serverside-applied
    logger.go:42: 15:45:39 | start-from-backup/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconapgupgrades.pgv2.percona.com serverside-applied
    logger.go:42: 15:45:39 | start-from-backup/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/pgadmins.postgres-operator.crunchydata.com serverside-applied
    logger.go:42: 15:45:40 | start-from-backup/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/pgupgrades.postgres-operator.crunchydata.com serverside-applied
    logger.go:42: 15:45:41 | start-from-backup/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/postgresclusters.postgres-operator.crunchydata.com serverside-applied
    logger.go:42: 15:45:41 | start-from-backup/0-deploy-operator | + kubectl -n pg-operator apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pg-operator_PR-1021/deploy/cw-rbac.yaml
    logger.go:42: 15:45:42 | start-from-backup/0-deploy-operator | serviceaccount/percona-postgresql-operator serverside-applied
    logger.go:42: 15:45:42 | start-from-backup/0-deploy-operator | clusterrole.rbac.authorization.k8s.io/percona-postgresql-operator serverside-applied
    logger.go:42: 15:45:42 | start-from-backup/0-deploy-operator | clusterrolebinding.rbac.authorization.k8s.io/percona-postgresql-operator serverside-applied
    logger.go:42: 15:45:42 | start-from-backup/0-deploy-operator | + local disable_telemetry=true
    logger.go:42: 15:45:42 | start-from-backup/0-deploy-operator | + '[' start-from-backup == telemetry-transfer ']'
    logger.go:42: 15:45:42 | start-from-backup/0-deploy-operator | + yq eval '.spec.template.spec.containers[0].image = "perconalab/percona-postgresql-operator:PR-1021-eed29e68b"' /mnt/jenkins/workspace/cloud-pg-operator_PR-1021/deploy/cw-operator.yaml
    logger.go:42: 15:45:42 | start-from-backup/0-deploy-operator | + yq eval '(.spec.template.spec.containers[] | select(.name=="operator") | .env[] | select(.name=="DISABLE_TELEMETRY") | .value) = "true"' -
    logger.go:42: 15:45:42 | start-from-backup/0-deploy-operator | + kubectl -n pg-operator apply -f -
    logger.go:42: 15:45:43 | start-from-backup/0-deploy-operator | deployment.apps/percona-postgresql-operator created
    logger.go:42: 15:45:43 | start-from-backup/0-deploy-operator | + deploy_client
    logger.go:42: 15:45:43 | start-from-backup/0-deploy-operator | + kubectl -n kuttl-test-resolved-akita apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-1021/e2e-tests/conf/client.yaml
    logger.go:42: 15:45:44 | start-from-backup/0-deploy-operator | deployment.apps/pg-client created
    logger.go:42: 15:45:44 | start-from-backup/0-deploy-operator | + deploy_s3_secrets
    logger.go:42: 15:45:44 | start-from-backup/0-deploy-operator | + set +o xtrace
    logger.go:42: 15:45:44 | start-from-backup/0-deploy-operator | secret/start-from-backup-pgbackrest-secrets created
    logger.go:42: 15:45:45 | start-from-backup/0-deploy-operator | running command: [sh -c kubectl assert exist-enhanced deployment percona-postgresql-operator -n ${OPERATOR_NS:-$NAMESPACE} --field-selector status.readyReplicas=1]
    logger.go:42: 15:45:45 | start-from-backup/0-deploy-operator | ASSERT deployment percona-postgresql-operator matching field criteria 'status.readyReplicas=1' should exist.
    logger.go:42: 15:45:45 | start-from-backup/0-deploy-operator | INFO   Found 1 resource(s).
    logger.go:42: 15:45:45 | start-from-backup/0-deploy-operator | NAME                          NAMESPACE     COL0
    logger.go:42: 15:45:45 | start-from-backup/0-deploy-operator | percona-postgresql-operator   pg-operator   1
    logger.go:42: 15:45:45 | start-from-backup/0-deploy-operator | ASSERT PASS
    logger.go:42: 15:45:45 | start-from-backup/0-deploy-operator | test step completed 0-deploy-operator
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | starting test step 1-create-cluster
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        get_cr "start-from-backup" ${RANDOM}| kubectl -n "${NAMESPACE}" apply -f -]
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | + source ../../functions
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ realpath ../../..
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | ++ CERT_MANAGER_VER=1.15.3
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | ++++ pwd
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1021/e2e-tests/tests/start-from-backup
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | ++ test_name=start-from-backup
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1021/e2e-tests/vars.sh
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/deploy
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/deploy
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/e2e-tests
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/e2e-tests
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/e2e-tests/conf
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/e2e-tests/conf
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ export TEMP_DIR=/tmp/kuttl/pg/start-from-backup
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ TEMP_DIR=/tmp/kuttl/pg/start-from-backup
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ export GIT_BRANCH=PR-1021
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ GIT_BRANCH=PR-1021
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ export VERSION=PR-1021-eed29e68b
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ VERSION=PR-1021-eed29e68b
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ IMAGE_BASE=perconalab/percona-postgresql-operator
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1021-eed29e68b
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1021-eed29e68b
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ export PG_VER=17
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ PG_VER=17
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-ppg17-pgbouncer
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-ppg17-pgbouncer
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ export IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg17-postgres
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg17-postgres
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-ppg17-pgbackrest
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-ppg17-pgbackrest
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ export BUCKET=pg-operator-testing
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ BUCKET=pg-operator-testing
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ export PMM_SERVER_VERSION=9.9.9
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ PMM_SERVER_VERSION=9.9.9
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ export PGOV1_TAG=1.4.0
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ PGOV1_TAG=1.4.0
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ export PGOV1_VER=14
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ PGOV1_VER=14
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | ++++ which gdate
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | ++++ which date
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ date=/usr/bin/date
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | ++++ which gsed
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | ++++ which sed
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ sed=/usr/bin/sed
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | +++ command -v oc
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | ++ oc get projects
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | + get_cr start-from-backup 5908
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | + local cr_name=start-from-backup
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | + kubectl -n kuttl-test-resolved-akita apply -f -
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | + '[' -z start-from-backup ']'
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | + local repo_path=5908
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | + yq eval '
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | 		.metadata.name = "start-from-backup" |
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | 		.metadata.labels = {"e2e":"start-from-backup"} |
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | 		.spec.postgresVersion = 17 |
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | 		.spec.users += [{"name":"postgres","password":{"type":"AlphaNumeric"}}] |
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | 		.spec.users += [{"name":"start-from-backup","password":{"type":"AlphaNumeric"}}] |
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | 		.spec.image = "perconalab/percona-postgresql-operator:main-ppg17-postgres" |
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | 		.spec.backups.pgbackrest.image = "perconalab/percona-postgresql-operator:main-ppg17-pgbackrest" |
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | 		.spec.proxy.pgBouncer.image = "perconalab/percona-postgresql-operator:main-ppg17-pgbouncer" |
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | 		.spec.pmm.image = "perconalab/pmm-client:dev-latest" |
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | 		.spec.pmm.secret = "start-from-backup-pmm-secret"
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | 		' /mnt/jenkins/workspace/cloud-pg-operator_PR-1021/deploy/cr.yaml
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | + [[ -n '' ]]
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | + case $test_name in
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | + yq eval -i '
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | 				.spec.backups.pgbackrest.configuration = [{"secret":{"name":"start-from-backup-pgbackrest-secrets"}}] |
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | 				.spec.backups.pgbackrest.manual.repoName = "repo1" |
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | 				.spec.backups.pgbackrest.manual.options = ["--type=full"] |
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | 				.spec.backups.pgbackrest.global.repo1-path = "/backrestrepo/postgres-operator/5908/repo1" |
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | 				.spec.backups.pgbackrest.repos = [{"name":"repo1","s3":{"bucket":"pg-operator-testing","endpoint":"s3.amazonaws.com","region":"us-east-1"}}]
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | 				' /tmp/kuttl/pg/start-from-backup/cr.yaml
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | + [[ start-from-backup == \d\e\m\a\n\d\-\b\a\c\k\u\p ]]
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | + [[ start-from-backup == \s\t\a\r\t\-\f\r\o\m\-\b\a\c\k\u\p ]]
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | + yq eval -i '
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | 					.spec.dataSource.pgbackrest.configuration = [{"secret":{"name":"start-from-backup-pgbackrest-secrets"}}] |
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | 					.spec.dataSource.pgbackrest.stanza = "db" |
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | 					.spec.dataSource.pgbackrest.global.repo1-path = "/cluster-source/demand-backup-ppg17/repo1" |
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | 					.spec.dataSource.pgbackrest.repo = {"name":"repo1","s3":{"bucket":"pg-operator-testing","endpoint":"s3.amazonaws.com","region":"us-east-1"}}
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | 					' /tmp/kuttl/pg/start-from-backup/cr.yaml
    logger.go:42: 15:45:45 | start-from-backup/1-create-cluster | + cat /tmp/kuttl/pg/start-from-backup/cr.yaml
    logger.go:42: 15:45:46 | start-from-backup/1-create-cluster | perconapgcluster.pgv2.percona.com/start-from-backup created
    logger.go:42: 15:49:35 | start-from-backup/1-create-cluster | test step completed 1-create-cluster
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | starting test step 2-read-from-primary
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        data=$(run_psql_local '\c myapp \\\ SELECT * from myApp;' "postgres:$(get_psql_user_pass start-from-backup-pguser-postgres)@$(get_psql_user_host start-from-backup-pguser-postgres)")
        
        kubectl create configmap -n "${NAMESPACE}" 02-read-from-primary --from-literal=data="${data}"]
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | + source ../../functions
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ realpath ../../..
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | ++ CERT_MANAGER_VER=1.15.3
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | ++++ pwd
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1021/e2e-tests/tests/start-from-backup
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | ++ test_name=start-from-backup
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1021/e2e-tests/vars.sh
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/deploy
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/deploy
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/e2e-tests
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/e2e-tests
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/e2e-tests/conf
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/e2e-tests/conf
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ export TEMP_DIR=/tmp/kuttl/pg/start-from-backup
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ TEMP_DIR=/tmp/kuttl/pg/start-from-backup
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ export GIT_BRANCH=PR-1021
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ GIT_BRANCH=PR-1021
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ export VERSION=PR-1021-eed29e68b
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ VERSION=PR-1021-eed29e68b
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ IMAGE_BASE=perconalab/percona-postgresql-operator
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1021-eed29e68b
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1021-eed29e68b
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ export PG_VER=17
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ PG_VER=17
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-ppg17-pgbouncer
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-ppg17-pgbouncer
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ export IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg17-postgres
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg17-postgres
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-ppg17-pgbackrest
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-ppg17-pgbackrest
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ export BUCKET=pg-operator-testing
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ BUCKET=pg-operator-testing
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ export PMM_SERVER_VERSION=9.9.9
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ PMM_SERVER_VERSION=9.9.9
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ export PGOV1_TAG=1.4.0
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ PGOV1_TAG=1.4.0
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ export PGOV1_VER=14
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ PGOV1_VER=14
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | ++++ which gdate
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | ++++ which date
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ date=/usr/bin/date
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | ++++ which gsed
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | ++++ which sed
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ sed=/usr/bin/sed
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ command -v oc
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | ++ oc get projects
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ get_psql_user_pass start-from-backup-pguser-postgres
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ local secret_name=start-from-backup-pguser-postgres
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ kubectl -n kuttl-test-resolved-akita get secret/start-from-backup-pguser-postgres '--template={{.data.password | base64decode}}'
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ get_psql_user_host start-from-backup-pguser-postgres
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ local secret_name=start-from-backup-pguser-postgres
    logger.go:42: 15:49:35 | start-from-backup/2-read-from-primary | +++ kubectl -n kuttl-test-resolved-akita get secret/start-from-backup-pguser-postgres '--template={{.data.host | base64decode }}'
    logger.go:42: 15:49:36 | start-from-backup/2-read-from-primary | ++ run_psql_local '\c myapp \\\ SELECT * from myApp;' postgres:GepcB3kZvAhacWmHpBr9EZZ1@start-from-backup-primary.kuttl-test-resolved-akita.svc
    logger.go:42: 15:49:36 | start-from-backup/2-read-from-primary | ++ local 'command=\c myapp \\\ SELECT * from myApp;'
    logger.go:42: 15:49:36 | start-from-backup/2-read-from-primary | ++ local uri=postgres:GepcB3kZvAhacWmHpBr9EZZ1@start-from-backup-primary.kuttl-test-resolved-akita.svc
    logger.go:42: 15:49:36 | start-from-backup/2-read-from-primary | ++ local driver=postgres
    logger.go:42: 15:49:36 | start-from-backup/2-read-from-primary | +++ get_client_pod
    logger.go:42: 15:49:36 | start-from-backup/2-read-from-primary | +++ kubectl -n kuttl-test-resolved-akita get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}'
    logger.go:42: 15:49:36 | start-from-backup/2-read-from-primary | ++ kubectl -n kuttl-test-resolved-akita exec pg-client-6cc584874-xr4zw -- bash -c 'printf '\''\c myapp \\\ SELECT * from myApp;\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://'\''postgres:GepcB3kZvAhacWmHpBr9EZZ1@start-from-backup-primary.kuttl-test-resolved-akita.svc'\'''
    logger.go:42: 15:49:38 | start-from-backup/2-read-from-primary | + data=' 100500
    logger.go:42: 15:49:38 | start-from-backup/2-read-from-primary |  100501'
    logger.go:42: 15:49:38 | start-from-backup/2-read-from-primary | + kubectl create configmap -n kuttl-test-resolved-akita 02-read-from-primary '--from-literal=data= 100500
    logger.go:42: 15:49:38 | start-from-backup/2-read-from-primary |  100501'
    logger.go:42: 15:49:39 | start-from-backup/2-read-from-primary | configmap/02-read-from-primary created
[controller-runtime] log.SetLogger(...) was never called; logs will not be displayed.
Detected at:
	>  goroutine 52 [running]:
	>  runtime/debug.Stack()
	>  	/nix/store/wkbckbd30nlhq4dxzg64q6y4vm1xx4fk-go-1.22.1/share/go/src/runtime/debug/stack.go:24 +0x5e
	>  sigs.k8s.io/controller-runtime/pkg/log.eventuallyFulfillRoot()
	>  	/home/mowsiany/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.3/pkg/log/log.go:60 +0xcd
	>  sigs.k8s.io/controller-runtime/pkg/log.(*delegatingLogSink).WithName(0xc0002a9c00, {0x184a055, 0x14})
	>  	/home/mowsiany/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.3/pkg/log/deleg.go:147 +0x3e
	>  github.com/go-logr/logr.Logger.WithName({{0x1acb7d8, 0xc0002a9c00}, 0x0}, {0x184a055?, 0xc0006abf80?})
	>  	/home/mowsiany/go/pkg/mod/github.com/go-logr/logr@v1.2.4/logr.go:336 +0x36
	>  sigs.k8s.io/controller-runtime/pkg/client.newClient(0x131ead3?, {0x0, 0xc00043c230, {0x1accd90, 0xc0005fe040}, 0x0, {0x0, 0x0}, 0x0})
	>  	/home/mowsiany/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.3/pkg/client/client.go:122 +0xf1
	>  sigs.k8s.io/controller-runtime/pkg/client.New(0xc0005db688?, {0x0, 0xc00043c230, {0x1accd90, 0xc0005fe040}, 0x0, {0x0, 0x0}, 0x0})
	>  	/home/mowsiany/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.3/pkg/client/client.go:103 +0x7d
	>  github.com/kudobuilder/kuttl/pkg/test/utils.NewRetryClient(0xc0005db688, {0x0, 0xc00043c230, {0x1accd90, 0xc0005fe040}, 0x0, {0x0, 0x0}, 0x0})
	>  	/home/mowsiany/go/src/github.com/kudobuilder/kuttl/pkg/test/utils/kubernetes.go:177 +0x127
	>  github.com/kudobuilder/kuttl/pkg/test.(*Harness).Client(0xc000608008, 0x4a?)
	>  	/home/mowsiany/go/src/github.com/kudobuilder/kuttl/pkg/test/harness.go:323 +0x18e
	>  github.com/kudobuilder/kuttl/pkg/test.(*Step).Create(0xc0001c85b0, 0xc0000acd00, {0xc000500840, 0x19})
	>  	/home/mowsiany/go/src/github.com/kudobuilder/kuttl/pkg/test/step.go:177 +0x63
	>  github.com/kudobuilder/kuttl/pkg/test.(*Step).Run(0xc0001c85b0, 0xc0000acd00, {0xc000500840, 0x19})
	>  	/home/mowsiany/go/src/github.com/kudobuilder/kuttl/pkg/test/step.go:457 +0x24a
	>  github.com/kudobuilder/kuttl/pkg/test.(*Case).Run(0xc000603e00, 0xc0000acd00, 0xc0001e2750)
	>  	/home/mowsiany/go/src/github.com/kudobuilder/kuttl/pkg/test/case.go:373 +0xaeb
	>  github.com/kudobuilder/kuttl/pkg/test.(*Harness).RunTests.func1.1(0xc0000acd00)
	>  	/home/mowsiany/go/src/github.com/kudobuilder/kuttl/pkg/test/harness.go:401 +0x12e
	>  testing.tRunner(0xc0000acd00, 0xc000010ee8)
	>  	/nix/store/wkbckbd30nlhq4dxzg64q6y4vm1xx4fk-go-1.22.1/share/go/src/testing/testing.go:1689 +0xfb
	>  created by testing.(*T).Run in goroutine 51
	>  	/nix/store/wkbckbd30nlhq4dxzg64q6y4vm1xx4fk-go-1.22.1/share/go/src/testing/testing.go:1742 +0x390
    logger.go:42: 15:49:39 | start-from-backup/2-read-from-primary | test step completed 2-read-from-primary
    logger.go:42: 15:49:39 | start-from-backup/99-remove-cluster-gracefully | starting test step 99-remove-cluster-gracefully
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | running command: [sh -c set -o errexit
        set -o xtrace
        
        source ../../functions
        
        destroy_operator]
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | + source ../../functions
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ realpath ../../..
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | ++ CERT_MANAGER_VER=1.15.3
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | ++++ pwd
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1021/e2e-tests/tests/start-from-backup
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | ++ test_name=start-from-backup
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1021/e2e-tests/vars.sh
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/deploy
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/deploy
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/e2e-tests
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/e2e-tests
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/e2e-tests/conf
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/e2e-tests/conf
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ export TEMP_DIR=/tmp/kuttl/pg/start-from-backup
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ TEMP_DIR=/tmp/kuttl/pg/start-from-backup
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | ++++ git rev-parse --abbrev-ref HEAD
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ export GIT_BRANCH=PR-1021
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ GIT_BRANCH=PR-1021
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ export VERSION=PR-1021-eed29e68b
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ VERSION=PR-1021-eed29e68b
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ IMAGE_BASE=perconalab/percona-postgresql-operator
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1021-eed29e68b
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1021-eed29e68b
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ export PG_VER=17
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ PG_VER=17
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-ppg17-pgbouncer
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-ppg17-pgbouncer
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ export IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg17-postgres
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ IMAGE_POSTGRESQL=perconalab/percona-postgresql-operator:main-ppg17-postgres
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-ppg17-pgbackrest
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-ppg17-pgbackrest
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ export BUCKET=pg-operator-testing
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ BUCKET=pg-operator-testing
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ export PMM_SERVER_VERSION=9.9.9
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ PMM_SERVER_VERSION=9.9.9
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ export PGOV1_TAG=1.4.0
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ PGOV1_TAG=1.4.0
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ export PGOV1_VER=14
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ PGOV1_VER=14
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | ++++ which gdate
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | ++++ which date
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ date=/usr/bin/date
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | ++++ which gsed
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1021/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin)
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | ++++ which sed
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ sed=/usr/bin/sed
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | +++ command -v oc
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | ++ oc get projects
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | + destroy_operator
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | + kubectl -n pg-operator delete deployment percona-postgresql-operator --force --grace-period=0
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | deployment.apps "percona-postgresql-operator" force deleted
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | + [[ -n pg-operator ]]
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | + kubectl delete namespace pg-operator --force --grace-period=0
    logger.go:42: 15:49:46 | start-from-backup/99-remove-cluster-gracefully | Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.
    logger.go:42: 15:49:47 | start-from-backup/99-remove-cluster-gracefully | namespace "pg-operator" force deleted
    logger.go:42: 15:49:53 | start-from-backup/99-remove-cluster-gracefully | test step completed 99-remove-cluster-gracefully
    logger.go:42: 15:49:53 | start-from-backup | start-from-backup events from ns kuttl-test-resolved-akita:
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:45:44 +0000 UTC	Normal	Pod pg-client-6cc584874-xr4zw		Scheduled	Successfully assigned kuttl-test-resolved-akita/pg-client-6cc584874-xr4zw to gke-jen-pg-1021-eed29e68-default-pool-b571bffb-6jb2	default-scheduler	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:45:44 +0000 UTC	Normal	ReplicaSet.apps pg-client-6cc584874		SuccessfulCreate	Created pod: pg-client-6cc584874-xr4zw	replicaset-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:45:44 +0000 UTC	Normal	Deployment.apps pg-client		ScalingReplicaSet	Scaled up replica set pg-client-6cc584874 to 1	deployment-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:45:45 +0000 UTC	Normal	Pod pg-client-6cc584874-xr4zw.spec.containers{pg-client}		Pulling	Pulling image "perconalab/percona-distribution-postgresql:15"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:45:45 +0000 UTC	Normal	Pod pg-client-6cc584874-xr4zw.spec.containers{pg-client}		Pulled	Successfully pulled image "perconalab/percona-distribution-postgresql:15" in 142ms (142ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:45:45 +0000 UTC	Normal	Pod pg-client-6cc584874-xr4zw.spec.containers{pg-client}		Created	Created container pg-client	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:45:46 +0000 UTC	Normal	Pod pg-client-6cc584874-xr4zw.spec.containers{pg-client}		Started	Started container pg-client	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:45:46 +0000 UTC	Normal	Pod start-from-backup-patroni-version-check		Scheduled	Successfully assigned kuttl-test-resolved-akita/start-from-backup-patroni-version-check to gke-jen-pg-1021-eed29e68-default-pool-b571bffb-6jb2	default-scheduler	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:45:47 +0000 UTC	Normal	Pod start-from-backup-patroni-version-check.spec.containers{patroni-version-check}		Pulled	Container image "perconalab/percona-postgresql-operator:main-ppg17-postgres" already present on machine	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:45:47 +0000 UTC	Normal	Pod start-from-backup-patroni-version-check.spec.containers{patroni-version-check}		Created	Created container patroni-version-check	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:45:47 +0000 UTC	Normal	Pod start-from-backup-patroni-version-check.spec.containers{patroni-version-check}		Started	Started container patroni-version-check	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:45:52 +0000 UTC	Normal	Pod start-from-backup-patroni-version-check.spec.containers{patroni-version-check}		Killing	Stopping container patroni-version-check	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:45:53 +0000 UTC	Normal	PersistentVolumeClaim start-from-backup-instance1-7f74-pgdata		WaitForFirstConsumer	waiting for first consumer to be created before binding	persistentvolume-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:45:53 +0000 UTC	Normal	PersistentVolumeClaim start-from-backup-instance1-7f74-pgdata		WaitForPodScheduled	waiting for pod start-from-backup-pgbackrest-restore-pcdjm to be scheduled	persistentvolume-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:45:53 +0000 UTC	Warning	Pod start-from-backup-pgbackrest-restore-pcdjm		FailedScheduling	running PreBind plugin "VolumeBinding": Operation cannot be fulfilled on persistentvolumeclaims "start-from-backup-instance1-7f74-pgdata": the object has been modified; please apply your changes to the latest version and try again	default-scheduler	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:45:53 +0000 UTC	Normal	Job.batch start-from-backup-pgbackrest-restore		SuccessfulCreate	Created pod: start-from-backup-pgbackrest-restore-pcdjm	job-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:45:55 +0000 UTC	Normal	PersistentVolumeClaim start-from-backup-instance1-7f74-pgdata		ExternalProvisioning	Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.	persistentvolume-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:45:55 +0000 UTC	Normal	PersistentVolumeClaim start-from-backup-instance1-7f74-pgdata		Provisioning	External provisioner is provisioning volume for claim "kuttl-test-resolved-akita/start-from-backup-instance1-7f74-pgdata"	pd.csi.storage.gke.io_gke-5d308b4d6eba408e9c5a-7be6-19bc-vm_4e673af0-6aea-4936-b4ba-2c51a1e706f1	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:45:59 +0000 UTC	Normal	PersistentVolumeClaim start-from-backup-instance1-7f74-pgdata		ProvisioningSucceeded	Successfully provisioned volume pvc-48b8c11d-299c-4e47-9401-fd97f4baec0a	pd.csi.storage.gke.io_gke-5d308b4d6eba408e9c5a-7be6-19bc-vm_4e673af0-6aea-4936-b4ba-2c51a1e706f1	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:45:59 +0000 UTC	Normal	Pod start-from-backup-pgbackrest-restore-pcdjm		Scheduled	Successfully assigned kuttl-test-resolved-akita/start-from-backup-pgbackrest-restore-pcdjm to gke-jen-pg-1021-eed29e68-default-pool-b571bffb-6jb2	default-scheduler	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:46:06 +0000 UTC	Normal	Pod start-from-backup-pgbackrest-restore-pcdjm		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-48b8c11d-299c-4e47-9401-fd97f4baec0a" 	attachdetach-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:46:09 +0000 UTC	Normal	Pod start-from-backup-pgbackrest-restore-pcdjm.spec.initContainers{nss-wrapper-init}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-pgbackrest"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:46:09 +0000 UTC	Normal	Pod start-from-backup-pgbackrest-restore-pcdjm.spec.initContainers{nss-wrapper-init}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-pgbackrest" in 143ms (143ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:46:09 +0000 UTC	Normal	Pod start-from-backup-pgbackrest-restore-pcdjm.spec.initContainers{nss-wrapper-init}		Created	Created container nss-wrapper-init	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:46:09 +0000 UTC	Normal	Pod start-from-backup-pgbackrest-restore-pcdjm.spec.initContainers{nss-wrapper-init}		Started	Started container nss-wrapper-init	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:46:10 +0000 UTC	Normal	Pod start-from-backup-pgbackrest-restore-pcdjm.spec.containers{pgbackrest-restore}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-postgres"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:46:10 +0000 UTC	Normal	Pod start-from-backup-pgbackrest-restore-pcdjm.spec.containers{pgbackrest-restore}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-postgres" in 140ms (140ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:46:10 +0000 UTC	Normal	Pod start-from-backup-pgbackrest-restore-pcdjm.spec.containers{pgbackrest-restore}		Created	Created container pgbackrest-restore	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:46:10 +0000 UTC	Normal	Pod start-from-backup-pgbackrest-restore-pcdjm.spec.containers{pgbackrest-restore}		Started	Started container pgbackrest-restore	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:30 +0000 UTC	Normal	StatefulSet.apps start-from-backup-instance1-7f74		SuccessfulCreate	create Pod start-from-backup-instance1-7f74-0 in StatefulSet start-from-backup-instance1-7f74 successful	statefulset-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:30 +0000 UTC	Normal	PersistentVolumeClaim start-from-backup-instance1-c44t-pgdata		WaitForFirstConsumer	waiting for first consumer to be created before binding	persistentvolume-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:30 +0000 UTC	Normal	Job.batch start-from-backup-pgbackrest-restore		Completed	Job completed	job-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:31 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0		Scheduled	Successfully assigned kuttl-test-resolved-akita/start-from-backup-instance1-7f74-0 to gke-jen-pg-1021-eed29e68-default-pool-b571bffb-6jb2	default-scheduler	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:31 +0000 UTC	Warning	Pod start-from-backup-instance1-7f74-0		FailedMount	MountVolume.SetUp failed for volume "pgbackrest-config" : secret "start-from-backup-pgbackrest" not found	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:31 +0000 UTC	Normal	PersistentVolumeClaim start-from-backup-instance1-bh4l-pgdata		WaitForFirstConsumer	waiting for first consumer to be created before binding	persistentvolume-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:31 +0000 UTC	Normal	Pod start-from-backup-repo-host-0		Scheduled	Successfully assigned kuttl-test-resolved-akita/start-from-backup-repo-host-0 to gke-jen-pg-1021-eed29e68-default-pool-b571bffb-htmz	default-scheduler	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:31 +0000 UTC	Warning	StatefulSet.apps start-from-backup-repo-host		FailedCreate	create Pod start-from-backup-repo-host-0 in StatefulSet start-from-backup-repo-host failed error: pods "start-from-backup-repo-host-0" is forbidden: error looking up service account kuttl-test-resolved-akita/start-from-backup-pgbackrest: serviceaccount "start-from-backup-pgbackrest" not found	statefulset-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:31 +0000 UTC	Normal	StatefulSet.apps start-from-backup-repo-host		SuccessfulCreate	create Pod start-from-backup-repo-host-0 in StatefulSet start-from-backup-repo-host successful	statefulset-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:31 +0000 UTC	Normal	PostgresCluster.postgres-operator.crunchydata.com start-from-backup		RepoHostCreated	created pgBackRest repository host StatefulSet/start-from-backup-repo-host	postgrescluster-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:32 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-2wqch		Scheduled	Successfully assigned kuttl-test-resolved-akita/start-from-backup-pgbouncer-7bfc966845-2wqch to gke-jen-pg-1021-eed29e68-default-pool-b571bffb-d886	default-scheduler	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:32 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-2wqch.spec.containers{pgbouncer}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-pgbouncer"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:32 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-4g7wj		Scheduled	Successfully assigned kuttl-test-resolved-akita/start-from-backup-pgbouncer-7bfc966845-4g7wj to gke-jen-pg-1021-eed29e68-default-pool-b571bffb-htmz	default-scheduler	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:32 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-4g7wj.spec.containers{pgbouncer}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-pgbouncer"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:32 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-8ptxp		Scheduled	Successfully assigned kuttl-test-resolved-akita/start-from-backup-pgbouncer-7bfc966845-8ptxp to gke-jen-pg-1021-eed29e68-default-pool-b571bffb-6jb2	default-scheduler	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:32 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-8ptxp.spec.containers{pgbouncer}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-pgbouncer"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:32 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-8ptxp.spec.containers{pgbouncer}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-pgbouncer" in 106ms (106ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:32 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-8ptxp.spec.containers{pgbouncer}		Created	Created container pgbouncer	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:32 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-8ptxp.spec.containers{pgbouncer}		Started	Started container pgbouncer	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:32 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-8ptxp.spec.containers{pgbouncer-config}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-pgbouncer"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:32 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-8ptxp.spec.containers{pgbouncer-config}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-pgbouncer" in 120ms (120ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:32 +0000 UTC	Normal	ReplicaSet.apps start-from-backup-pgbouncer-7bfc966845		SuccessfulCreate	Created pod: start-from-backup-pgbouncer-7bfc966845-8ptxp	replicaset-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:32 +0000 UTC	Normal	ReplicaSet.apps start-from-backup-pgbouncer-7bfc966845		SuccessfulCreate	Created pod: start-from-backup-pgbouncer-7bfc966845-4g7wj	replicaset-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:32 +0000 UTC	Normal	ReplicaSet.apps start-from-backup-pgbouncer-7bfc966845		SuccessfulCreate	Created pod: start-from-backup-pgbouncer-7bfc966845-2wqch	replicaset-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:32 +0000 UTC	Normal	Deployment.apps start-from-backup-pgbouncer		ScalingReplicaSet	Scaled up replica set start-from-backup-pgbouncer-7bfc966845 to 3	deployment-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:32 +0000 UTC	Normal	PodDisruptionBudget.policy start-from-backup-pgbouncer		NoPods	No matching pods found	controllermanager	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:32 +0000 UTC	Normal	Pod start-from-backup-repo-host-0.spec.initContainers{nss-wrapper-init}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-pgbackrest"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:32 +0000 UTC	Normal	Pod start-from-backup-repo-host-0.spec.initContainers{nss-wrapper-init}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-pgbackrest" in 136ms (136ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:32 +0000 UTC	Normal	Pod start-from-backup-repo-host-0.spec.initContainers{nss-wrapper-init}		Created	Created container nss-wrapper-init	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:32 +0000 UTC	Normal	Pod start-from-backup-repo-host-0.spec.initContainers{nss-wrapper-init}		Started	Started container nss-wrapper-init	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:33 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-2wqch.spec.containers{pgbouncer}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-pgbouncer" in 124ms (124ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:33 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-2wqch.spec.containers{pgbouncer}		Created	Created container pgbouncer	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:33 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-2wqch.spec.containers{pgbouncer}		Started	Started container pgbouncer	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:33 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-2wqch.spec.containers{pgbouncer-config}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-pgbouncer"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:33 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-2wqch.spec.containers{pgbouncer-config}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-pgbouncer" in 130ms (130ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:33 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-2wqch.spec.containers{pgbouncer-config}		Created	Created container pgbouncer-config	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:33 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-2wqch.spec.containers{pgbouncer-config}		Started	Started container pgbouncer-config	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:33 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-4g7wj.spec.containers{pgbouncer}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-pgbouncer" in 130ms (130ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:33 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-4g7wj.spec.containers{pgbouncer}		Created	Created container pgbouncer	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:33 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-4g7wj.spec.containers{pgbouncer}		Started	Started container pgbouncer	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:33 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-4g7wj.spec.containers{pgbouncer-config}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-pgbouncer"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:33 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-4g7wj.spec.containers{pgbouncer-config}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-pgbouncer" in 177ms (177ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:33 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-4g7wj.spec.containers{pgbouncer-config}		Created	Created container pgbouncer-config	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:33 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-4g7wj.spec.containers{pgbouncer-config}		Started	Started container pgbouncer-config	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:33 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-8ptxp.spec.containers{pgbouncer-config}		Created	Created container pgbouncer-config	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:33 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-8ptxp.spec.containers{pgbouncer-config}		Started	Started container pgbouncer-config	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:33 +0000 UTC	Normal	Pod start-from-backup-repo-host-0.spec.containers{pgbackrest}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-pgbackrest"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:33 +0000 UTC	Normal	Pod start-from-backup-repo-host-0.spec.containers{pgbackrest}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-pgbackrest" in 186ms (186ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:33 +0000 UTC	Normal	Pod start-from-backup-repo-host-0.spec.containers{pgbackrest}		Created	Created container pgbackrest	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:33 +0000 UTC	Normal	Pod start-from-backup-repo-host-0.spec.containers{pgbackrest}		Started	Started container pgbackrest	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:33 +0000 UTC	Normal	Pod start-from-backup-repo-host-0.spec.containers{pgbackrest-config}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-pgbackrest"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:33 +0000 UTC	Normal	Pod start-from-backup-repo-host-0.spec.containers{pgbackrest-config}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-pgbackrest" in 131ms (131ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:33 +0000 UTC	Normal	Pod start-from-backup-repo-host-0.spec.containers{pgbackrest-config}		Created	Created container pgbackrest-config	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:33 +0000 UTC	Normal	Pod start-from-backup-repo-host-0.spec.containers{pgbackrest-config}		Started	Started container pgbackrest-config	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:41 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-48b8c11d-299c-4e47-9401-fd97f4baec0a" 	attachdetach-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:42 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0.spec.initContainers{postgres-startup}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-postgres"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:43 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0.spec.initContainers{postgres-startup}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-postgres" in 134ms (134ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:43 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0.spec.initContainers{postgres-startup}		Created	Created container postgres-startup	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:43 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0.spec.initContainers{postgres-startup}		Started	Started container postgres-startup	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:43 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0.spec.initContainers{nss-wrapper-init}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-postgres"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:43 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0.spec.initContainers{nss-wrapper-init}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-postgres" in 137ms (137ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:43 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0.spec.initContainers{nss-wrapper-init}		Created	Created container nss-wrapper-init	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:43 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0.spec.initContainers{nss-wrapper-init}		Started	Started container nss-wrapper-init	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:44 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0.spec.containers{database}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-postgres"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:44 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0.spec.containers{database}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-postgres" in 114ms (114ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:44 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0.spec.containers{database}		Created	Created container database	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:44 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0.spec.containers{database}		Started	Started container database	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:44 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0.spec.containers{replication-cert-copy}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-postgres"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:44 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0.spec.containers{replication-cert-copy}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-postgres" in 133ms (133ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:44 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0.spec.containers{replication-cert-copy}		Created	Created container replication-cert-copy	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:44 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0.spec.containers{replication-cert-copy}		Started	Started container replication-cert-copy	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:44 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0.spec.containers{pgbackrest}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-pgbackrest"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:45 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0.spec.containers{pgbackrest}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-pgbackrest" in 130ms (130ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:45 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0.spec.containers{pgbackrest}		Created	Created container pgbackrest	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:45 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0.spec.containers{pgbackrest}		Started	Started container pgbackrest	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:45 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0.spec.containers{pgbackrest-config}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-pgbackrest"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:45 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0.spec.containers{pgbackrest-config}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-pgbackrest" in 121ms (121ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:45 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0.spec.containers{pgbackrest-config}		Created	Created container pgbackrest-config	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:45 +0000 UTC	Normal	Pod start-from-backup-instance1-7f74-0.spec.containers{pgbackrest-config}		Started	Started container pgbackrest-config	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:49 +0000 UTC	Normal	Pod start-from-backup-backup-zfl9-trc87		Scheduled	Successfully assigned kuttl-test-resolved-akita/start-from-backup-backup-zfl9-trc87 to gke-jen-pg-1021-eed29e68-default-pool-b571bffb-d886	default-scheduler	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:49 +0000 UTC	Normal	Pod start-from-backup-backup-zfl9-trc87.spec.containers{pgbackrest}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-pgbackrest"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:49 +0000 UTC	Normal	Pod start-from-backup-backup-zfl9-trc87.spec.containers{pgbackrest}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-pgbackrest" in 117ms (117ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:49 +0000 UTC	Normal	Pod start-from-backup-backup-zfl9-trc87.spec.containers{pgbackrest}		Created	Created container pgbackrest	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:49 +0000 UTC	Normal	Pod start-from-backup-backup-zfl9-trc87.spec.containers{pgbackrest}		Started	Started container pgbackrest	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:49 +0000 UTC	Normal	Job.batch start-from-backup-backup-zfl9		SuccessfulCreate	Created pod: start-from-backup-backup-zfl9-trc87	job-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:49 +0000 UTC	Normal	PostgresCluster.postgres-operator.crunchydata.com start-from-backup		StanzasCreated	pgBackRest stanza creation completed successfully	postgrescluster-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:50 +0000 UTC	Normal	StatefulSet.apps start-from-backup-instance1-c44t		SuccessfulCreate	create Pod start-from-backup-instance1-c44t-0 in StatefulSet start-from-backup-instance1-c44t successful	statefulset-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:51 +0000 UTC	Normal	PersistentVolumeClaim start-from-backup-instance1-bh4l-pgdata		ExternalProvisioning	Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.	persistentvolume-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:51 +0000 UTC	Normal	PersistentVolumeClaim start-from-backup-instance1-bh4l-pgdata		Provisioning	External provisioner is provisioning volume for claim "kuttl-test-resolved-akita/start-from-backup-instance1-bh4l-pgdata"	pd.csi.storage.gke.io_gke-5d308b4d6eba408e9c5a-7be6-19bc-vm_4e673af0-6aea-4936-b4ba-2c51a1e706f1	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:51 +0000 UTC	Normal	StatefulSet.apps start-from-backup-instance1-bh4l		SuccessfulCreate	create Pod start-from-backup-instance1-bh4l-0 in StatefulSet start-from-backup-instance1-bh4l successful	statefulset-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:51 +0000 UTC	Normal	PersistentVolumeClaim start-from-backup-instance1-c44t-pgdata		Provisioning	External provisioner is provisioning volume for claim "kuttl-test-resolved-akita/start-from-backup-instance1-c44t-pgdata"	pd.csi.storage.gke.io_gke-5d308b4d6eba408e9c5a-7be6-19bc-vm_4e673af0-6aea-4936-b4ba-2c51a1e706f1	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:51 +0000 UTC	Normal	PersistentVolumeClaim start-from-backup-instance1-c44t-pgdata		ExternalProvisioning	Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.	persistentvolume-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:51 +0000 UTC	Warning	PostgresCluster.postgres-operator.crunchydata.com start-from-backup		UnableToCreateStanzas	command terminated with exit code 50: 2025-01-24 15:47:51.784 P00  ERROR: [050]: unable to acquire lock on file '/tmp/pgbackrest/db-backup-1.lock': Resource temporarily unavailable
                                            HINT: is another pgBackRest process running?
        2025-01-24 15:47:51.796 P00  ERROR: [050]: unable to acquire lock on file '/tmp/pgbackrest/db-backup-1.lock': Resource temporarily unavailable
                                            HINT: is another pgBackRest process running?
         	postgrescluster-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:54 +0000 UTC	Normal	PersistentVolumeClaim start-from-backup-instance1-bh4l-pgdata		ProvisioningSucceeded	Successfully provisioned volume pvc-ef0fa122-1ef2-4a99-84fe-e90d2bff28ce	pd.csi.storage.gke.io_gke-5d308b4d6eba408e9c5a-7be6-19bc-vm_4e673af0-6aea-4936-b4ba-2c51a1e706f1	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:54 +0000 UTC	Normal	PersistentVolumeClaim start-from-backup-instance1-c44t-pgdata		ProvisioningSucceeded	Successfully provisioned volume pvc-15f0f42d-95e2-4b62-bc17-4483d64618cd	pd.csi.storage.gke.io_gke-5d308b4d6eba408e9c5a-7be6-19bc-vm_4e673af0-6aea-4936-b4ba-2c51a1e706f1	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:55 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0		Scheduled	Successfully assigned kuttl-test-resolved-akita/start-from-backup-instance1-bh4l-0 to gke-jen-pg-1021-eed29e68-default-pool-b571bffb-htmz	default-scheduler	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:47:55 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0		Scheduled	Successfully assigned kuttl-test-resolved-akita/start-from-backup-instance1-c44t-0 to gke-jen-pg-1021-eed29e68-default-pool-b571bffb-d886	default-scheduler	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:02 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-ef0fa122-1ef2-4a99-84fe-e90d2bff28ce" 	attachdetach-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:02 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0		SuccessfulAttachVolume	AttachVolume.Attach succeeded for volume "pvc-15f0f42d-95e2-4b62-bc17-4483d64618cd" 	attachdetach-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:03 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0.spec.initContainers{postgres-startup}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-postgres"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:03 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0.spec.initContainers{postgres-startup}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-postgres" in 146ms (146ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:03 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0.spec.initContainers{postgres-startup}		Created	Created container postgres-startup	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:03 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0.spec.initContainers{postgres-startup}		Started	Started container postgres-startup	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:04 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0.spec.initContainers{postgres-startup}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-postgres"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:04 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0.spec.initContainers{postgres-startup}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-postgres" in 151ms (151ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:04 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0.spec.initContainers{postgres-startup}		Created	Created container postgres-startup	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:04 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0.spec.initContainers{postgres-startup}		Started	Started container postgres-startup	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:04 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0.spec.initContainers{nss-wrapper-init}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-postgres"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:04 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0.spec.initContainers{nss-wrapper-init}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-postgres" in 122ms (122ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:04 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0.spec.initContainers{nss-wrapper-init}		Created	Created container nss-wrapper-init	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:04 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0.spec.initContainers{nss-wrapper-init}		Started	Started container nss-wrapper-init	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:05 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0.spec.initContainers{nss-wrapper-init}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-postgres"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:05 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0.spec.initContainers{nss-wrapper-init}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-postgres" in 136ms (136ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:05 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0.spec.initContainers{nss-wrapper-init}		Created	Created container nss-wrapper-init	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:05 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0.spec.initContainers{nss-wrapper-init}		Started	Started container nss-wrapper-init	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:05 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0.spec.containers{database}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-postgres"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:05 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0.spec.containers{database}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-postgres" in 160ms (160ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:05 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0.spec.containers{database}		Created	Created container database	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:05 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0.spec.containers{database}		Started	Started container database	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:05 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0.spec.containers{replication-cert-copy}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-postgres"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:05 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0.spec.containers{replication-cert-copy}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-postgres" in 135ms (135ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:05 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0.spec.containers{replication-cert-copy}		Created	Created container replication-cert-copy	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:05 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0.spec.containers{replication-cert-copy}		Started	Started container replication-cert-copy	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:05 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0.spec.containers{pgbackrest}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-pgbackrest"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:05 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0.spec.containers{pgbackrest}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-pgbackrest" in 115ms (115ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:05 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0.spec.containers{pgbackrest}		Created	Created container pgbackrest	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:05 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0.spec.containers{pgbackrest}		Started	Started container pgbackrest	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:05 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0.spec.containers{pgbackrest-config}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-pgbackrest"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:06 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0.spec.containers{database}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-postgres"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:06 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0.spec.containers{database}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-postgres" in 137ms (137ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:06 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0.spec.containers{database}		Created	Created container database	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:06 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0.spec.containers{database}		Started	Started container database	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:06 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0.spec.containers{replication-cert-copy}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-postgres"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:06 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0.spec.containers{replication-cert-copy}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-postgres" in 135ms (135ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:06 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0.spec.containers{replication-cert-copy}		Created	Created container replication-cert-copy	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:06 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0.spec.containers{replication-cert-copy}		Started	Started container replication-cert-copy	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:06 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0.spec.containers{pgbackrest}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-pgbackrest"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:06 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0.spec.containers{pgbackrest-config}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-pgbackrest" in 130ms (130ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:06 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0.spec.containers{pgbackrest-config}		Created	Created container pgbackrest-config	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:06 +0000 UTC	Normal	Pod start-from-backup-instance1-c44t-0.spec.containers{pgbackrest-config}		Started	Started container pgbackrest-config	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:07 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0.spec.containers{pgbackrest}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-pgbackrest" in 136ms (136ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:07 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0.spec.containers{pgbackrest}		Created	Created container pgbackrest	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:07 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0.spec.containers{pgbackrest}		Started	Started container pgbackrest	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:07 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0.spec.containers{pgbackrest-config}		Pulling	Pulling image "perconalab/percona-postgresql-operator:main-ppg17-pgbackrest"	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:07 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0.spec.containers{pgbackrest-config}		Pulled	Successfully pulled image "perconalab/percona-postgresql-operator:main-ppg17-pgbackrest" in 131ms (131ms including waiting)	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:07 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0.spec.containers{pgbackrest-config}		Created	Created container pgbackrest-config	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:07 +0000 UTC	Normal	Pod start-from-backup-instance1-bh4l-0.spec.containers{pgbackrest-config}		Started	Started container pgbackrest-config	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:08 +0000 UTC	Warning	Pod start-from-backup-instance1-c44t-0.spec.containers{database}		Unhealthy	Readiness probe failed: HTTP probe failed with statuscode: 503	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:48:09 +0000 UTC	Warning	Pod start-from-backup-instance1-bh4l-0.spec.containers{database}		Unhealthy	Readiness probe failed: HTTP probe failed with statuscode: 503	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:49:33 +0000 UTC	Normal	Job.batch start-from-backup-backup-zfl9		Completed	Job completed	job-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:49:40 +0000 UTC	Normal	StatefulSet.apps start-from-backup-instance1-bh4l		SuccessfulDelete	delete Pod start-from-backup-instance1-bh4l-0 in StatefulSet start-from-backup-instance1-bh4l successful	statefulset-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:49:40 +0000 UTC	Normal	StatefulSet.apps start-from-backup-instance1-c44t		SuccessfulDelete	delete Pod start-from-backup-instance1-c44t-0 in StatefulSet start-from-backup-instance1-c44t successful	statefulset-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:49:42 +0000 UTC	Normal	StatefulSet.apps start-from-backup-instance1-7f74		SuccessfulDelete	delete Pod start-from-backup-instance1-7f74-0 in StatefulSet start-from-backup-instance1-7f74 successful	statefulset-controller	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:49:45 +0000 UTC	Normal	PodDisruptionBudget.policy start-from-backup-set-instance1		NoPods	No matching pods found	controllermanager	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:49:47 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-2wqch.spec.containers{pgbouncer}		Killing	Stopping container pgbouncer	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:49:47 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-2wqch.spec.containers{pgbouncer-config}		Killing	Stopping container pgbouncer-config	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:49:47 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-4g7wj.spec.containers{pgbouncer}		Killing	Stopping container pgbouncer	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:49:47 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-4g7wj.spec.containers{pgbouncer-config}		Killing	Stopping container pgbouncer-config	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:49:47 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-8ptxp.spec.containers{pgbouncer}		Killing	Stopping container pgbouncer	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:49:47 +0000 UTC	Normal	Pod start-from-backup-pgbouncer-7bfc966845-8ptxp.spec.containers{pgbouncer-config}		Killing	Stopping container pgbouncer-config	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:49:47 +0000 UTC	Normal	Pod start-from-backup-repo-host-0.spec.containers{pgbackrest}		Killing	Stopping container pgbackrest	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | 2025-01-24 15:49:47 +0000 UTC	Normal	Pod start-from-backup-repo-host-0.spec.containers{pgbackrest-config}		Killing	Stopping container pgbackrest-config	kubelet	
    logger.go:42: 15:49:53 | start-from-backup | Deleting namespace: kuttl-test-resolved-akita
=== NAME  kuttl
    harness.go:407: run tests finished
    harness.go:515: cleaning up
    harness.go:572: removing temp folder: ""
--- PASS: kuttl (296.76s)
    --- PASS: kuttl/harness (0.00s)
        --- PASS: kuttl/harness/start-from-backup (296.32s)
PASS