=== RUN kuttl harness.go:459: starting setup harness.go:254: running tests using configured kubeconfig. harness.go:277: Successful connection to cluster at: https://34.9.166.233 harness.go:362: running tests harness.go:74: going to run test suite with timeout of 180 seconds for each step harness.go:374: testsuite: e2e-tests/tests has 24 tests === RUN kuttl/harness === RUN kuttl/harness/start-from-backup === PAUSE kuttl/harness/start-from-backup === CONT kuttl/harness/start-from-backup logger.go:42: 19:38:32 | start-from-backup | Ignoring cr.yaml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 19:38:32 | start-from-backup | Creating namespace: kuttl-test-topical-oarfish logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | starting test step 0-deploy-operator logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | running command: [sh -c set -o errexit set -o xtrace source ../../functions init_temp_dir # do this only in the first TestStep deploy_operator deploy_client deploy_s3_secrets] logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | + source ../../functions logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ realpath ../../.. logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | ++ CERT_MANAGER_VER=1.17.1 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | ++++ pwd logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/tests/start-from-backup logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | ++ test_name=start-from-backup logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/vars.sh logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/deploy logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/deploy logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/conf logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/conf logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ export TEMP_DIR=/tmp/kuttl/pg/start-from-backup logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ TEMP_DIR=/tmp/kuttl/pg/start-from-backup logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ export GIT_BRANCH=PR-1212 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ GIT_BRANCH=PR-1212 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ export VERSION=PR-1212-7085fc864 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ VERSION=PR-1212-7085fc864 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1212-7085fc864 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1212-7085fc864 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ export PG_VER=17 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ PG_VER=17 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ export IMAGE_POSTGRESQL=ademajazzserve/postgres:11 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ IMAGE_POSTGRESQL=ademajazzserve/postgres:11 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ export BUCKET=pg-operator-testing logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ BUCKET=pg-operator-testing logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ export IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ export IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ export PGOV1_TAG=1.4.0 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ PGOV1_TAG=1.4.0 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ export PGOV1_VER=14 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ PGOV1_VER=14 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | ++++ which gdate logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | ++++ which date logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ date=/usr/bin/date logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | ++++ which gsed logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | ++++ which sed logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ sed=/usr/bin/sed logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | +++ command -v oc logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | ++ oc get projects logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | + init_temp_dir logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | + rm -rf /tmp/kuttl/pg/start-from-backup logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | + mkdir -p /tmp/kuttl/pg/start-from-backup logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | + deploy_operator logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | + local cw_prefix= logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | + destroy_operator logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | + kubectl -n pg-operator delete deployment percona-postgresql-operator --force --grace-period=0 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | Error from server (NotFound): deployments.apps "percona-postgresql-operator" not found logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | + true logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | + [[ -n pg-operator ]] logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | + kubectl delete namespace pg-operator --force --grace-period=0 logger.go:42: 19:38:32 | start-from-backup/0-deploy-operator | Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. logger.go:42: 19:38:33 | start-from-backup/0-deploy-operator | Error from server (NotFound): namespaces "pg-operator" not found logger.go:42: 19:38:33 | start-from-backup/0-deploy-operator | + true logger.go:42: 19:38:33 | start-from-backup/0-deploy-operator | + [[ -n pg-operator ]] logger.go:42: 19:38:33 | start-from-backup/0-deploy-operator | + create_namespace pg-operator logger.go:42: 19:38:33 | start-from-backup/0-deploy-operator | + local namespace=pg-operator logger.go:42: 19:38:33 | start-from-backup/0-deploy-operator | + [[ -n '' ]] logger.go:42: 19:38:33 | start-from-backup/0-deploy-operator | + kubectl delete namespace pg-operator --ignore-not-found logger.go:42: 19:38:33 | start-from-backup/0-deploy-operator | + kubectl wait --for=delete namespace pg-operator logger.go:42: 19:38:34 | start-from-backup/0-deploy-operator | + kubectl create namespace pg-operator logger.go:42: 19:38:34 | start-from-backup/0-deploy-operator | namespace/pg-operator created logger.go:42: 19:38:34 | start-from-backup/0-deploy-operator | + cw_prefix=cw- logger.go:42: 19:38:34 | start-from-backup/0-deploy-operator | + kubectl -n pg-operator apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pg-operator_PR-1212/deploy/crd.yaml logger.go:42: 19:38:35 | start-from-backup/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/crunchybridgeclusters.postgres-operator.crunchydata.com serverside-applied logger.go:42: 19:38:35 | start-from-backup/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconapgbackups.pgv2.percona.com serverside-applied logger.go:42: 19:38:37 | start-from-backup/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconapgclusters.pgv2.percona.com serverside-applied logger.go:42: 19:38:38 | start-from-backup/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconapgrestores.pgv2.percona.com serverside-applied logger.go:42: 19:38:38 | start-from-backup/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconapgupgrades.pgv2.percona.com serverside-applied logger.go:42: 19:38:38 | start-from-backup/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/pgadmins.postgres-operator.crunchydata.com serverside-applied logger.go:42: 19:38:39 | start-from-backup/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/pgupgrades.postgres-operator.crunchydata.com serverside-applied logger.go:42: 19:38:41 | start-from-backup/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/postgresclusters.postgres-operator.crunchydata.com serverside-applied logger.go:42: 19:38:41 | start-from-backup/0-deploy-operator | + kubectl -n pg-operator apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pg-operator_PR-1212/deploy/cw-rbac.yaml logger.go:42: 19:38:42 | start-from-backup/0-deploy-operator | serviceaccount/percona-postgresql-operator serverside-applied logger.go:42: 19:38:42 | start-from-backup/0-deploy-operator | clusterrole.rbac.authorization.k8s.io/percona-postgresql-operator serverside-applied logger.go:42: 19:38:42 | start-from-backup/0-deploy-operator | clusterrolebinding.rbac.authorization.k8s.io/percona-postgresql-operator serverside-applied logger.go:42: 19:38:42 | start-from-backup/0-deploy-operator | + local disable_telemetry=true logger.go:42: 19:38:42 | start-from-backup/0-deploy-operator | + '[' start-from-backup == telemetry-transfer ']' logger.go:42: 19:38:42 | start-from-backup/0-deploy-operator | + yq eval '.spec.template.spec.containers[0].image = "perconalab/percona-postgresql-operator:PR-1212-7085fc864"' /mnt/jenkins/workspace/cloud-pg-operator_PR-1212/deploy/cw-operator.yaml logger.go:42: 19:38:42 | start-from-backup/0-deploy-operator | + kubectl -n pg-operator apply -f - logger.go:42: 19:38:42 | start-from-backup/0-deploy-operator | + yq eval '(.spec.template.spec.containers[] | select(.name=="operator") | .env[] | select(.name=="DISABLE_TELEMETRY") | .value) = "true"' - logger.go:42: 19:38:43 | start-from-backup/0-deploy-operator | deployment.apps/percona-postgresql-operator created logger.go:42: 19:38:43 | start-from-backup/0-deploy-operator | + deploy_client logger.go:42: 19:38:43 | start-from-backup/0-deploy-operator | + kubectl -n kuttl-test-topical-oarfish apply -f /mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/conf/client.yaml logger.go:42: 19:38:44 | start-from-backup/0-deploy-operator | deployment.apps/pg-client created logger.go:42: 19:38:44 | start-from-backup/0-deploy-operator | + deploy_s3_secrets logger.go:42: 19:38:44 | start-from-backup/0-deploy-operator | + set +o xtrace logger.go:42: 19:38:44 | start-from-backup/0-deploy-operator | secret/start-from-backup-pgbackrest-secrets created logger.go:42: 19:38:45 | start-from-backup/0-deploy-operator | running command: [sh -c kubectl assert exist-enhanced deployment percona-postgresql-operator -n ${OPERATOR_NS:-$NAMESPACE} --field-selector status.readyReplicas=1] logger.go:42: 19:38:45 | start-from-backup/0-deploy-operator | ASSERT deployment percona-postgresql-operator matching field criteria 'status.readyReplicas=1' should exist. logger.go:42: 19:38:46 | start-from-backup/0-deploy-operator | INFO Found 1 resource(s). logger.go:42: 19:38:46 | start-from-backup/0-deploy-operator | NAME NAMESPACE COL0 logger.go:42: 19:38:46 | start-from-backup/0-deploy-operator | percona-postgresql-operator pg-operator 1 logger.go:42: 19:38:46 | start-from-backup/0-deploy-operator | ASSERT PASS logger.go:42: 19:38:46 | start-from-backup/0-deploy-operator | test step completed 0-deploy-operator logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | starting test step 1-create-source-cluster logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | running command: [sh -c set -o errexit set -o xtrace source ../../functions get_cr "start-from-backup-source" "demand-backup-ppg$PG_VER" \ | kubectl -n "${NAMESPACE}" apply -f -] logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | + source ../../functions logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ realpath ../../.. logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | ++ CERT_MANAGER_VER=1.17.1 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | ++++ pwd logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/tests/start-from-backup logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | ++ test_name=start-from-backup logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/vars.sh logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/deploy logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/deploy logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/conf logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/conf logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ export TEMP_DIR=/tmp/kuttl/pg/start-from-backup logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ TEMP_DIR=/tmp/kuttl/pg/start-from-backup logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ export GIT_BRANCH=PR-1212 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ GIT_BRANCH=PR-1212 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ export VERSION=PR-1212-7085fc864 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ VERSION=PR-1212-7085fc864 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1212-7085fc864 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1212-7085fc864 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ export PG_VER=17 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ PG_VER=17 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ export IMAGE_POSTGRESQL=ademajazzserve/postgres:11 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ IMAGE_POSTGRESQL=ademajazzserve/postgres:11 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ export BUCKET=pg-operator-testing logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ BUCKET=pg-operator-testing logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ export IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ export IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ export PGOV1_TAG=1.4.0 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ PGOV1_TAG=1.4.0 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ export PGOV1_VER=14 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ PGOV1_VER=14 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | ++++ which gdate logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | ++++ which date logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ date=/usr/bin/date logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | ++++ which gsed logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | ++++ which sed logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ sed=/usr/bin/sed logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | +++ command -v oc logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | ++ oc get projects logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | + get_cr start-from-backup-source demand-backup-ppg17 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | + local cr_name=start-from-backup-source logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | + '[' -z start-from-backup-source ']' logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | + local repo_path=demand-backup-ppg17 logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | + yq eval ' logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | .metadata.name = "start-from-backup-source" | logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | .metadata.labels = {"e2e":"start-from-backup-source"} | logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | .spec.postgresVersion = 17 | logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | .spec.users += [{"name":"postgres","password":{"type":"AlphaNumeric"}}] | logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | .spec.users += [{"name":"start-from-backup-source","password":{"type":"AlphaNumeric"}}] | logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | .spec.image = "ademajazzserve/postgres:11" | logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | .spec.initContainer.image = "perconalab/percona-postgresql-operator:PR-1212-7085fc864" | logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | .spec.backups.pgbackrest.image = "perconalab/percona-postgresql-operator:main-pgbackrest17" | logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | .spec.proxy.pgBouncer.image = "perconalab/percona-postgresql-operator:main-pgbouncer17" | logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | .spec.pmm.image = "perconalab/pmm-client:dev-latest" | logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | .spec.pmm.secret = "start-from-backup-source-pmm-secret" | logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | .spec.pmm.customClusterName = "start-from-backup-source-pmm-custom-name" | logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | .spec.pmm.postgresParams = "--environment=dev-postgres" logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | ' /mnt/jenkins/workspace/cloud-pg-operator_PR-1212/deploy/cr.yaml logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | + kubectl -n kuttl-test-topical-oarfish apply -f - logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | + [[ -n '' ]] logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | + case $test_name in logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | + yq eval -i ' logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | .spec.backups.pgbackrest.configuration = [{"secret":{"name":"start-from-backup-pgbackrest-secrets"}}] | logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | .spec.backups.pgbackrest.manual.repoName = "repo1" | logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | .spec.backups.pgbackrest.manual.options = ["--type=full"] | logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | .spec.backups.pgbackrest.global.repo1-path = "/backrestrepo/postgres-operator/demand-backup-ppg17/repo1" | logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | .spec.backups.pgbackrest.repos = [{"name":"repo1","s3":{"bucket":"pg-operator-testing","endpoint":"s3.amazonaws.com","region":"us-east-1"}}] logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | ' /tmp/kuttl/pg/start-from-backup/cr.yaml logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | + [[ start-from-backup == \d\e\m\a\n\d\-\b\a\c\k\u\p ]] logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | + [[ start-from-backup == \s\t\a\r\t\-\f\r\o\m\-\b\a\c\k\u\p ]] logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | + [[ ! start-from-backup-source =~ source ]] logger.go:42: 19:38:46 | start-from-backup/1-create-source-cluster | + cat /tmp/kuttl/pg/start-from-backup/cr.yaml logger.go:42: 19:38:47 | start-from-backup/1-create-source-cluster | perconapgcluster.pgv2.percona.com/start-from-backup-source created logger.go:42: 19:40:45 | start-from-backup/1-create-source-cluster | test step completed 1-create-source-cluster logger.go:42: 19:40:45 | start-from-backup/2-write-data | starting test step 2-write-data logger.go:42: 19:40:45 | start-from-backup/2-write-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions run_psql_local \ 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' \ "postgres:$(get_psql_user_pass start-from-backup-source-pguser-postgres)@$(get_psql_user_host start-from-backup-source-pguser-postgres)" run_psql_local \ '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500), (100501)' \ "postgres:$(get_psql_user_pass start-from-backup-source-pguser-postgres)@$(get_psql_user_host start-from-backup-source-pguser-postgres)"] logger.go:42: 19:40:45 | start-from-backup/2-write-data | + source ../../functions logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ realpath ../../.. logger.go:42: 19:40:45 | start-from-backup/2-write-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212 logger.go:42: 19:40:45 | start-from-backup/2-write-data | ++ CERT_MANAGER_VER=1.17.1 logger.go:42: 19:40:45 | start-from-backup/2-write-data | ++++ pwd logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/tests/start-from-backup logger.go:42: 19:40:45 | start-from-backup/2-write-data | ++ test_name=start-from-backup logger.go:42: 19:40:45 | start-from-backup/2-write-data | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/vars.sh logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212 logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212 logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/deploy logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/deploy logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/conf logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/conf logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ export TEMP_DIR=/tmp/kuttl/pg/start-from-backup logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ TEMP_DIR=/tmp/kuttl/pg/start-from-backup logger.go:42: 19:40:45 | start-from-backup/2-write-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ export GIT_BRANCH=PR-1212 logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ GIT_BRANCH=PR-1212 logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ export VERSION=PR-1212-7085fc864 logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ VERSION=PR-1212-7085fc864 logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1212-7085fc864 logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1212-7085fc864 logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ export PG_VER=17 logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ PG_VER=17 logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ export IMAGE_POSTGRESQL=ademajazzserve/postgres:11 logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ IMAGE_POSTGRESQL=ademajazzserve/postgres:11 logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ export BUCKET=pg-operator-testing logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ BUCKET=pg-operator-testing logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ export IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ export IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ export PGOV1_TAG=1.4.0 logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ PGOV1_TAG=1.4.0 logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ export PGOV1_VER=14 logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ PGOV1_VER=14 logger.go:42: 19:40:45 | start-from-backup/2-write-data | ++++ which gdate logger.go:42: 19:40:45 | start-from-backup/2-write-data | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 19:40:45 | start-from-backup/2-write-data | ++++ which date logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ date=/usr/bin/date logger.go:42: 19:40:45 | start-from-backup/2-write-data | ++++ which gsed logger.go:42: 19:40:45 | start-from-backup/2-write-data | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 19:40:45 | start-from-backup/2-write-data | ++++ which sed logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ sed=/usr/bin/sed logger.go:42: 19:40:45 | start-from-backup/2-write-data | +++ command -v oc logger.go:42: 19:40:45 | start-from-backup/2-write-data | ++ oc get projects logger.go:42: 19:40:45 | start-from-backup/2-write-data | ++ get_psql_user_pass start-from-backup-source-pguser-postgres logger.go:42: 19:40:45 | start-from-backup/2-write-data | ++ local secret_name=start-from-backup-source-pguser-postgres logger.go:42: 19:40:45 | start-from-backup/2-write-data | ++ kubectl -n kuttl-test-topical-oarfish get secret/start-from-backup-source-pguser-postgres '--template={{.data.password | base64decode}}' logger.go:42: 19:40:46 | start-from-backup/2-write-data | ++ get_psql_user_host start-from-backup-source-pguser-postgres logger.go:42: 19:40:46 | start-from-backup/2-write-data | ++ local secret_name=start-from-backup-source-pguser-postgres logger.go:42: 19:40:46 | start-from-backup/2-write-data | ++ kubectl -n kuttl-test-topical-oarfish get secret/start-from-backup-source-pguser-postgres '--template={{.data.host | base64decode }}' logger.go:42: 19:40:47 | start-from-backup/2-write-data | + run_psql_local 'CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' postgres:dUieZDm7774QUDAeEkacTgv4@start-from-backup-source-primary.kuttl-test-topical-oarfish.svc logger.go:42: 19:40:47 | start-from-backup/2-write-data | + local 'command=CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' logger.go:42: 19:40:47 | start-from-backup/2-write-data | + local uri=postgres:dUieZDm7774QUDAeEkacTgv4@start-from-backup-source-primary.kuttl-test-topical-oarfish.svc logger.go:42: 19:40:47 | start-from-backup/2-write-data | + local driver=postgres logger.go:42: 19:40:47 | start-from-backup/2-write-data | ++ get_client_pod logger.go:42: 19:40:47 | start-from-backup/2-write-data | ++ kubectl -n kuttl-test-topical-oarfish get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 19:40:47 | start-from-backup/2-write-data | + kubectl -n kuttl-test-topical-oarfish exec pg-client-84d6c45668-g89hv -- bash -c 'printf '\''CREATE DATABASE myapp; \c myapp \\\ CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://'\''postgres:dUieZDm7774QUDAeEkacTgv4@start-from-backup-source-primary.kuttl-test-topical-oarfish.svc'\''' logger.go:42: 19:40:48 | start-from-backup/2-write-data | ++ get_psql_user_pass start-from-backup-source-pguser-postgres logger.go:42: 19:40:48 | start-from-backup/2-write-data | ++ local secret_name=start-from-backup-source-pguser-postgres logger.go:42: 19:40:48 | start-from-backup/2-write-data | ++ kubectl -n kuttl-test-topical-oarfish get secret/start-from-backup-source-pguser-postgres '--template={{.data.password | base64decode}}' logger.go:42: 19:40:48 | start-from-backup/2-write-data | ++ get_psql_user_host start-from-backup-source-pguser-postgres logger.go:42: 19:40:48 | start-from-backup/2-write-data | ++ local secret_name=start-from-backup-source-pguser-postgres logger.go:42: 19:40:48 | start-from-backup/2-write-data | ++ kubectl -n kuttl-test-topical-oarfish get secret/start-from-backup-source-pguser-postgres '--template={{.data.host | base64decode }}' logger.go:42: 19:40:49 | start-from-backup/2-write-data | + run_psql_local '\c myapp \\\ INSERT INTO myApp (id) VALUES (100500), (100501)' postgres:dUieZDm7774QUDAeEkacTgv4@start-from-backup-source-primary.kuttl-test-topical-oarfish.svc logger.go:42: 19:40:49 | start-from-backup/2-write-data | + local 'command=\c myapp \\\ INSERT INTO myApp (id) VALUES (100500), (100501)' logger.go:42: 19:40:49 | start-from-backup/2-write-data | + local uri=postgres:dUieZDm7774QUDAeEkacTgv4@start-from-backup-source-primary.kuttl-test-topical-oarfish.svc logger.go:42: 19:40:49 | start-from-backup/2-write-data | + local driver=postgres logger.go:42: 19:40:49 | start-from-backup/2-write-data | ++ get_client_pod logger.go:42: 19:40:49 | start-from-backup/2-write-data | ++ kubectl -n kuttl-test-topical-oarfish get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 19:40:49 | start-from-backup/2-write-data | + kubectl -n kuttl-test-topical-oarfish exec pg-client-84d6c45668-g89hv -- bash -c 'printf '\''\c myapp \\\ INSERT INTO myApp (id) VALUES (100500), (100501)\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://'\''postgres:dUieZDm7774QUDAeEkacTgv4@start-from-backup-source-primary.kuttl-test-topical-oarfish.svc'\''' logger.go:42: 19:40:50 | start-from-backup/2-write-data | test step completed 2-write-data logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | starting test step 3-read-from-primary logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | running command: [sh -c set -o errexit set -o xtrace source ../../functions data=$(run_psql_local '\c myapp \\\ SELECT * from myApp;' "postgres:$(get_psql_user_pass start-from-backup-source-pguser-postgres)@$(get_psql_user_host start-from-backup-source-pguser-postgres)") kubectl create configmap -n "${NAMESPACE}" 03-read-from-primary --from-literal=data="${data}"] logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | + source ../../functions logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ realpath ../../.. logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212 logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | ++ CERT_MANAGER_VER=1.17.1 logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | ++++ pwd logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/tests/start-from-backup logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | ++ test_name=start-from-backup logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/vars.sh logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212 logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212 logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/deploy logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/deploy logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/conf logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/conf logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ export TEMP_DIR=/tmp/kuttl/pg/start-from-backup logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ TEMP_DIR=/tmp/kuttl/pg/start-from-backup logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ export GIT_BRANCH=PR-1212 logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ GIT_BRANCH=PR-1212 logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ export VERSION=PR-1212-7085fc864 logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ VERSION=PR-1212-7085fc864 logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1212-7085fc864 logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1212-7085fc864 logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ export PG_VER=17 logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ PG_VER=17 logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ export IMAGE_POSTGRESQL=ademajazzserve/postgres:11 logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ IMAGE_POSTGRESQL=ademajazzserve/postgres:11 logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ export BUCKET=pg-operator-testing logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ BUCKET=pg-operator-testing logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ export IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ export IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ export PGOV1_TAG=1.4.0 logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ PGOV1_TAG=1.4.0 logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ export PGOV1_VER=14 logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ PGOV1_VER=14 logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | ++++ which gdate logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | ++++ which date logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ date=/usr/bin/date logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | ++++ which gsed logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | ++++ which sed logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ sed=/usr/bin/sed logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ command -v oc logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | ++ oc get projects logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ get_psql_user_pass start-from-backup-source-pguser-postgres logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ local secret_name=start-from-backup-source-pguser-postgres logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ kubectl -n kuttl-test-topical-oarfish get secret/start-from-backup-source-pguser-postgres '--template={{.data.password | base64decode}}' logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ get_psql_user_host start-from-backup-source-pguser-postgres logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ local secret_name=start-from-backup-source-pguser-postgres logger.go:42: 19:40:50 | start-from-backup/3-read-from-primary | +++ kubectl -n kuttl-test-topical-oarfish get secret/start-from-backup-source-pguser-postgres '--template={{.data.host | base64decode }}' logger.go:42: 19:40:51 | start-from-backup/3-read-from-primary | ++ run_psql_local '\c myapp \\\ SELECT * from myApp;' postgres:dUieZDm7774QUDAeEkacTgv4@start-from-backup-source-primary.kuttl-test-topical-oarfish.svc logger.go:42: 19:40:51 | start-from-backup/3-read-from-primary | ++ local 'command=\c myapp \\\ SELECT * from myApp;' logger.go:42: 19:40:51 | start-from-backup/3-read-from-primary | ++ local uri=postgres:dUieZDm7774QUDAeEkacTgv4@start-from-backup-source-primary.kuttl-test-topical-oarfish.svc logger.go:42: 19:40:51 | start-from-backup/3-read-from-primary | ++ local driver=postgres logger.go:42: 19:40:51 | start-from-backup/3-read-from-primary | +++ get_client_pod logger.go:42: 19:40:51 | start-from-backup/3-read-from-primary | +++ kubectl -n kuttl-test-topical-oarfish get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 19:40:51 | start-from-backup/3-read-from-primary | ++ kubectl -n kuttl-test-topical-oarfish exec pg-client-84d6c45668-g89hv -- bash -c 'printf '\''\c myapp \\\ SELECT * from myApp;\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://'\''postgres:dUieZDm7774QUDAeEkacTgv4@start-from-backup-source-primary.kuttl-test-topical-oarfish.svc'\''' logger.go:42: 19:40:52 | start-from-backup/3-read-from-primary | + data=' 100500 logger.go:42: 19:40:52 | start-from-backup/3-read-from-primary | 100501' logger.go:42: 19:40:52 | start-from-backup/3-read-from-primary | + kubectl create configmap -n kuttl-test-topical-oarfish 03-read-from-primary '--from-literal=data= 100500 logger.go:42: 19:40:52 | start-from-backup/3-read-from-primary | 100501' logger.go:42: 19:40:52 | start-from-backup/3-read-from-primary | configmap/03-read-from-primary created logger.go:42: 19:40:53 | start-from-backup/3-read-from-primary | test step completed 3-read-from-primary logger.go:42: 19:40:53 | start-from-backup/4-create-backup-s3 | starting test step 4-create-backup-s3 logger.go:42: 19:40:53 | start-from-backup/4-create-backup-s3 | PerconaPGBackup:kuttl-test-topical-oarfish/start-from-backup-source-full-s3 created logger.go:42: 19:44:15 | start-from-backup/4-create-backup-s3 | test step completed 4-create-backup-s3 logger.go:42: 19:44:15 | start-from-backup/5-delete-source-cluster | starting test step 5-delete-source-cluster logger.go:42: 19:44:24 | start-from-backup/5-delete-source-cluster | test step completed 5-delete-source-cluster logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | starting test step 6-create-cluster logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | running command: [sh -c set -o errexit set -o xtrace source ../../functions get_cr "start-from-backup" ${RANDOM}| kubectl -n "${NAMESPACE}" apply -f -] logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | + source ../../functions logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ realpath ../../.. logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | ++ CERT_MANAGER_VER=1.17.1 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | ++++ pwd logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/tests/start-from-backup logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | ++ test_name=start-from-backup logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/vars.sh logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/deploy logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/deploy logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/conf logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/conf logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ export TEMP_DIR=/tmp/kuttl/pg/start-from-backup logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ TEMP_DIR=/tmp/kuttl/pg/start-from-backup logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ export GIT_BRANCH=PR-1212 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ GIT_BRANCH=PR-1212 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ export VERSION=PR-1212-7085fc864 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ VERSION=PR-1212-7085fc864 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1212-7085fc864 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1212-7085fc864 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ export PG_VER=17 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ PG_VER=17 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ export IMAGE_POSTGRESQL=ademajazzserve/postgres:11 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ IMAGE_POSTGRESQL=ademajazzserve/postgres:11 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ export BUCKET=pg-operator-testing logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ BUCKET=pg-operator-testing logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ export IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ export IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ export PGOV1_TAG=1.4.0 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ PGOV1_TAG=1.4.0 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ export PGOV1_VER=14 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ PGOV1_VER=14 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | ++++ which gdate logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | ++++ which date logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ date=/usr/bin/date logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | ++++ which gsed logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | ++++ which sed logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ sed=/usr/bin/sed logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | +++ command -v oc logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | ++ oc get projects logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | + get_cr start-from-backup 18948 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | + local cr_name=start-from-backup logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | + '[' -z start-from-backup ']' logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | + local repo_path=18948 logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | + yq eval ' logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | .metadata.name = "start-from-backup" | logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | .metadata.labels = {"e2e":"start-from-backup"} | logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | .spec.postgresVersion = 17 | logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | .spec.users += [{"name":"postgres","password":{"type":"AlphaNumeric"}}] | logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | .spec.users += [{"name":"start-from-backup","password":{"type":"AlphaNumeric"}}] | logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | .spec.image = "ademajazzserve/postgres:11" | logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | .spec.initContainer.image = "perconalab/percona-postgresql-operator:PR-1212-7085fc864" | logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | .spec.backups.pgbackrest.image = "perconalab/percona-postgresql-operator:main-pgbackrest17" | logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | .spec.proxy.pgBouncer.image = "perconalab/percona-postgresql-operator:main-pgbouncer17" | logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | .spec.pmm.image = "perconalab/pmm-client:dev-latest" | logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | .spec.pmm.secret = "start-from-backup-pmm-secret" | logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | .spec.pmm.customClusterName = "start-from-backup-pmm-custom-name" | logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | .spec.pmm.postgresParams = "--environment=dev-postgres" logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | ' /mnt/jenkins/workspace/cloud-pg-operator_PR-1212/deploy/cr.yaml logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | + kubectl -n kuttl-test-topical-oarfish apply -f - logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | + [[ -n '' ]] logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | + case $test_name in logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | + yq eval -i ' logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | .spec.backups.pgbackrest.configuration = [{"secret":{"name":"start-from-backup-pgbackrest-secrets"}}] | logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | .spec.backups.pgbackrest.manual.repoName = "repo1" | logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | .spec.backups.pgbackrest.manual.options = ["--type=full"] | logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | .spec.backups.pgbackrest.global.repo1-path = "/backrestrepo/postgres-operator/18948/repo1" | logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | .spec.backups.pgbackrest.repos = [{"name":"repo1","s3":{"bucket":"pg-operator-testing","endpoint":"s3.amazonaws.com","region":"us-east-1"}}] logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | ' /tmp/kuttl/pg/start-from-backup/cr.yaml logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | + [[ start-from-backup == \d\e\m\a\n\d\-\b\a\c\k\u\p ]] logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | + [[ start-from-backup == \s\t\a\r\t\-\f\r\o\m\-\b\a\c\k\u\p ]] logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | + [[ ! start-from-backup =~ source ]] logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | + yq eval -i ' logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | .spec.dataSource.pgbackrest.configuration = [{"secret":{"name":"start-from-backup-pgbackrest-secrets"}}] | logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | .spec.dataSource.pgbackrest.stanza = "db" | logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | .spec.dataSource.pgbackrest.global.repo1-path = "/backrestrepo/postgres-operator/demand-backup-ppg17/repo1" | logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | .spec.dataSource.pgbackrest.repo = {"name":"repo1","s3":{"bucket":"pg-operator-testing","endpoint":"s3.amazonaws.com","region":"us-east-1"}} logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | ' /tmp/kuttl/pg/start-from-backup/cr.yaml logger.go:42: 19:44:24 | start-from-backup/6-create-cluster | + cat /tmp/kuttl/pg/start-from-backup/cr.yaml logger.go:42: 19:44:25 | start-from-backup/6-create-cluster | perconapgcluster.pgv2.percona.com/start-from-backup created logger.go:42: 19:48:06 | start-from-backup/6-create-cluster | test step completed 6-create-cluster logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | starting test step 7-read-from-primary logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | running command: [sh -c set -o errexit set -o xtrace source ../../functions data=$(run_psql_local '\c myapp \\\ SELECT * from myApp;' "postgres:$(get_psql_user_pass start-from-backup-pguser-postgres)@$(get_psql_user_host start-from-backup-pguser-postgres)") kubectl create configmap -n "${NAMESPACE}" 07-read-from-primary --from-literal=data="${data}"] logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | + source ../../functions logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ realpath ../../.. logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212 logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | ++ CERT_MANAGER_VER=1.17.1 logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | ++++ pwd logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/tests/start-from-backup logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | ++ test_name=start-from-backup logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/vars.sh logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212 logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212 logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/deploy logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/deploy logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/conf logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/conf logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ export TEMP_DIR=/tmp/kuttl/pg/start-from-backup logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ TEMP_DIR=/tmp/kuttl/pg/start-from-backup logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ export GIT_BRANCH=PR-1212 logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ GIT_BRANCH=PR-1212 logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ export VERSION=PR-1212-7085fc864 logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ VERSION=PR-1212-7085fc864 logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1212-7085fc864 logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1212-7085fc864 logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ export PG_VER=17 logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ PG_VER=17 logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ export IMAGE_POSTGRESQL=ademajazzserve/postgres:11 logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ IMAGE_POSTGRESQL=ademajazzserve/postgres:11 logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ export BUCKET=pg-operator-testing logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ BUCKET=pg-operator-testing logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ export IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ export IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ export PGOV1_TAG=1.4.0 logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ PGOV1_TAG=1.4.0 logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ export PGOV1_VER=14 logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ PGOV1_VER=14 logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | ++++ which gdate logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | ++++ which date logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ date=/usr/bin/date logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | ++++ which gsed logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | ++++ which sed logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ sed=/usr/bin/sed logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ command -v oc logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | ++ oc get projects logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ get_psql_user_pass start-from-backup-pguser-postgres logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ local secret_name=start-from-backup-pguser-postgres logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ kubectl -n kuttl-test-topical-oarfish get secret/start-from-backup-pguser-postgres '--template={{.data.password | base64decode}}' logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ get_psql_user_host start-from-backup-pguser-postgres logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ local secret_name=start-from-backup-pguser-postgres logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ kubectl -n kuttl-test-topical-oarfish get secret/start-from-backup-pguser-postgres '--template={{.data.host | base64decode }}' logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | ++ run_psql_local '\c myapp \\\ SELECT * from myApp;' postgres:Q4rpaSspb1T6WKHPc7gxMWkQ@start-from-backup-primary.kuttl-test-topical-oarfish.svc logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | ++ local 'command=\c myapp \\\ SELECT * from myApp;' logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | ++ local uri=postgres:Q4rpaSspb1T6WKHPc7gxMWkQ@start-from-backup-primary.kuttl-test-topical-oarfish.svc logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | ++ local driver=postgres logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ get_client_pod logger.go:42: 19:48:06 | start-from-backup/7-read-from-primary | +++ kubectl -n kuttl-test-topical-oarfish get pods --selector=name=pg-client -o 'jsonpath={.items[].metadata.name}' logger.go:42: 19:48:07 | start-from-backup/7-read-from-primary | ++ kubectl -n kuttl-test-topical-oarfish exec pg-client-84d6c45668-g89hv -- bash -c 'printf '\''\c myapp \\\ SELECT * from myApp;\n'\'' | psql -v ON_ERROR_STOP=1 -t -q postgres://'\''postgres:Q4rpaSspb1T6WKHPc7gxMWkQ@start-from-backup-primary.kuttl-test-topical-oarfish.svc'\''' logger.go:42: 19:48:08 | start-from-backup/7-read-from-primary | + data=' 100500 logger.go:42: 19:48:08 | start-from-backup/7-read-from-primary | 100501' logger.go:42: 19:48:08 | start-from-backup/7-read-from-primary | + kubectl create configmap -n kuttl-test-topical-oarfish 07-read-from-primary '--from-literal=data= 100500 logger.go:42: 19:48:08 | start-from-backup/7-read-from-primary | 100501' logger.go:42: 19:48:08 | start-from-backup/7-read-from-primary | configmap/07-read-from-primary created logger.go:42: 19:48:09 | start-from-backup/7-read-from-primary | test step completed 7-read-from-primary logger.go:42: 19:48:09 | start-from-backup/99-remove-cluster-gracefully | starting test step 99-remove-cluster-gracefully logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | running command: [sh -c set -o errexit set -o xtrace source ../../functions remove_all_finalizers destroy_operator] logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | + source ../../functions logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ realpath ../../.. logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212 logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | ++ CERT_MANAGER_VER=1.17.1 logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | ++++ pwd logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ basename /mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/tests/start-from-backup logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | ++ test_name=start-from-backup logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | ++ source /mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/vars.sh logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212 logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212 logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/deploy logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/deploy logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/conf logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/e2e-tests/conf logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ export TEMP_DIR=/tmp/kuttl/pg/start-from-backup logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ TEMP_DIR=/tmp/kuttl/pg/start-from-backup logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ export GIT_BRANCH=PR-1212 logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ GIT_BRANCH=PR-1212 logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ export VERSION=PR-1212-7085fc864 logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ VERSION=PR-1212-7085fc864 logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ export IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ IMAGE_BASE=perconalab/percona-postgresql-operator logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ export IMAGE=perconalab/percona-postgresql-operator:PR-1212-7085fc864 logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ IMAGE=perconalab/percona-postgresql-operator:PR-1212-7085fc864 logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ export PG_VER=17 logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ PG_VER=17 logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ export IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ IMAGE_PGBOUNCER=perconalab/percona-postgresql-operator:main-pgbouncer17 logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ export IMAGE_POSTGRESQL=ademajazzserve/postgres:11 logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ IMAGE_POSTGRESQL=ademajazzserve/postgres:11 logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ export IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ IMAGE_BACKREST=perconalab/percona-postgresql-operator:main-pgbackrest17 logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ export IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ IMAGE_UPGRADE=perconalab/percona-postgresql-operator:main-upgrade logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ export BUCKET=pg-operator-testing logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ BUCKET=pg-operator-testing logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ export PMM_SERVER_VERSION=9.9.9 logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ PMM_SERVER_VERSION=9.9.9 logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:dev-latest logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ export IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ IMAGE_PMM3_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ export IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ IMAGE_PMM3_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ export PGOV1_TAG=1.4.0 logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ PGOV1_TAG=1.4.0 logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ export PGOV1_VER=14 logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ PGOV1_VER=14 logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | ++++ which gdate logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | which: no gdate in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | ++++ which date logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ date=/usr/bin/date logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | ++++ which gsed logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | which: no gsed in (/mnt/jenkins/workspace/cloud-pg-operator_PR-1212/bin/:/home/ec2-user/.krew/bin:/usr/local/bin:/usr/bin) logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | ++++ which sed logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ sed=/usr/bin/sed logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | +++ command -v oc logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | ++ oc get projects logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | + remove_all_finalizers logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | + resource_types=("pg-restore" "pg-backup" "pg") logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | + for resource in '"${resource_types[@]}"' logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | + echo 'removing all finalizers for pg-restore resources' logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | removing all finalizers for pg-restore resources logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | + kubectl -n kuttl-test-topical-oarfish get pg-restore -o json logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | + IFS= logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | + read -r name logger.go:42: 19:48:16 | start-from-backup/99-remove-cluster-gracefully | + jq '.items[] | .metadata.name' -r logger.go:42: 19:48:17 | start-from-backup/99-remove-cluster-gracefully | + kubectl -n kuttl-test-topical-oarfish delete pg-restore start-from-backup-bootstrap --wait=0 logger.go:42: 19:48:17 | start-from-backup/99-remove-cluster-gracefully | perconapgrestore.pgv2.percona.com "start-from-backup-bootstrap" deleted logger.go:42: 19:48:17 | start-from-backup/99-remove-cluster-gracefully | ++ yq '.metadata.finalizers | length' logger.go:42: 19:48:17 | start-from-backup/99-remove-cluster-gracefully | ++ kubectl -n kuttl-test-topical-oarfish get pg-restore start-from-backup-bootstrap -o yaml logger.go:42: 19:48:18 | start-from-backup/99-remove-cluster-gracefully | Error from server (NotFound): perconapgrestores.pgv2.percona.com "start-from-backup-bootstrap" not found logger.go:42: 19:48:18 | start-from-backup/99-remove-cluster-gracefully | + [[ 0 == \0 ]] logger.go:42: 19:48:18 | start-from-backup/99-remove-cluster-gracefully | + continue logger.go:42: 19:48:18 | start-from-backup/99-remove-cluster-gracefully | + IFS= logger.go:42: 19:48:18 | start-from-backup/99-remove-cluster-gracefully | + read -r name logger.go:42: 19:48:18 | start-from-backup/99-remove-cluster-gracefully | + for resource in '"${resource_types[@]}"' logger.go:42: 19:48:18 | start-from-backup/99-remove-cluster-gracefully | + echo 'removing all finalizers for pg-backup resources' logger.go:42: 19:48:18 | start-from-backup/99-remove-cluster-gracefully | removing all finalizers for pg-backup resources logger.go:42: 19:48:18 | start-from-backup/99-remove-cluster-gracefully | + kubectl -n kuttl-test-topical-oarfish get pg-backup -o json logger.go:42: 19:48:18 | start-from-backup/99-remove-cluster-gracefully | + jq '.items[] | .metadata.name' -r logger.go:42: 19:48:18 | start-from-backup/99-remove-cluster-gracefully | + IFS= logger.go:42: 19:48:18 | start-from-backup/99-remove-cluster-gracefully | + read -r name logger.go:42: 19:48:18 | start-from-backup/99-remove-cluster-gracefully | + kubectl -n kuttl-test-topical-oarfish delete pg-backup start-from-backup-backup-fcqt-jzpxj --wait=0 logger.go:42: 19:48:18 | start-from-backup/99-remove-cluster-gracefully | perconapgbackup.pgv2.percona.com "start-from-backup-backup-fcqt-jzpxj" deleted logger.go:42: 19:48:18 | start-from-backup/99-remove-cluster-gracefully | ++ kubectl -n kuttl-test-topical-oarfish get pg-backup start-from-backup-backup-fcqt-jzpxj -o yaml logger.go:42: 19:48:18 | start-from-backup/99-remove-cluster-gracefully | ++ yq '.metadata.finalizers | length' logger.go:42: 19:48:19 | start-from-backup/99-remove-cluster-gracefully | Error from server (NotFound): perconapgbackups.pgv2.percona.com "start-from-backup-backup-fcqt-jzpxj" not found logger.go:42: 19:48:19 | start-from-backup/99-remove-cluster-gracefully | + [[ 0 == \0 ]] logger.go:42: 19:48:19 | start-from-backup/99-remove-cluster-gracefully | + continue logger.go:42: 19:48:19 | start-from-backup/99-remove-cluster-gracefully | + IFS= logger.go:42: 19:48:19 | start-from-backup/99-remove-cluster-gracefully | + read -r name logger.go:42: 19:48:19 | start-from-backup/99-remove-cluster-gracefully | + kubectl -n kuttl-test-topical-oarfish delete pg-backup start-from-backup-source-backup-djg9-66tg4 --wait=0 logger.go:42: 19:48:19 | start-from-backup/99-remove-cluster-gracefully | perconapgbackup.pgv2.percona.com "start-from-backup-source-backup-djg9-66tg4" deleted logger.go:42: 19:48:19 | start-from-backup/99-remove-cluster-gracefully | ++ yq '.metadata.finalizers | length' logger.go:42: 19:48:19 | start-from-backup/99-remove-cluster-gracefully | ++ kubectl -n kuttl-test-topical-oarfish get pg-backup start-from-backup-source-backup-djg9-66tg4 -o yaml logger.go:42: 19:48:20 | start-from-backup/99-remove-cluster-gracefully | Error from server (NotFound): perconapgbackups.pgv2.percona.com "start-from-backup-source-backup-djg9-66tg4" not found logger.go:42: 19:48:20 | start-from-backup/99-remove-cluster-gracefully | + [[ 0 == \0 ]] logger.go:42: 19:48:20 | start-from-backup/99-remove-cluster-gracefully | + continue logger.go:42: 19:48:20 | start-from-backup/99-remove-cluster-gracefully | + IFS= logger.go:42: 19:48:20 | start-from-backup/99-remove-cluster-gracefully | + read -r name logger.go:42: 19:48:20 | start-from-backup/99-remove-cluster-gracefully | + kubectl -n kuttl-test-topical-oarfish delete pg-backup start-from-backup-source-full-s3 --wait=0 logger.go:42: 19:48:20 | start-from-backup/99-remove-cluster-gracefully | perconapgbackup.pgv2.percona.com "start-from-backup-source-full-s3" deleted logger.go:42: 19:48:20 | start-from-backup/99-remove-cluster-gracefully | ++ yq '.metadata.finalizers | length' logger.go:42: 19:48:20 | start-from-backup/99-remove-cluster-gracefully | ++ kubectl -n kuttl-test-topical-oarfish get pg-backup start-from-backup-source-full-s3 -o yaml logger.go:42: 19:48:21 | start-from-backup/99-remove-cluster-gracefully | Error from server (NotFound): perconapgbackups.pgv2.percona.com "start-from-backup-source-full-s3" not found logger.go:42: 19:48:21 | start-from-backup/99-remove-cluster-gracefully | + [[ 0 == \0 ]] logger.go:42: 19:48:21 | start-from-backup/99-remove-cluster-gracefully | + continue logger.go:42: 19:48:21 | start-from-backup/99-remove-cluster-gracefully | + IFS= logger.go:42: 19:48:21 | start-from-backup/99-remove-cluster-gracefully | + read -r name logger.go:42: 19:48:21 | start-from-backup/99-remove-cluster-gracefully | + for resource in '"${resource_types[@]}"' logger.go:42: 19:48:21 | start-from-backup/99-remove-cluster-gracefully | + echo 'removing all finalizers for pg resources' logger.go:42: 19:48:21 | start-from-backup/99-remove-cluster-gracefully | removing all finalizers for pg resources logger.go:42: 19:48:21 | start-from-backup/99-remove-cluster-gracefully | + kubectl -n kuttl-test-topical-oarfish get pg -o json logger.go:42: 19:48:21 | start-from-backup/99-remove-cluster-gracefully | + jq '.items[] | .metadata.name' -r logger.go:42: 19:48:21 | start-from-backup/99-remove-cluster-gracefully | + IFS= logger.go:42: 19:48:21 | start-from-backup/99-remove-cluster-gracefully | + read -r name logger.go:42: 19:48:21 | start-from-backup/99-remove-cluster-gracefully | + destroy_operator logger.go:42: 19:48:21 | start-from-backup/99-remove-cluster-gracefully | + kubectl -n pg-operator delete deployment percona-postgresql-operator --force --grace-period=0 logger.go:42: 19:48:21 | start-from-backup/99-remove-cluster-gracefully | Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. logger.go:42: 19:48:22 | start-from-backup/99-remove-cluster-gracefully | deployment.apps "percona-postgresql-operator" force deleted logger.go:42: 19:48:22 | start-from-backup/99-remove-cluster-gracefully | + [[ -n pg-operator ]] logger.go:42: 19:48:22 | start-from-backup/99-remove-cluster-gracefully | + kubectl delete namespace pg-operator --force --grace-period=0 logger.go:42: 19:48:22 | start-from-backup/99-remove-cluster-gracefully | Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. logger.go:42: 19:48:22 | start-from-backup/99-remove-cluster-gracefully | namespace "pg-operator" force deleted logger.go:42: 19:48:29 | start-from-backup/99-remove-cluster-gracefully | test step completed 99-remove-cluster-gracefully logger.go:42: 19:48:30 | start-from-backup | start-from-backup events from ns kuttl-test-topical-oarfish: logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:44 +0000 UTC Normal Pod pg-client-84d6c45668-g89hv Binding Scheduled Successfully assigned kuttl-test-topical-oarfish/pg-client-84d6c45668-g89hv to gke-jen-pg-1212-7085fc86-default-pool-e7360247-95hk default-scheduler logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:44 +0000 UTC Normal ReplicaSet.apps pg-client-84d6c45668 SuccessfulCreate Created pod: pg-client-84d6c45668-g89hv replicaset-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:44 +0000 UTC Normal Deployment.apps pg-client ScalingReplicaSet Scaled up replica set pg-client-84d6c45668 to 1 deployment-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:45 +0000 UTC Normal Pod pg-client-84d6c45668-g89hv.spec.containers{pg-client} Pulling Pulling image "perconalab/percona-distribution-postgresql:16" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:45 +0000 UTC Normal Pod pg-client-84d6c45668-g89hv.spec.containers{pg-client} Pulled Successfully pulled image "perconalab/percona-distribution-postgresql:16" in 123ms (123ms including waiting). Image size: 319911229 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:45 +0000 UTC Normal Pod pg-client-84d6c45668-g89hv.spec.containers{pg-client} Created Created container: pg-client kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:45 +0000 UTC Normal Pod pg-client-84d6c45668-g89hv.spec.containers{pg-client} Started Started container pg-client kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:47 +0000 UTC Normal Pod start-from-backup-source-patroni-version-check Binding Scheduled Successfully assigned kuttl-test-topical-oarfish/start-from-backup-source-patroni-version-check to gke-jen-pg-1212-7085fc86-default-pool-e7360247-txhw default-scheduler logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:47 +0000 UTC Normal Pod start-from-backup-source-patroni-version-check.spec.containers{patroni-version-check} Pulled Container image "ademajazzserve/postgres:11" already present on machine kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:47 +0000 UTC Normal Pod start-from-backup-source-patroni-version-check.spec.containers{patroni-version-check} Created Created container: patroni-version-check kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:47 +0000 UTC Normal Pod start-from-backup-source-patroni-version-check.spec.containers{patroni-version-check} Started Started container patroni-version-check kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:52 +0000 UTC Normal Pod start-from-backup-source-patroni-version-check.spec.containers{patroni-version-check} Killing Stopping container patroni-version-check kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:53 +0000 UTC Normal PersistentVolumeClaim start-from-backup-source-instance1-bg8r-pgdata WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:54 +0000 UTC Normal PersistentVolumeClaim start-from-backup-source-instance1-52kg-pgdata WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:54 +0000 UTC Normal PersistentVolumeClaim start-from-backup-source-instance1-52kg-pgdata ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:54 +0000 UTC Normal PersistentVolumeClaim start-from-backup-source-instance1-52kg-pgdata Provisioning External provisioner is provisioning volume for claim "kuttl-test-topical-oarfish/start-from-backup-source-instance1-52kg-pgdata" pd.csi.storage.gke.io_gke-68b65f74e6184ab28538-8a06-06eb-vm_6170685c-858a-4534-b150-8e7bf62d24ba logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:54 +0000 UTC Normal StatefulSet.apps start-from-backup-source-instance1-52kg SuccessfulCreate create Pod start-from-backup-source-instance1-52kg-0 in StatefulSet start-from-backup-source-instance1-52kg successful statefulset-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:54 +0000 UTC Normal PersistentVolumeClaim start-from-backup-source-instance1-bg8r-pgdata ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:54 +0000 UTC Normal PersistentVolumeClaim start-from-backup-source-instance1-bg8r-pgdata Provisioning External provisioner is provisioning volume for claim "kuttl-test-topical-oarfish/start-from-backup-source-instance1-bg8r-pgdata" pd.csi.storage.gke.io_gke-68b65f74e6184ab28538-8a06-06eb-vm_6170685c-858a-4534-b150-8e7bf62d24ba logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:54 +0000 UTC Normal StatefulSet.apps start-from-backup-source-instance1-bg8r SuccessfulCreate create Pod start-from-backup-source-instance1-bg8r-0 in StatefulSet start-from-backup-source-instance1-bg8r successful statefulset-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:54 +0000 UTC Normal PersistentVolumeClaim start-from-backup-source-instance1-fzb9-pgdata WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:54 +0000 UTC Normal StatefulSet.apps start-from-backup-source-instance1-fzb9 SuccessfulCreate create Pod start-from-backup-source-instance1-fzb9-0 in StatefulSet start-from-backup-source-instance1-fzb9 successful statefulset-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:55 +0000 UTC Normal PersistentVolumeClaim start-from-backup-source-instance1-fzb9-pgdata ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:55 +0000 UTC Normal PersistentVolumeClaim start-from-backup-source-instance1-fzb9-pgdata Provisioning External provisioner is provisioning volume for claim "kuttl-test-topical-oarfish/start-from-backup-source-instance1-fzb9-pgdata" pd.csi.storage.gke.io_gke-68b65f74e6184ab28538-8a06-06eb-vm_6170685c-858a-4534-b150-8e7bf62d24ba logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:55 +0000 UTC Normal PodDisruptionBudget.policy start-from-backup-source-pgbouncer NoPods No matching pods found controllermanager logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:55 +0000 UTC Normal Deployment.apps start-from-backup-source-pgbouncer ScalingReplicaSet Scaled up replica set start-from-backup-source-pgbouncer-74b88bf448 to 3 deployment-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:55 +0000 UTC Normal Pod start-from-backup-source-repo-host-0 Binding Scheduled Successfully assigned kuttl-test-topical-oarfish/start-from-backup-source-repo-host-0 to gke-jen-pg-1212-7085fc86-default-pool-e7360247-95hk default-scheduler logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:55 +0000 UTC Normal StatefulSet.apps start-from-backup-source-repo-host SuccessfulCreate create Pod start-from-backup-source-repo-host-0 in StatefulSet start-from-backup-source-repo-host successful statefulset-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:55 +0000 UTC Normal PostgresCluster.postgres-operator.crunchydata.com start-from-backup-source RepoHostCreated created pgBackRest repository host StatefulSet/start-from-backup-source-repo-host postgrescluster-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:56 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-9kltg Binding Scheduled Successfully assigned kuttl-test-topical-oarfish/start-from-backup-source-pgbouncer-74b88bf448-9kltg to gke-jen-pg-1212-7085fc86-default-pool-e7360247-95hk default-scheduler logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:56 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-9kltg.spec.containers{pgbouncer} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbouncer17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:56 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-9kltg.spec.containers{pgbouncer} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbouncer17" in 112ms (112ms including waiting). Image size: 82601372 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:56 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-ldzqt Binding Scheduled Successfully assigned kuttl-test-topical-oarfish/start-from-backup-source-pgbouncer-74b88bf448-ldzqt to gke-jen-pg-1212-7085fc86-default-pool-e7360247-txhw default-scheduler logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:56 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-ldzqt.spec.containers{pgbouncer} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbouncer17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:56 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-ldzqt.spec.containers{pgbouncer} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbouncer17" in 131ms (131ms including waiting). Image size: 82601372 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:56 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-ldzqt.spec.containers{pgbouncer} Created Created container: pgbouncer kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:56 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-ldzqt.spec.containers{pgbouncer} Started Started container pgbouncer kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:56 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-ldzqt.spec.containers{pgbouncer-config} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbouncer17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:56 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-mjqf5 Binding Scheduled Successfully assigned kuttl-test-topical-oarfish/start-from-backup-source-pgbouncer-74b88bf448-mjqf5 to gke-jen-pg-1212-7085fc86-default-pool-e7360247-25vj default-scheduler logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:56 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-mjqf5.spec.containers{pgbouncer} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbouncer17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:56 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-mjqf5.spec.containers{pgbouncer} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbouncer17" in 128ms (128ms including waiting). Image size: 82601372 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:56 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-mjqf5.spec.containers{pgbouncer} Created Created container: pgbouncer kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:56 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-mjqf5.spec.containers{pgbouncer} Started Started container pgbouncer kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:56 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-mjqf5.spec.containers{pgbouncer-config} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbouncer17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:56 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-mjqf5.spec.containers{pgbouncer-config} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbouncer17" in 134ms (134ms including waiting). Image size: 82601372 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:56 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-mjqf5.spec.containers{pgbouncer-config} Created Created container: pgbouncer-config kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:56 +0000 UTC Normal ReplicaSet.apps start-from-backup-source-pgbouncer-74b88bf448 SuccessfulCreate Created pod: start-from-backup-source-pgbouncer-74b88bf448-mjqf5 replicaset-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:56 +0000 UTC Normal ReplicaSet.apps start-from-backup-source-pgbouncer-74b88bf448 SuccessfulCreate Created pod: start-from-backup-source-pgbouncer-74b88bf448-ldzqt replicaset-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:56 +0000 UTC Normal ReplicaSet.apps start-from-backup-source-pgbouncer-74b88bf448 SuccessfulCreate Created pod: start-from-backup-source-pgbouncer-74b88bf448-9kltg replicaset-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:56 +0000 UTC Normal Pod start-from-backup-source-repo-host-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:56 +0000 UTC Normal Pod start-from-backup-source-repo-host-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbackrest17" in 113ms (113ms including waiting). Image size: 151981060 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:56 +0000 UTC Normal Pod start-from-backup-source-repo-host-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:56 +0000 UTC Normal Pod start-from-backup-source-repo-host-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:57 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-9kltg.spec.containers{pgbouncer} Created Created container: pgbouncer kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:57 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-9kltg.spec.containers{pgbouncer} Started Started container pgbouncer kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:57 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-9kltg.spec.containers{pgbouncer-config} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbouncer17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:57 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-9kltg.spec.containers{pgbouncer-config} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbouncer17" in 121ms (121ms including waiting). Image size: 82601372 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:57 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-9kltg.spec.containers{pgbouncer-config} Created Created container: pgbouncer-config kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:57 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-9kltg.spec.containers{pgbouncer-config} Started Started container pgbouncer-config kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:57 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-ldzqt.spec.containers{pgbouncer-config} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbouncer17" in 119ms (119ms including waiting). Image size: 82601372 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:57 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-ldzqt.spec.containers{pgbouncer-config} Created Created container: pgbouncer-config kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:57 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-ldzqt.spec.containers{pgbouncer-config} Started Started container pgbouncer-config kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:57 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-mjqf5.spec.containers{pgbouncer-config} Started Started container pgbouncer-config kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:57 +0000 UTC Normal Pod start-from-backup-source-repo-host-0.spec.containers{pgbackrest} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:57 +0000 UTC Normal Pod start-from-backup-source-repo-host-0.spec.containers{pgbackrest} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbackrest17" in 105ms (106ms including waiting). Image size: 151981060 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:57 +0000 UTC Normal Pod start-from-backup-source-repo-host-0.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:57 +0000 UTC Normal Pod start-from-backup-source-repo-host-0.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:57 +0000 UTC Normal Pod start-from-backup-source-repo-host-0.spec.containers{pgbackrest-config} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:57 +0000 UTC Normal Pod start-from-backup-source-repo-host-0.spec.containers{pgbackrest-config} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbackrest17" in 94ms (94ms including waiting). Image size: 151981060 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:57 +0000 UTC Normal Pod start-from-backup-source-repo-host-0.spec.containers{pgbackrest-config} Created Created container: pgbackrest-config kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:57 +0000 UTC Normal Pod start-from-backup-source-repo-host-0.spec.containers{pgbackrest-config} Started Started container pgbackrest-config kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:58 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0 Binding Scheduled Successfully assigned kuttl-test-topical-oarfish/start-from-backup-source-instance1-52kg-0 to gke-jen-pg-1212-7085fc86-default-pool-e7360247-25vj default-scheduler logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:58 +0000 UTC Normal PersistentVolumeClaim start-from-backup-source-instance1-52kg-pgdata ProvisioningSucceeded Successfully provisioned volume pvc-c2d34153-e195-4b3c-bad5-fc317a4fb06c pd.csi.storage.gke.io_gke-68b65f74e6184ab28538-8a06-06eb-vm_6170685c-858a-4534-b150-8e7bf62d24ba logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:58 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0 Binding Scheduled Successfully assigned kuttl-test-topical-oarfish/start-from-backup-source-instance1-bg8r-0 to gke-jen-pg-1212-7085fc86-default-pool-e7360247-95hk default-scheduler logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:58 +0000 UTC Normal PersistentVolumeClaim start-from-backup-source-instance1-bg8r-pgdata ProvisioningSucceeded Successfully provisioned volume pvc-78cc6b26-c800-4a82-bfe5-3b1a48f0fc11 pd.csi.storage.gke.io_gke-68b65f74e6184ab28538-8a06-06eb-vm_6170685c-858a-4534-b150-8e7bf62d24ba logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:58 +0000 UTC Normal PersistentVolumeClaim start-from-backup-source-instance1-fzb9-pgdata ProvisioningSucceeded Successfully provisioned volume pvc-a2856225-ae7d-47e5-96a5-8e38ebde506f pd.csi.storage.gke.io_gke-68b65f74e6184ab28538-8a06-06eb-vm_6170685c-858a-4534-b150-8e7bf62d24ba logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:38:59 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0 Binding Scheduled Successfully assigned kuttl-test-topical-oarfish/start-from-backup-source-instance1-fzb9-0 to gke-jen-pg-1212-7085fc86-default-pool-e7360247-txhw default-scheduler logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:05 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-78cc6b26-c800-4a82-bfe5-3b1a48f0fc11" attachdetach-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:06 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-c2d34153-e195-4b3c-bad5-fc317a4fb06c" attachdetach-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:06 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-a2856225-ae7d-47e5-96a5-8e38ebde506f" attachdetach-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:07 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.initContainers{postgres-startup} Pulling Pulling image "ademajazzserve/postgres:11" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:07 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.initContainers{postgres-startup} Pulled Successfully pulled image "ademajazzserve/postgres:11" in 150ms (150ms including waiting). Image size: 478209084 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:07 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.initContainers{postgres-startup} Created Created container: postgres-startup kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:07 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.initContainers{postgres-startup} Started Started container postgres-startup kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:07 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.initContainers{postgres-startup} Pulling Pulling image "ademajazzserve/postgres:11" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:07 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.initContainers{postgres-startup} Pulled Successfully pulled image "ademajazzserve/postgres:11" in 139ms (139ms including waiting). Image size: 478209084 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:07 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.initContainers{postgres-startup} Created Created container: postgres-startup kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:07 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.initContainers{postgres-startup} Started Started container postgres-startup kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:08 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.initContainers{database-init} Pulling Pulling image "perconalab/percona-postgresql-operator:PR-1212-7085fc864" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:08 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.initContainers{database-init} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:PR-1212-7085fc864" in 127ms (127ms including waiting). Image size: 95747635 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:08 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.initContainers{database-init} Created Created container: database-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:08 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.initContainers{database-init} Pulling Pulling image "perconalab/percona-postgresql-operator:PR-1212-7085fc864" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:08 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.initContainers{database-init} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:PR-1212-7085fc864" in 126ms (126ms including waiting). Image size: 95747635 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:08 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.initContainers{database-init} Created Created container: database-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:08 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.initContainers{database-init} Started Started container database-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:08 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.initContainers{postgres-startup} Pulling Pulling image "ademajazzserve/postgres:11" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:08 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.initContainers{postgres-startup} Pulled Successfully pulled image "ademajazzserve/postgres:11" in 143ms (143ms including waiting). Image size: 478209084 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:08 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.initContainers{postgres-startup} Created Created container: postgres-startup kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:09 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.initContainers{database-init} Started Started container database-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:09 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "ademajazzserve/postgres:11" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:09 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "ademajazzserve/postgres:11" in 125ms (125ms including waiting). Image size: 478209084 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:09 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:09 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "ademajazzserve/postgres:11" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:09 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "ademajazzserve/postgres:11" in 125ms (125ms including waiting). Image size: 478209084 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:09 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:09 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:09 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.initContainers{postgres-startup} Started Started container postgres-startup kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:09 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.initContainers{database-init} Pulling Pulling image "perconalab/percona-postgresql-operator:PR-1212-7085fc864" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:09 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.initContainers{database-init} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:PR-1212-7085fc864" in 130ms (130ms including waiting). Image size: 95747635 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:09 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.initContainers{database-init} Created Created container: database-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:09 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.initContainers{database-init} Started Started container database-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:10 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:10 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.containers{database} Pulling Pulling image "ademajazzserve/postgres:11" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:10 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.containers{database} Pulled Successfully pulled image "ademajazzserve/postgres:11" in 145ms (145ms including waiting). Image size: 478209084 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:10 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.containers{database} Created Created container: database kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:10 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.containers{database} Pulling Pulling image "ademajazzserve/postgres:11" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:10 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.containers{database} Pulled Successfully pulled image "ademajazzserve/postgres:11" in 153ms (153ms including waiting). Image size: 478209084 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:10 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.containers{database} Created Created container: database kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:10 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.containers{database} Started Started container database kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:10 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.containers{replication-cert-copy} Pulling Pulling image "ademajazzserve/postgres:11" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:10 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.containers{replication-cert-copy} Pulled Successfully pulled image "ademajazzserve/postgres:11" in 167ms (167ms including waiting). Image size: 478209084 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:10 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.containers{replication-cert-copy} Created Created container: replication-cert-copy kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:10 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "ademajazzserve/postgres:11" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:10 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "ademajazzserve/postgres:11" in 145ms (145ms including waiting). Image size: 478209084 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:10 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:10 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:11 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.containers{database} Started Started container database kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:11 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.containers{replication-cert-copy} Pulling Pulling image "ademajazzserve/postgres:11" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:11 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.containers{replication-cert-copy} Pulled Successfully pulled image "ademajazzserve/postgres:11" in 147ms (147ms including waiting). Image size: 478209084 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:11 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.containers{replication-cert-copy} Created Created container: replication-cert-copy kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:11 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.containers{replication-cert-copy} Started Started container replication-cert-copy kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:11 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.containers{pgbackrest} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:11 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.containers{pgbackrest} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbackrest17" in 130ms (130ms including waiting). Image size: 151981060 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:11 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:11 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:11 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.containers{pgbackrest-config} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:11 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.containers{replication-cert-copy} Started Started container replication-cert-copy kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:11 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.containers{pgbackrest} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:11 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.containers{pgbackrest} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbackrest17" in 107ms (107ms including waiting). Image size: 151981060 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:11 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:11 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:11 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.containers{pgbackrest-config} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:11 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.containers{database} Pulling Pulling image "ademajazzserve/postgres:11" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:11 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.containers{database} Pulled Successfully pulled image "ademajazzserve/postgres:11" in 143ms (143ms including waiting). Image size: 478209084 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:11 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.containers{database} Created Created container: database kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:11 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.containers{database} Started Started container database kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:11 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.containers{replication-cert-copy} Pulling Pulling image "ademajazzserve/postgres:11" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:11 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.containers{replication-cert-copy} Pulled Successfully pulled image "ademajazzserve/postgres:11" in 144ms (144ms including waiting). Image size: 478209084 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:11 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.containers{replication-cert-copy} Created Created container: replication-cert-copy kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:11 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.containers{replication-cert-copy} Started Started container replication-cert-copy kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:11 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.containers{pgbackrest} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:12 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.containers{pgbackrest} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbackrest17" in 114ms (114ms including waiting). Image size: 151981060 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:12 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:12 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:12 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.containers{pgbackrest-config} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:20 +0000 UTC Normal Pod start-from-backup-source-backup-djg9-8klt9 Binding Scheduled Successfully assigned kuttl-test-topical-oarfish/start-from-backup-source-backup-djg9-8klt9 to gke-jen-pg-1212-7085fc86-default-pool-e7360247-txhw default-scheduler logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:20 +0000 UTC Normal Job.batch start-from-backup-source-backup-djg9 SuccessfulCreate Created pod: start-from-backup-source-backup-djg9-8klt9 job-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:20 +0000 UTC Normal PostgresCluster.postgres-operator.crunchydata.com start-from-backup-source StanzasCreated pgBackRest stanza creation completed successfully postgrescluster-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:21 +0000 UTC Normal Pod start-from-backup-source-backup-djg9-8klt9.spec.initContainers{pgbackrest-init} Pulling Pulling image "perconalab/percona-postgresql-operator:PR-1212-7085fc864" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:21 +0000 UTC Normal Pod start-from-backup-source-backup-djg9-8klt9.spec.initContainers{pgbackrest-init} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:PR-1212-7085fc864" in 151ms (151ms including waiting). Image size: 95747635 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:21 +0000 UTC Normal Pod start-from-backup-source-backup-djg9-8klt9.spec.initContainers{pgbackrest-init} Created Created container: pgbackrest-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:21 +0000 UTC Normal Pod start-from-backup-source-backup-djg9-8klt9.spec.initContainers{pgbackrest-init} Started Started container pgbackrest-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:22 +0000 UTC Normal Pod start-from-backup-source-backup-djg9-8klt9.spec.containers{pgbackrest} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:22 +0000 UTC Normal Pod start-from-backup-source-backup-djg9-8klt9.spec.containers{pgbackrest} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbackrest17" in 146ms (146ms including waiting). Image size: 151981060 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:22 +0000 UTC Normal Pod start-from-backup-source-backup-djg9-8klt9.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:22 +0000 UTC Normal Pod start-from-backup-source-backup-djg9-8klt9.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:39:25 +0000 UTC Warning PostgresCluster.postgres-operator.crunchydata.com start-from-backup-source UnableToCreateStanzas command terminated with exit code 50: 2025-07-04 19:39:25.410 P00 ERROR: [050]: unable to acquire lock on file '/tmp/pgbackrest/db-backup-1.lock': Resource temporarily unavailable HINT: is another pgBackRest process running? 2025-07-04 19:39:25.487 P00 ERROR: [050]: unable to acquire lock on file '/tmp/pgbackrest/db-backup-1.lock': Resource temporarily unavailable HINT: is another pgBackRest process running? postgrescluster-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:40:44 +0000 UTC Normal Job.batch start-from-backup-source-backup-djg9 Completed Job completed job-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:40:58 +0000 UTC Normal Pod start-from-backup-source-backup-h468-fz8vm Binding Scheduled Successfully assigned kuttl-test-topical-oarfish/start-from-backup-source-backup-h468-fz8vm to gke-jen-pg-1212-7085fc86-default-pool-e7360247-txhw default-scheduler logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:40:58 +0000 UTC Normal Pod start-from-backup-source-backup-h468-fz8vm.spec.initContainers{pgbackrest-init} Pulling Pulling image "perconalab/percona-postgresql-operator:PR-1212-7085fc864" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:40:58 +0000 UTC Normal Pod start-from-backup-source-backup-h468-fz8vm.spec.initContainers{pgbackrest-init} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:PR-1212-7085fc864" in 160ms (160ms including waiting). Image size: 95747635 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:40:58 +0000 UTC Normal Pod start-from-backup-source-backup-h468-fz8vm.spec.initContainers{pgbackrest-init} Created Created container: pgbackrest-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:40:58 +0000 UTC Normal Pod start-from-backup-source-backup-h468-fz8vm.spec.initContainers{pgbackrest-init} Started Started container pgbackrest-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:40:58 +0000 UTC Normal Job.batch start-from-backup-source-backup-h468 SuccessfulCreate Created pod: start-from-backup-source-backup-h468-fz8vm job-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:40:59 +0000 UTC Normal Pod start-from-backup-source-backup-h468-fz8vm.spec.containers{pgbackrest} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:40:59 +0000 UTC Normal Pod start-from-backup-source-backup-h468-fz8vm.spec.containers{pgbackrest} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbackrest17" in 120ms (120ms including waiting). Image size: 151981060 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:40:59 +0000 UTC Normal Pod start-from-backup-source-backup-h468-fz8vm.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:40:59 +0000 UTC Normal Pod start-from-backup-source-backup-h468-fz8vm.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:01 +0000 UTC Normal Job.batch start-from-backup-source-backup-h468 Completed Job completed job-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:18 +0000 UTC Normal Pod start-from-backup-source-instance1-52kg-0.spec.containers{database} Killing Stopping container database kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:18 +0000 UTC Normal StatefulSet.apps start-from-backup-source-instance1-52kg SuccessfulDelete delete Pod start-from-backup-source-instance1-52kg-0 in StatefulSet start-from-backup-source-instance1-52kg successful statefulset-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:18 +0000 UTC Normal Pod start-from-backup-source-instance1-fzb9-0.spec.containers{database} Killing Stopping container database kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:18 +0000 UTC Normal StatefulSet.apps start-from-backup-source-instance1-fzb9 SuccessfulDelete delete Pod start-from-backup-source-instance1-fzb9-0 in StatefulSet start-from-backup-source-instance1-fzb9 successful statefulset-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:20 +0000 UTC Normal Pod start-from-backup-source-instance1-bg8r-0.spec.containers{database} Killing Stopping container database kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:20 +0000 UTC Normal StatefulSet.apps start-from-backup-source-instance1-bg8r SuccessfulDelete delete Pod start-from-backup-source-instance1-bg8r-0 in StatefulSet start-from-backup-source-instance1-bg8r successful statefulset-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:24 +0000 UTC Normal PodDisruptionBudget.policy start-from-backup-source-set-instance1 NoPods No matching pods found controllermanager logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:25 +0000 UTC Normal Pod start-from-backup-patroni-version-check Binding Scheduled Successfully assigned kuttl-test-topical-oarfish/start-from-backup-patroni-version-check to gke-jen-pg-1212-7085fc86-default-pool-e7360247-95hk default-scheduler logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:25 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-9kltg.spec.containers{pgbouncer} Killing Stopping container pgbouncer kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:25 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-9kltg.spec.containers{pgbouncer-config} Killing Stopping container pgbouncer-config kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:25 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-ldzqt.spec.containers{pgbouncer} Killing Stopping container pgbouncer kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:25 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-ldzqt.spec.containers{pgbouncer-config} Killing Stopping container pgbouncer-config kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:25 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-mjqf5.spec.containers{pgbouncer} Killing Stopping container pgbouncer kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:25 +0000 UTC Normal Pod start-from-backup-source-pgbouncer-74b88bf448-mjqf5.spec.containers{pgbouncer-config} Killing Stopping container pgbouncer-config kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:25 +0000 UTC Normal Pod start-from-backup-source-repo-host-0.spec.containers{pgbackrest} Killing Stopping container pgbackrest kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:25 +0000 UTC Normal Pod start-from-backup-source-repo-host-0.spec.containers{pgbackrest-config} Killing Stopping container pgbackrest-config kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:26 +0000 UTC Normal Pod start-from-backup-patroni-version-check.spec.containers{patroni-version-check} Pulled Container image "ademajazzserve/postgres:11" already present on machine kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:26 +0000 UTC Normal Pod start-from-backup-patroni-version-check.spec.containers{patroni-version-check} Created Created container: patroni-version-check kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:26 +0000 UTC Normal Pod start-from-backup-patroni-version-check.spec.containers{patroni-version-check} Started Started container patroni-version-check kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:32 +0000 UTC Normal PersistentVolumeClaim start-from-backup-instance1-7f74-pgdata WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:32 +0000 UTC Normal Pod start-from-backup-patroni-version-check.spec.containers{patroni-version-check} Killing Stopping container patroni-version-check kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:33 +0000 UTC Normal PersistentVolumeClaim start-from-backup-instance1-7f74-pgdata ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:33 +0000 UTC Normal PersistentVolumeClaim start-from-backup-instance1-7f74-pgdata Provisioning External provisioner is provisioning volume for claim "kuttl-test-topical-oarfish/start-from-backup-instance1-7f74-pgdata" pd.csi.storage.gke.io_gke-68b65f74e6184ab28538-8a06-06eb-vm_6170685c-858a-4534-b150-8e7bf62d24ba logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:33 +0000 UTC Normal Job.batch start-from-backup-pgbackrest-restore SuccessfulCreate Created pod: start-from-backup-pgbackrest-restore-6ngv4 job-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:36 +0000 UTC Normal PersistentVolumeClaim start-from-backup-instance1-7f74-pgdata ProvisioningSucceeded Successfully provisioned volume pvc-ed367c63-51c9-47c8-9289-6dc38d47e625 pd.csi.storage.gke.io_gke-68b65f74e6184ab28538-8a06-06eb-vm_6170685c-858a-4534-b150-8e7bf62d24ba logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:37 +0000 UTC Normal Pod start-from-backup-pgbackrest-restore-6ngv4 Binding Scheduled Successfully assigned kuttl-test-topical-oarfish/start-from-backup-pgbackrest-restore-6ngv4 to gke-jen-pg-1212-7085fc86-default-pool-e7360247-txhw default-scheduler logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:44 +0000 UTC Normal Pod start-from-backup-pgbackrest-restore-6ngv4 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-ed367c63-51c9-47c8-9289-6dc38d47e625" attachdetach-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:46 +0000 UTC Normal Pod start-from-backup-pgbackrest-restore-6ngv4.spec.initContainers{nss-wrapper-init} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:46 +0000 UTC Normal Pod start-from-backup-pgbackrest-restore-6ngv4.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbackrest17" in 131ms (131ms including waiting). Image size: 151981060 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:46 +0000 UTC Normal Pod start-from-backup-pgbackrest-restore-6ngv4.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:46 +0000 UTC Normal Pod start-from-backup-pgbackrest-restore-6ngv4.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:46 +0000 UTC Normal Pod start-from-backup-pgbackrest-restore-6ngv4.spec.containers{pgbackrest-restore} Pulling Pulling image "ademajazzserve/postgres:11" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:46 +0000 UTC Normal Pod start-from-backup-pgbackrest-restore-6ngv4.spec.containers{pgbackrest-restore} Pulled Successfully pulled image "ademajazzserve/postgres:11" in 138ms (138ms including waiting). Image size: 478209084 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:46 +0000 UTC Normal Pod start-from-backup-pgbackrest-restore-6ngv4.spec.containers{pgbackrest-restore} Created Created container: pgbackrest-restore kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:44:46 +0000 UTC Normal Pod start-from-backup-pgbackrest-restore-6ngv4.spec.containers{pgbackrest-restore} Started Started container pgbackrest-restore kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:54 +0000 UTC Normal Job.batch start-from-backup-pgbackrest-restore Completed Job completed job-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:55 +0000 UTC Normal Pod start-from-backup-instance1-7f74-0 Binding Scheduled Successfully assigned kuttl-test-topical-oarfish/start-from-backup-instance1-7f74-0 to gke-jen-pg-1212-7085fc86-default-pool-e7360247-95hk default-scheduler logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:55 +0000 UTC Warning Pod start-from-backup-instance1-7f74-0 FailedMount MountVolume.SetUp failed for volume "pgbackrest-config" : secret "start-from-backup-pgbackrest" not found kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:55 +0000 UTC Normal StatefulSet.apps start-from-backup-instance1-7f74 SuccessfulCreate create Pod start-from-backup-instance1-7f74-0 in StatefulSet start-from-backup-instance1-7f74 successful statefulset-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:55 +0000 UTC Normal PersistentVolumeClaim start-from-backup-instance1-qbrp-pgdata WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:55 +0000 UTC Normal PersistentVolumeClaim start-from-backup-instance1-qd5d-pgdata WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:56 +0000 UTC Normal Pod start-from-backup-repo-host-0 Binding Scheduled Successfully assigned kuttl-test-topical-oarfish/start-from-backup-repo-host-0 to gke-jen-pg-1212-7085fc86-default-pool-e7360247-txhw default-scheduler logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:56 +0000 UTC Normal StatefulSet.apps start-from-backup-repo-host SuccessfulCreate create Pod start-from-backup-repo-host-0 in StatefulSet start-from-backup-repo-host successful statefulset-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:56 +0000 UTC Normal PostgresCluster.postgres-operator.crunchydata.com start-from-backup RepoHostCreated created pgBackRest repository host StatefulSet/start-from-backup-repo-host postgrescluster-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:57 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-4jbnw Binding Scheduled Successfully assigned kuttl-test-topical-oarfish/start-from-backup-pgbouncer-5bc5f8d79d-4jbnw to gke-jen-pg-1212-7085fc86-default-pool-e7360247-txhw default-scheduler logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:57 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-4jbnw.spec.containers{pgbouncer} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbouncer17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:57 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-plck7 Binding Scheduled Successfully assigned kuttl-test-topical-oarfish/start-from-backup-pgbouncer-5bc5f8d79d-plck7 to gke-jen-pg-1212-7085fc86-default-pool-e7360247-95hk default-scheduler logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:57 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-plck7.spec.containers{pgbouncer} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbouncer17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:57 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-q5hnz Binding Scheduled Successfully assigned kuttl-test-topical-oarfish/start-from-backup-pgbouncer-5bc5f8d79d-q5hnz to gke-jen-pg-1212-7085fc86-default-pool-e7360247-25vj default-scheduler logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:57 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-q5hnz.spec.containers{pgbouncer} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbouncer17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:57 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-q5hnz.spec.containers{pgbouncer} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbouncer17" in 115ms (115ms including waiting). Image size: 82601372 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:57 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-q5hnz.spec.containers{pgbouncer} Created Created container: pgbouncer kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:57 +0000 UTC Normal ReplicaSet.apps start-from-backup-pgbouncer-5bc5f8d79d SuccessfulCreate Created pod: start-from-backup-pgbouncer-5bc5f8d79d-q5hnz replicaset-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:57 +0000 UTC Normal ReplicaSet.apps start-from-backup-pgbouncer-5bc5f8d79d SuccessfulCreate Created pod: start-from-backup-pgbouncer-5bc5f8d79d-4jbnw replicaset-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:57 +0000 UTC Normal ReplicaSet.apps start-from-backup-pgbouncer-5bc5f8d79d SuccessfulCreate Created pod: start-from-backup-pgbouncer-5bc5f8d79d-plck7 replicaset-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:57 +0000 UTC Normal Deployment.apps start-from-backup-pgbouncer ScalingReplicaSet Scaled up replica set start-from-backup-pgbouncer-5bc5f8d79d to 3 deployment-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:57 +0000 UTC Normal PodDisruptionBudget.policy start-from-backup-pgbouncer NoPods No matching pods found controllermanager logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:57 +0000 UTC Normal Pod start-from-backup-repo-host-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:57 +0000 UTC Normal Pod start-from-backup-repo-host-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbackrest17" in 108ms (108ms including waiting). Image size: 151981060 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:57 +0000 UTC Normal Pod start-from-backup-repo-host-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:57 +0000 UTC Normal Pod start-from-backup-repo-host-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-4jbnw.spec.containers{pgbouncer} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbouncer17" in 127ms (128ms including waiting). Image size: 82601372 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-4jbnw.spec.containers{pgbouncer} Created Created container: pgbouncer kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-4jbnw.spec.containers{pgbouncer} Started Started container pgbouncer kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-4jbnw.spec.containers{pgbouncer-config} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbouncer17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-4jbnw.spec.containers{pgbouncer-config} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbouncer17" in 109ms (109ms including waiting). Image size: 82601372 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-4jbnw.spec.containers{pgbouncer-config} Created Created container: pgbouncer-config kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-4jbnw.spec.containers{pgbouncer-config} Started Started container pgbouncer-config kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-plck7.spec.containers{pgbouncer} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbouncer17" in 133ms (133ms including waiting). Image size: 82601372 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-plck7.spec.containers{pgbouncer} Created Created container: pgbouncer kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-plck7.spec.containers{pgbouncer} Started Started container pgbouncer kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-plck7.spec.containers{pgbouncer-config} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbouncer17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-plck7.spec.containers{pgbouncer-config} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbouncer17" in 110ms (110ms including waiting). Image size: 82601372 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-plck7.spec.containers{pgbouncer-config} Created Created container: pgbouncer-config kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-plck7.spec.containers{pgbouncer-config} Started Started container pgbouncer-config kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-q5hnz.spec.containers{pgbouncer} Started Started container pgbouncer kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-q5hnz.spec.containers{pgbouncer-config} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbouncer17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-q5hnz.spec.containers{pgbouncer-config} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbouncer17" in 129ms (129ms including waiting). Image size: 82601372 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-q5hnz.spec.containers{pgbouncer-config} Created Created container: pgbouncer-config kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-q5hnz.spec.containers{pgbouncer-config} Started Started container pgbouncer-config kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-repo-host-0.spec.containers{pgbackrest} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-repo-host-0.spec.containers{pgbackrest} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbackrest17" in 111ms (111ms including waiting). Image size: 151981060 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-repo-host-0.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-repo-host-0.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-repo-host-0.spec.containers{pgbackrest-config} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-repo-host-0.spec.containers{pgbackrest-config} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbackrest17" in 109ms (109ms including waiting). Image size: 151981060 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:58 +0000 UTC Normal Pod start-from-backup-repo-host-0.spec.containers{pgbackrest-config} Created Created container: pgbackrest-config kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:45:59 +0000 UTC Normal Pod start-from-backup-repo-host-0.spec.containers{pgbackrest-config} Started Started container pgbackrest-config kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:10 +0000 UTC Normal Pod start-from-backup-instance1-7f74-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-ed367c63-51c9-47c8-9289-6dc38d47e625" attachdetach-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:11 +0000 UTC Normal Pod start-from-backup-instance1-7f74-0.spec.initContainers{postgres-startup} Pulling Pulling image "ademajazzserve/postgres:11" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:11 +0000 UTC Normal Pod start-from-backup-instance1-7f74-0.spec.initContainers{postgres-startup} Pulled Successfully pulled image "ademajazzserve/postgres:11" in 126ms (126ms including waiting). Image size: 478209084 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:11 +0000 UTC Normal Pod start-from-backup-instance1-7f74-0.spec.initContainers{postgres-startup} Created Created container: postgres-startup kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:11 +0000 UTC Normal Pod start-from-backup-instance1-7f74-0.spec.initContainers{postgres-startup} Started Started container postgres-startup kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:12 +0000 UTC Normal Pod start-from-backup-instance1-7f74-0.spec.initContainers{database-init} Pulling Pulling image "perconalab/percona-postgresql-operator:PR-1212-7085fc864" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:12 +0000 UTC Normal Pod start-from-backup-instance1-7f74-0.spec.initContainers{database-init} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:PR-1212-7085fc864" in 117ms (117ms including waiting). Image size: 95747635 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:12 +0000 UTC Normal Pod start-from-backup-instance1-7f74-0.spec.initContainers{database-init} Created Created container: database-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:12 +0000 UTC Normal Pod start-from-backup-instance1-7f74-0.spec.initContainers{database-init} Started Started container database-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:13 +0000 UTC Normal Pod start-from-backup-instance1-7f74-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "ademajazzserve/postgres:11" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:13 +0000 UTC Normal Pod start-from-backup-instance1-7f74-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "ademajazzserve/postgres:11" in 128ms (128ms including waiting). Image size: 478209084 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:13 +0000 UTC Normal Pod start-from-backup-instance1-7f74-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:13 +0000 UTC Normal Pod start-from-backup-instance1-7f74-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:14 +0000 UTC Normal Pod start-from-backup-instance1-7f74-0.spec.containers{database} Pulling Pulling image "ademajazzserve/postgres:11" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:14 +0000 UTC Normal Pod start-from-backup-instance1-7f74-0.spec.containers{database} Pulled Successfully pulled image "ademajazzserve/postgres:11" in 136ms (136ms including waiting). Image size: 478209084 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:14 +0000 UTC Normal Pod start-from-backup-instance1-7f74-0.spec.containers{database} Created Created container: database kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:14 +0000 UTC Normal Pod start-from-backup-instance1-7f74-0.spec.containers{database} Started Started container database kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:14 +0000 UTC Normal Pod start-from-backup-instance1-7f74-0.spec.containers{replication-cert-copy} Pulling Pulling image "ademajazzserve/postgres:11" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:15 +0000 UTC Normal Pod start-from-backup-instance1-7f74-0.spec.containers{replication-cert-copy} Pulled Successfully pulled image "ademajazzserve/postgres:11" in 118ms (118ms including waiting). Image size: 478209084 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:15 +0000 UTC Normal Pod start-from-backup-instance1-7f74-0.spec.containers{replication-cert-copy} Created Created container: replication-cert-copy kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:15 +0000 UTC Normal Pod start-from-backup-instance1-7f74-0.spec.containers{replication-cert-copy} Started Started container replication-cert-copy kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:15 +0000 UTC Normal Pod start-from-backup-instance1-7f74-0.spec.containers{pgbackrest} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:15 +0000 UTC Normal Pod start-from-backup-instance1-7f74-0.spec.containers{pgbackrest} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbackrest17" in 125ms (125ms including waiting). Image size: 151981060 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:15 +0000 UTC Normal Pod start-from-backup-instance1-7f74-0.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:20 +0000 UTC Normal Pod start-from-backup-backup-fcqt-b6ktj Binding Scheduled Successfully assigned kuttl-test-topical-oarfish/start-from-backup-backup-fcqt-b6ktj to gke-jen-pg-1212-7085fc86-default-pool-e7360247-25vj default-scheduler logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:20 +0000 UTC Normal Job.batch start-from-backup-backup-fcqt SuccessfulCreate Created pod: start-from-backup-backup-fcqt-b6ktj job-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:20 +0000 UTC Normal PostgresCluster.postgres-operator.crunchydata.com start-from-backup StanzasCreated pgBackRest stanza creation completed successfully postgrescluster-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:21 +0000 UTC Normal Pod start-from-backup-backup-fcqt-b6ktj.spec.initContainers{pgbackrest-init} Pulling Pulling image "perconalab/percona-postgresql-operator:PR-1212-7085fc864" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:21 +0000 UTC Normal Pod start-from-backup-backup-fcqt-b6ktj.spec.initContainers{pgbackrest-init} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:PR-1212-7085fc864" in 135ms (135ms including waiting). Image size: 95747635 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:21 +0000 UTC Normal Pod start-from-backup-backup-fcqt-b6ktj.spec.initContainers{pgbackrest-init} Created Created container: pgbackrest-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:21 +0000 UTC Normal Pod start-from-backup-backup-fcqt-b6ktj.spec.initContainers{pgbackrest-init} Started Started container pgbackrest-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:21 +0000 UTC Normal Pod start-from-backup-backup-fcqt-b6ktj.spec.containers{pgbackrest} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:21 +0000 UTC Normal Pod start-from-backup-backup-fcqt-b6ktj.spec.containers{pgbackrest} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbackrest17" in 109ms (109ms including waiting). Image size: 151981060 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:21 +0000 UTC Normal Pod start-from-backup-backup-fcqt-b6ktj.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:22 +0000 UTC Normal Pod start-from-backup-backup-fcqt-b6ktj.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:22 +0000 UTC Normal StatefulSet.apps start-from-backup-instance1-qbrp SuccessfulCreate create Pod start-from-backup-instance1-qbrp-0 in StatefulSet start-from-backup-instance1-qbrp successful statefulset-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:23 +0000 UTC Normal PersistentVolumeClaim start-from-backup-instance1-qbrp-pgdata Provisioning External provisioner is provisioning volume for claim "kuttl-test-topical-oarfish/start-from-backup-instance1-qbrp-pgdata" pd.csi.storage.gke.io_gke-68b65f74e6184ab28538-8a06-06eb-vm_6170685c-858a-4534-b150-8e7bf62d24ba logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:23 +0000 UTC Normal PersistentVolumeClaim start-from-backup-instance1-qbrp-pgdata ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:23 +0000 UTC Normal PersistentVolumeClaim start-from-backup-instance1-qd5d-pgdata Provisioning External provisioner is provisioning volume for claim "kuttl-test-topical-oarfish/start-from-backup-instance1-qd5d-pgdata" pd.csi.storage.gke.io_gke-68b65f74e6184ab28538-8a06-06eb-vm_6170685c-858a-4534-b150-8e7bf62d24ba logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:23 +0000 UTC Normal PersistentVolumeClaim start-from-backup-instance1-qd5d-pgdata ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'pd.csi.storage.gke.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:23 +0000 UTC Normal StatefulSet.apps start-from-backup-instance1-qd5d SuccessfulCreate create Pod start-from-backup-instance1-qd5d-0 in StatefulSet start-from-backup-instance1-qd5d successful statefulset-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:24 +0000 UTC Warning PostgresCluster.postgres-operator.crunchydata.com start-from-backup UnableToCreateStanzas command terminated with exit code 50: 2025-07-04 19:46:24.680 P00 ERROR: [050]: unable to acquire lock on file '/tmp/pgbackrest/db-backup-1.lock': Resource temporarily unavailable HINT: is another pgBackRest process running? 2025-07-04 19:46:24.757 P00 ERROR: [050]: unable to acquire lock on file '/tmp/pgbackrest/db-backup-1.lock': Resource temporarily unavailable HINT: is another pgBackRest process running? postgrescluster-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:26 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0 Binding Scheduled Successfully assigned kuttl-test-topical-oarfish/start-from-backup-instance1-qbrp-0 to gke-jen-pg-1212-7085fc86-default-pool-e7360247-25vj default-scheduler logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:26 +0000 UTC Normal PersistentVolumeClaim start-from-backup-instance1-qbrp-pgdata ProvisioningSucceeded Successfully provisioned volume pvc-267eaca2-1c6b-494c-b21a-1dd0048ce44a pd.csi.storage.gke.io_gke-68b65f74e6184ab28538-8a06-06eb-vm_6170685c-858a-4534-b150-8e7bf62d24ba logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:27 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0 Binding Scheduled Successfully assigned kuttl-test-topical-oarfish/start-from-backup-instance1-qd5d-0 to gke-jen-pg-1212-7085fc86-default-pool-e7360247-txhw default-scheduler logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:27 +0000 UTC Normal PersistentVolumeClaim start-from-backup-instance1-qd5d-pgdata ProvisioningSucceeded Successfully provisioned volume pvc-7c99d080-3c19-44dd-ac0e-5d9ecbc4e395 pd.csi.storage.gke.io_gke-68b65f74e6184ab28538-8a06-06eb-vm_6170685c-858a-4534-b150-8e7bf62d24ba logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:35 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-267eaca2-1c6b-494c-b21a-1dd0048ce44a" attachdetach-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:35 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-7c99d080-3c19-44dd-ac0e-5d9ecbc4e395" attachdetach-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:36 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0.spec.initContainers{postgres-startup} Pulling Pulling image "ademajazzserve/postgres:11" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:36 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0.spec.initContainers{postgres-startup} Pulled Successfully pulled image "ademajazzserve/postgres:11" in 135ms (135ms including waiting). Image size: 478209084 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:36 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0.spec.initContainers{postgres-startup} Created Created container: postgres-startup kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:36 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0.spec.initContainers{postgres-startup} Started Started container postgres-startup kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:36 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0.spec.initContainers{database-init} Pulling Pulling image "perconalab/percona-postgresql-operator:PR-1212-7085fc864" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:36 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0.spec.initContainers{database-init} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:PR-1212-7085fc864" in 121ms (121ms including waiting). Image size: 95747635 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:36 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0.spec.initContainers{postgres-startup} Pulling Pulling image "ademajazzserve/postgres:11" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:37 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0.spec.initContainers{database-init} Created Created container: database-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:37 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0.spec.initContainers{database-init} Started Started container database-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:37 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "ademajazzserve/postgres:11" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:37 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "ademajazzserve/postgres:11" in 140ms (140ms including waiting). Image size: 478209084 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:37 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0.spec.initContainers{postgres-startup} Pulled Successfully pulled image "ademajazzserve/postgres:11" in 135ms (135ms including waiting). Image size: 478209084 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:37 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0.spec.initContainers{postgres-startup} Created Created container: postgres-startup kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:37 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0.spec.initContainers{postgres-startup} Started Started container postgres-startup kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:37 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0.spec.initContainers{database-init} Pulling Pulling image "perconalab/percona-postgresql-operator:PR-1212-7085fc864" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:37 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0.spec.initContainers{database-init} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:PR-1212-7085fc864" in 158ms (158ms including waiting). Image size: 95747635 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:37 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0.spec.initContainers{database-init} Created Created container: database-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:37 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0.spec.initContainers{database-init} Started Started container database-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:38 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:38 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:38 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0.spec.containers{database} Pulling Pulling image "ademajazzserve/postgres:11" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:38 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0.spec.initContainers{nss-wrapper-init} Pulling Pulling image "ademajazzserve/postgres:11" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:38 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0.spec.initContainers{nss-wrapper-init} Pulled Successfully pulled image "ademajazzserve/postgres:11" in 118ms (118ms including waiting). Image size: 478209084 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:38 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0.spec.initContainers{nss-wrapper-init} Created Created container: nss-wrapper-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:38 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0.spec.initContainers{nss-wrapper-init} Started Started container nss-wrapper-init kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:39 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0.spec.containers{database} Pulled Successfully pulled image "ademajazzserve/postgres:11" in 161ms (161ms including waiting). Image size: 478209084 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:39 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0.spec.containers{database} Created Created container: database kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:39 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0.spec.containers{database} Started Started container database kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:39 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0.spec.containers{replication-cert-copy} Pulling Pulling image "ademajazzserve/postgres:11" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:39 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0.spec.containers{replication-cert-copy} Pulled Successfully pulled image "ademajazzserve/postgres:11" in 146ms (146ms including waiting). Image size: 478209084 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:39 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0.spec.containers{replication-cert-copy} Created Created container: replication-cert-copy kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:39 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0.spec.containers{replication-cert-copy} Started Started container replication-cert-copy kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:39 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0.spec.containers{pgbackrest} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:39 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0.spec.containers{pgbackrest} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbackrest17" in 113ms (113ms including waiting). Image size: 151981060 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:39 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:39 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:39 +0000 UTC Normal Pod start-from-backup-instance1-qbrp-0.spec.containers{pgbackrest-config} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:39 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0.spec.containers{database} Pulling Pulling image "ademajazzserve/postgres:11" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:39 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0.spec.containers{database} Pulled Successfully pulled image "ademajazzserve/postgres:11" in 123ms (123ms including waiting). Image size: 478209084 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:39 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0.spec.containers{database} Created Created container: database kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:39 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0.spec.containers{database} Started Started container database kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:39 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0.spec.containers{replication-cert-copy} Pulling Pulling image "ademajazzserve/postgres:11" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:40 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0.spec.containers{replication-cert-copy} Pulled Successfully pulled image "ademajazzserve/postgres:11" in 132ms (132ms including waiting). Image size: 478209084 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:40 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0.spec.containers{replication-cert-copy} Created Created container: replication-cert-copy kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:40 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0.spec.containers{replication-cert-copy} Started Started container replication-cert-copy kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:40 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0.spec.containers{pgbackrest} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:40 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0.spec.containers{pgbackrest} Pulled Successfully pulled image "perconalab/percona-postgresql-operator:main-pgbackrest17" in 190ms (190ms including waiting). Image size: 151981060 bytes. kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:40 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0.spec.containers{pgbackrest} Created Created container: pgbackrest kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:40 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0.spec.containers{pgbackrest} Started Started container pgbackrest kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:46:40 +0000 UTC Normal Pod start-from-backup-instance1-qd5d-0.spec.containers{pgbackrest-config} Pulling Pulling image "perconalab/percona-postgresql-operator:main-pgbackrest17" kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:48:04 +0000 UTC Normal Job.batch start-from-backup-backup-fcqt Completed Job completed job-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:48:10 +0000 UTC Normal StatefulSet.apps start-from-backup-instance1-qbrp SuccessfulDelete delete Pod start-from-backup-instance1-qbrp-0 in StatefulSet start-from-backup-instance1-qbrp successful statefulset-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:48:10 +0000 UTC Normal StatefulSet.apps start-from-backup-instance1-qd5d SuccessfulDelete delete Pod start-from-backup-instance1-qd5d-0 in StatefulSet start-from-backup-instance1-qd5d successful statefulset-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:48:12 +0000 UTC Normal StatefulSet.apps start-from-backup-instance1-7f74 SuccessfulDelete delete Pod start-from-backup-instance1-7f74-0 in StatefulSet start-from-backup-instance1-7f74 successful statefulset-controller logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:48:16 +0000 UTC Normal PodDisruptionBudget.policy start-from-backup-set-instance1 NoPods No matching pods found controllermanager logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:48:17 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-4jbnw.spec.containers{pgbouncer} Killing Stopping container pgbouncer kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:48:17 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-4jbnw.spec.containers{pgbouncer-config} Killing Stopping container pgbouncer-config kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:48:17 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-plck7.spec.containers{pgbouncer} Killing Stopping container pgbouncer kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:48:17 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-plck7.spec.containers{pgbouncer-config} Killing Stopping container pgbouncer-config kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:48:17 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-q5hnz.spec.containers{pgbouncer} Killing Stopping container pgbouncer kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:48:17 +0000 UTC Normal Pod start-from-backup-pgbouncer-5bc5f8d79d-q5hnz.spec.containers{pgbouncer-config} Killing Stopping container pgbouncer-config kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:48:17 +0000 UTC Normal Pod start-from-backup-repo-host-0.spec.containers{pgbackrest} Killing Stopping container pgbackrest kubelet logger.go:42: 19:48:30 | start-from-backup | 2025-07-04 19:48:17 +0000 UTC Normal Pod start-from-backup-repo-host-0.spec.containers{pgbackrest-config} Killing Stopping container pgbackrest-config kubelet logger.go:42: 19:48:30 | start-from-backup | Deleting namespace: kuttl-test-topical-oarfish === NAME kuttl harness.go:403: run tests finished harness.go:510: cleaning up harness.go:567: removing temp folder: "" --- PASS: kuttl (641.55s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/start-from-backup (640.84s) PASS