Log: /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2270/e2e-tests/logs/pitr-physical-backup-source.log Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 + main + create_infra pitr-physical-backup-source-28911 + local ns=pitr-physical-backup-source-28911 + [[ 1 == 1 ]] + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2270/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.q7CkypVUya ++ mktemp + local LAST_ERR=/tmp/tmp.iui3DvdTXL + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2270/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.q7CkypVUya customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.iui3DvdTXL + rm /tmp/tmp.q7CkypVUya /tmp/tmp.iui3DvdTXL + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2270/deploy/crd.yaml ++ grep -v '\-\-\-' grep: warning: stray \ before - grep: warning: stray \ before - + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.oCjMHbesYg ++ mktemp + local LAST_ERR=/tmp/tmp.dwzcCQa9nF + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.oCjMHbesYg + cat /tmp/tmp.dwzcCQa9nF + rm /tmp/tmp.oCjMHbesYg /tmp/tmp.dwzcCQa9nF + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.wAAgCY35Ep ++ mktemp + local LAST_ERR=/tmp/tmp.ytFHRpzvcT + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.wAAgCY35Ep + cat /tmp/tmp.ytFHRpzvcT + rm /tmp/tmp.wAAgCY35Ep /tmp/tmp.ytFHRpzvcT + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbs" + kubectl patch perconaservermongodbs.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbs" + : + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.bDzQo9SLO0 ++ mktemp + local LAST_ERR=/tmp/tmp.7SsfUBZ6Xn + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.bDzQo9SLO0 + cat /tmp/tmp.7SsfUBZ6Xn + rm /tmp/tmp.bDzQo9SLO0 /tmp/tmp.7SsfUBZ6Xn + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2270/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.ZmDMDwBwXp ++ mktemp + local LAST_ERR=/tmp/tmp.Xt3DAUcrlc + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2270/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ZmDMDwBwXp clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.Xt3DAUcrlc + rm /tmp/tmp.ZmDMDwBwXp /tmp/tmp.Xt3DAUcrlc + return 0 + check_crd_for_deletion PR-2270-9169a6cb + local git_tag=PR-2270-9169a6cb ++ curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/PR-2270-9169a6cb/deploy/crd.yaml ++ yq eval .metadata.name ++ /usr/sbin/sed ':a;N;$!ba;s/\n/ /g' ++ /usr/sbin/sed s/---//g + for crd_name in $(curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/${git_tag}/deploy/crd.yaml | yq eval '.metadata.name' | $sed 's/---//g' | $sed ':a;N;$!ba;s/\n/ /g') ++ kubectl_bin get crd/null -o 'jsonpath={.status.conditions[-1].type}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.P8KOwWO6qC +++ mktemp ++ local LAST_ERR=/tmp/tmp.VUUSqTyzVC ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.P8KOwWO6qC ++ cat /tmp/tmp.VUUSqTyzVC Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 0 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.P8KOwWO6qC ++ cat /tmp/tmp.VUUSqTyzVC Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 4 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.P8KOwWO6qC ++ cat /tmp/tmp.VUUSqTyzVC Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 8 ++ cat /tmp/tmp.P8KOwWO6qC ++ cat /tmp/tmp.VUUSqTyzVC Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ rm /tmp/tmp.P8KOwWO6qC /tmp/tmp.VUUSqTyzVC ++ return 1 + [[ '' == \T\e\r\m\i\n\a\t\i\n\g ]] + '[' -n psmdb-operator ']' + create_namespace psmdb-operator + local namespace=psmdb-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + '[' -n '' ']' + kubectl_bin get ns + grep -E -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' + awk '{print$1}' + xargs kubectl delete ns ++ mktemp + desc 'cleaned up old namespaces psmdb-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace psmdb-operator --ignore-not-found + local LAST_OUT=/tmp/tmp.pyTqIVPsMe ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.wKuzWnDJGn + local exit_status=0 + local timeout=4 ++ seq 0 2 + local LAST_OUT=/tmp/tmp.C8nniAvAvB ++ mktemp + for i in $(seq 0 2) + set +e + kubectl get ns + local LAST_ERR=/tmp/tmp.P7PI71byPb + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete namespace psmdb-operator --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.pyTqIVPsMe + cat /tmp/tmp.wKuzWnDJGn + rm /tmp/tmp.pyTqIVPsMe /tmp/tmp.wKuzWnDJGn + return 0 namespace "pitr-physical-backup-source-4418" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.C8nniAvAvB namespace "psmdb-operator" deleted + cat /tmp/tmp.P7PI71byPb + rm /tmp/tmp.C8nniAvAvB /tmp/tmp.P7PI71byPb + return 0 + kubectl_bin wait --for=delete namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.Y3MYQDvjxr ++ mktemp + local LAST_ERR=/tmp/tmp.uzoNO5lsYK + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Y3MYQDvjxr + cat /tmp/tmp.uzoNO5lsYK + rm /tmp/tmp.Y3MYQDvjxr /tmp/tmp.uzoNO5lsYK + return 0 + desc 'create namespace psmdb-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.oXsby4I0RG ++ mktemp + local LAST_ERR=/tmp/tmp.CHRmC6Sy6k + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.oXsby4I0RG namespace/psmdb-operator created + cat /tmp/tmp.CHRmC6Sy6k + rm /tmp/tmp.oXsby4I0RG /tmp/tmp.CHRmC6Sy6k + return 0 + set_kube_ctx psmdb-operator + local namespace=psmdb-operator ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.PlXF4M79Jl +++ mktemp ++ local LAST_ERR=/tmp/tmp.boXhUg8Rsi ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.PlXF4M79Jl ++ cat /tmp/tmp.boXhUg8Rsi ++ rm /tmp/tmp.PlXF4M79Jl /tmp/tmp.boXhUg8Rsi ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2270-9169a6cb-4-cluster15 --namespace=psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.BRXSEuY21J ++ mktemp + local LAST_ERR=/tmp/tmp.pbdi1x3pAa + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2270-9169a6cb-4-cluster15 --namespace=psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.BRXSEuY21J Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2270-9169a6cb-4-cluster15" modified. + cat /tmp/tmp.pbdi1x3pAa + rm /tmp/tmp.BRXSEuY21J /tmp/tmp.pbdi1x3pAa + return 0 + deploy_operator + desc 'start PSMDB operator: docker.io/perconalab/percona-server-mongodb-operator:PR-2270-9169a6cb' + set +o xtrace ----------------------------------------------------------------------------------- start PSMDB operator: docker.io/perconalab/percona-server-mongodb-operator:PR-2270-9169a6cb ----------------------------------------------------------------------------------- + local cr_file + '[' -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2270/e2e-tests/pitr-physical-backup-source/conf/crd.yaml ']' + cr_file=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2270/deploy/crd.yaml + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2270/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.xIZkYn1LjY ++ mktemp + local LAST_ERR=/tmp/tmp.56k7D9hPIA + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2270/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.xIZkYn1LjY customresourcedefinition.apiextensions.k8s.io/perconaservermongodbbackups.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbrestores.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com serverside-applied + cat /tmp/tmp.56k7D9hPIA + rm /tmp/tmp.xIZkYn1LjY /tmp/tmp.56k7D9hPIA + return 0 + '[' -n psmdb-operator ']' + apply_rbac cw-rbac + local operator_namespace=psmdb-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2270/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: psmdb-operator^' + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.odjswSzRQK ++ mktemp + local LAST_ERR=/tmp/tmp.fUocjMnJZ3 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.odjswSzRQK clusterrole.rbac.authorization.k8s.io/percona-server-mongodb-operator created serviceaccount/percona-server-mongodb-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-server-mongodb-operator created + cat /tmp/tmp.fUocjMnJZ3 + rm /tmp/tmp.odjswSzRQK /tmp/tmp.fUocjMnJZ3 + return 0 + yq eval ' (.spec.template.spec.containers[].image = "docker.io/perconalab/percona-server-mongodb-operator:PR-2270-9169a6cb") | ((.. | select(.[] == "DISABLE_TELEMETRY")) |= .value="true") | ((.. | select(.[] == "LOG_LEVEL")) |= .value="DEBUG")' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2270/deploy/cw-operator.yaml + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.SnZsnMm1Bd ++ mktemp + local LAST_ERR=/tmp/tmp.pbseHsJBHD + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.SnZsnMm1Bd deployment.apps/percona-server-mongodb-operator created + cat /tmp/tmp.pbseHsJBHD + rm /tmp/tmp.SnZsnMm1Bd /tmp/tmp.pbseHsJBHD + return 0 + sleep 20 ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.lhMiLDWizq +++ mktemp ++ local LAST_ERR=/tmp/tmp.T9GD8xznTE ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.lhMiLDWizq ++ cat /tmp/tmp.T9GD8xznTE ++ rm /tmp/tmp.lhMiLDWizq /tmp/tmp.T9GD8xznTE ++ return 0 + wait_operator_pod percona-server-mongodb-operator-dfd77fbfc-cdfb4 + local pod=percona-server-mongodb-operator-dfd77fbfc-cdfb4 + set +o xtrace waiting for pod/percona-server-mongodb-operator-dfd77fbfc-cdfb4 to be ready.OK + echo 'Print operator info from log' Print operator info from log + grep 'Manager starting up' ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.yX0ywiTnr4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.InLOhkUW1T ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.yX0ywiTnr4 ++ cat /tmp/tmp.InLOhkUW1T ++ rm /tmp/tmp.yX0ywiTnr4 /tmp/tmp.InLOhkUW1T ++ return 0 + kubectl_bin logs -n psmdb-operator percona-server-mongodb-operator-dfd77fbfc-cdfb4 ++ mktemp + local LAST_OUT=/tmp/tmp.czi337W5IZ ++ mktemp + local LAST_ERR=/tmp/tmp.t6BjhYq4S5 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl logs -n psmdb-operator percona-server-mongodb-operator-dfd77fbfc-cdfb4 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.czi337W5IZ + cat /tmp/tmp.t6BjhYq4S5 + rm /tmp/tmp.czi337W5IZ /tmp/tmp.t6BjhYq4S5 + return 0 2026-03-26T09:30:44.091Z INFO setup Manager starting up {"gitCommit": "9169a6cb07cda16a30cfe6ec0278ae77b9077338", "gitBranch": "PR-2270-9169a6cb", "buildTime": "", "goVersion": "go1.25.8", "os": "linux", "arch": "amd64"} + create_namespace pitr-physical-backup-source-28911 + local namespace=pitr-physical-backup-source-28911 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ sed s/NAMESPACE// ++ awk '-F ' '{print $2}' + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + grep -E -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' ++ mktemp + '[' -n '' ']' + desc 'cleaned up old namespaces pitr-physical-backup-source-28911' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pitr-physical-backup-source-28911 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pitr-physical-backup-source-28911 --ignore-not-found + xargs kubectl delete ns ++ mktemp + awk '{print$1}' + local LAST_OUT=/tmp/tmp.lvFgwSDaDI + local LAST_OUT=/tmp/tmp.5EW0JWyjwy ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.aRC6GmO6nI + local exit_status=0 + local timeout=4 + local LAST_ERR=/tmp/tmp.E0QsGpctDI + local exit_status=0 + local timeout=4 ++ seq 0 2 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete namespace pitr-physical-backup-source-28911 --ignore-not-found + for i in $(seq 0 2) + set +e + kubectl get ns + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.lvFgwSDaDI + cat /tmp/tmp.E0QsGpctDI + rm /tmp/tmp.lvFgwSDaDI /tmp/tmp.E0QsGpctDI + return 0 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.5EW0JWyjwy + cat /tmp/tmp.aRC6GmO6nI + rm /tmp/tmp.5EW0JWyjwy /tmp/tmp.aRC6GmO6nI + return 0 + kubectl_bin wait --for=delete namespace pitr-physical-backup-source-28911 ++ mktemp + local LAST_OUT=/tmp/tmp.2HL4XDsLJw ++ mktemp + local LAST_ERR=/tmp/tmp.aWMODmUPmF + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete namespace pitr-physical-backup-source-28911 error: resource(s) were provided, but no name was specified + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.2HL4XDsLJw + cat /tmp/tmp.aWMODmUPmF + rm /tmp/tmp.2HL4XDsLJw /tmp/tmp.aWMODmUPmF + return 0 + desc 'create namespace pitr-physical-backup-source-28911' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pitr-physical-backup-source-28911 ----------------------------------------------------------------------------------- + kubectl_bin create namespace pitr-physical-backup-source-28911 ++ mktemp + local LAST_OUT=/tmp/tmp.juLhexAxw0 ++ mktemp + local LAST_ERR=/tmp/tmp.Dsyrv9LsKh + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace pitr-physical-backup-source-28911 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.juLhexAxw0 namespace/pitr-physical-backup-source-28911 created + cat /tmp/tmp.Dsyrv9LsKh + rm /tmp/tmp.juLhexAxw0 /tmp/tmp.Dsyrv9LsKh + return 0 + set_kube_ctx pitr-physical-backup-source-28911 + local namespace=pitr-physical-backup-source-28911 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.X18FUbo1zX +++ mktemp ++ local LAST_ERR=/tmp/tmp.shuQi1fOEq ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.X18FUbo1zX ++ cat /tmp/tmp.shuQi1fOEq ++ rm /tmp/tmp.X18FUbo1zX /tmp/tmp.shuQi1fOEq ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2270-9169a6cb-4-cluster15 --namespace=pitr-physical-backup-source-28911 ++ mktemp + local LAST_OUT=/tmp/tmp.OMBvtpTN45 ++ mktemp + local LAST_ERR=/tmp/tmp.0ARqSopv2b + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2270-9169a6cb-4-cluster15 --namespace=pitr-physical-backup-source-28911 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.OMBvtpTN45 Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2270-9169a6cb-4-cluster15" modified. + cat /tmp/tmp.0ARqSopv2b + rm /tmp/tmp.OMBvtpTN45 /tmp/tmp.0ARqSopv2b + return 0 + deploy_minio + local cert_secret= + local service_name=minio-service + desc 'install MinIO: minio-service' + set +o xtrace ----------------------------------------------------------------------------------- install MinIO: minio-service ----------------------------------------------------------------------------------- + helm uninstall minio-service + : + helm repo remove minio "minio" has been removed from your repositories + helm repo add minio https://charts.min.io/ "minio" has been added to your repositories + local endpoint=http://minio-service:9000 + minio_args=('--version' '5.4.0' '--set' 'replicas=1' '--set' 'mode=standalone' '--set' 'resources.requests.memory=256Mi' '--set' 'rootUser=rootuser' '--set' 'rootPassword=rootpass123' '--set' 'users[0].accessKey=some-access-key' '--set' 'users[0].secretKey=some-secret-key' '--set' 'users[0].policy=consoleAdmin' '--set' 'service.type=ClusterIP' '--set' 'configPathmc=/tmp/' '--set' 'securityContext.enabled=false' '--set' 'persistence.size=2G' '--set' 'fullnameOverride=minio-service' '--set' 'serviceAccount.create=true' '--set' 'serviceAccount.name=minio-service-sa') + local minio_args + [[ -n '' ]] + retry 10 60 helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G --set fullnameOverride=minio-service --set serviceAccount.create=true --set serviceAccount.name=minio-service-sa minio/minio + local max=10 + local delay=60 + shift 2 + local n=1 + helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G --set fullnameOverride=minio-service --set serviceAccount.create=true --set serviceAccount.name=minio-service-sa minio/minio NAME: minio-service LAST DEPLOYED: Thu Mar 26 09:31:23 2026 NAMESPACE: pitr-physical-backup-source-28911 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: MinIO can be accessed via port 9000 on the following DNS name from within your cluster: minio-service.pitr-physical-backup-source-28911.cluster.local To access MinIO from localhost, run the below commands: 1. export POD_NAME=$(kubectl get pods --namespace pitr-physical-backup-source-28911 -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}") 2. kubectl port-forward $POD_NAME 9000 --namespace pitr-physical-backup-source-28911 Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/ You can now access MinIO server on http://localhost:9000. Follow the below steps to connect to MinIO server with mc client: 1. Download the MinIO mc client - https://min.io/docs/minio/linux/reference/minio-mc.html#quickstart 2. export MC_HOST_minio-service-local=http://$(kubectl get secret --namespace pitr-physical-backup-source-28911 minio-service -o jsonpath="{.data.rootUser}" | base64 --decode):$(kubectl get secret --namespace pitr-physical-backup-source-28911 minio-service -o jsonpath="{.data.rootPassword}" | base64 --decode)@localhost:9000 3. mc ls minio-service-local ++ kubectl_bin get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.iZfqymkNRv +++ mktemp ++ local LAST_ERR=/tmp/tmp.lTFMi5FRAJ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.iZfqymkNRv ++ cat /tmp/tmp.lTFMi5FRAJ ++ rm /tmp/tmp.iZfqymkNRv /tmp/tmp.lTFMi5FRAJ ++ return 0 + local MINIO_POD=minio-service-6d5f646cdc-xpsrf + wait_pod minio-service-6d5f646cdc-xpsrf + local pod=minio-service-6d5f646cdc-xpsrf + set +o xtrace waiting for pod/minio-service-6d5f646cdc-xpsrf to be ready.OK + '[' -n psmdb-operator ']' + kubectl_bin create svc -n psmdb-operator externalname minio-service --external-name=minio-service.pitr-physical-backup-source-28911.svc.cluster.local --tcp=9000 service/minio-service created + create_minio_bucket operator-testing http://minio-service:9000 + local bucket=operator-testing + local endpoint=http://minio-service:9000 + kubectl_bin run -i --rm aws-cli --image=docker.io/perconalab/awscli --restart=Never -- bash -c 'AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --no-verify-ssl --endpoint-url http://minio-service:9000 s3 mb s3://operator-testing' ++ mktemp + local LAST_OUT=/tmp/tmp.m8MjXBLZ7v ++ mktemp + local LAST_ERR=/tmp/tmp.vmqOd9cgos + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl run -i --rm aws-cli --image=docker.io/perconalab/awscli --restart=Never -- bash -c 'AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --no-verify-ssl --endpoint-url http://minio-service:9000 s3 mb s3://operator-testing' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.m8MjXBLZ7v make_bucket: operator-testing pod "aws-cli" deleted from pitr-physical-backup-source-28911 namespace + cat /tmp/tmp.vmqOd9cgos + rm /tmp/tmp.m8MjXBLZ7v /tmp/tmp.vmqOd9cgos + return 0 + desc 'create secrets and start client' + set +o xtrace ----------------------------------------------------------------------------------- create secrets and start client ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2270/e2e-tests/conf/secrets.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2270/e2e-tests/conf/client.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2270/e2e-tests/conf/minio-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.p74cObKP6i ++ mktemp + local LAST_ERR=/tmp/tmp.NQKhnc2Vpk + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2270/e2e-tests/conf/secrets.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2270/e2e-tests/conf/client.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2270/e2e-tests/conf/minio-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.p74cObKP6i secret/some-users created deployment.apps/psmdb-client created secret/minio-secret created + cat /tmp/tmp.NQKhnc2Vpk + rm /tmp/tmp.p74cObKP6i /tmp/tmp.NQKhnc2Vpk + return 0 + cluster=some-name + desc 'create first PSMDB cluster some-name' + set +o xtrace ----------------------------------------------------------------------------------- create first PSMDB cluster some-name ----------------------------------------------------------------------------------- + apply_cluster /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2270/e2e-tests/pitr-physical-backup-source/conf/some-name-rs0.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2270/e2e-tests/pitr-physical-backup-source/conf/some-name-rs0.yml + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2270/e2e-tests/pitr-physical-backup-source/conf/some-name-rs0.yml ++ mktemp + yq eval '(.spec | select(has("initImage"))).initImage = "docker.io/perconalab/percona-server-mongodb-operator:PR-2270-9169a6cb"' + /usr/sbin/sed -e s/NAME_SPACE/pitr-physical-backup-source-28911/g + yq eval '(.spec | select(has("backup"))).backup.image = "docker.io/perconalab/percona-server-mongodb-operator:main-backup"' + yq eval '(.spec | select(.image == null)).image = "docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0"' + local LAST_OUT=/tmp/tmp.pQ4VZiLxKx + yq eval '.spec.upgradeOptions.apply="Never"' ++ mktemp + yq eval '(.spec | select(has("pmm"))).pmm.image = "docker.io/percona/pmm-client:2.44.1-1"' + local LAST_ERR=/tmp/tmp.0HrMmTVgln + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.pQ4VZiLxKx perconaservermongodb.psmdb.percona.com/some-name created + cat /tmp/tmp.0HrMmTVgln + rm /tmp/tmp.pQ4VZiLxKx /tmp/tmp.0HrMmTVgln + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- + wait_for_running some-name-rs0 3 + local name=some-name-rs0 + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=rs0 + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-rs0-0 + local pod=some-name-rs0-0 + set +o xtrace waiting for pod/some-name-rs0-0 to be ready.......OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-rs0-1 + local pod=some-name-rs0-1 + set +o xtrace waiting for pod/some-name-rs0-1 to be ready.....OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VwUUytfUAI +++ mktemp ++ local LAST_ERR=/tmp/tmp.aNcUv8q7eW ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.VwUUytfUAI ++ cat /tmp/tmp.aNcUv8q7eW ++ rm /tmp/tmp.VwUUytfUAI /tmp/tmp.aNcUv8q7eW ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-rs0-2 + local pod=some-name-rs0-2 + set +o xtrace waiting for pod/some-name-rs0-2 to be ready.......OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qplTZZQAOW +++ mktemp ++ local LAST_ERR=/tmp/tmp.kmG7iN3eSW ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.qplTZZQAOW ++ cat /tmp/tmp.kmG7iN3eSW ++ rm /tmp/tmp.qplTZZQAOW /tmp/tmp.kmG7iN3eSW ++ return 0 + [[ '' == \t\r\u\e ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zR6LbuppvV +++ mktemp ++ local LAST_ERR=/tmp/tmp.J5folXQh3Y ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.zR6LbuppvV ++ cat /tmp/tmp.J5folXQh3Y ++ rm /tmp/tmp.zR6LbuppvV /tmp/tmp.J5folXQh3Y ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness................................................. + wait_for_running some-name-cfg 3 false + local name=some-name-cfg + let last_pod=2 + local check_cluster_readyness=false + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=cfg + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-cfg-0 + local pod=some-name-cfg-0 + set +o xtrace waiting for pod/some-name-cfg-0 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-cfg-1 + local pod=some-name-cfg-1 + set +o xtrace waiting for pod/some-name-cfg-1 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EcedRhfV9e +++ mktemp ++ local LAST_ERR=/tmp/tmp.6UF7VVmSUW ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.EcedRhfV9e ++ cat /tmp/tmp.6UF7VVmSUW ++ rm /tmp/tmp.EcedRhfV9e /tmp/tmp.6UF7VVmSUW ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-cfg-2 + local pod=some-name-cfg-2 + set +o xtrace waiting for pod/some-name-cfg-2 to be ready.OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4en1bvXAY1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.cigWrKRPRH ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4en1bvXAY1 ++ cat /tmp/tmp.cigWrKRPRH ++ rm /tmp/tmp.4en1bvXAY1 /tmp/tmp.cigWrKRPRH ++ return 0 + [[ '' == \t\r\u\e ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PQeCoRQalc +++ mktemp ++ local LAST_ERR=/tmp/tmp.gUSOjaUCBN ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.PQeCoRQalc ++ cat /tmp/tmp.gUSOjaUCBN ++ rm /tmp/tmp.PQeCoRQalc /tmp/tmp.gUSOjaUCBN ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ false == \t\r\u\e ]] + sleep 10 + write_initial_data + desc 'create user myApp' + set +o xtrace ----------------------------------------------------------------------------------- create user myApp ----------------------------------------------------------------------------------- + run_mongos 'db.createUser({user:"myApp",pwd:"myPass",roles:[{db:"myApp",role:"readWrite"}]})' userAdmin:userAdmin123456@some-name-mongos.pitr-physical-backup-source-28911 + local 'command=db.createUser({user:"myApp",pwd:"myPass",roles:[{db:"myApp",role:"readWrite"}]})' + local uri=userAdmin:userAdmin123456@some-name-mongos.pitr-physical-backup-source-28911 + local driver=mongodb + local suffix=.svc.cluster.local + local mongo_flag= + local port=27017 + local mongo_bin=mongo ++ echo .svc.cluster.local ++ awk -F: '{print $2}' + suffix_port= + [[ -z '' ]] + suffix=.svc.cluster.local:27017 ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bSI8DsRK3b +++ mktemp ++ local LAST_ERR=/tmp/tmp.5PvDieMRcG ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.bSI8DsRK3b ++ cat /tmp/tmp.5PvDieMRcG ++ rm /tmp/tmp.bSI8DsRK3b /tmp/tmp.5PvDieMRcG ++ return 0 + local client_container=psmdb-client-bb8b97679-cp4ft + kubectl_bin exec psmdb-client-bb8b97679-cp4ft -- bash -c 'printf '\''db.createUser({user:"myApp",pwd:"myPass",roles:[{db:"myApp",role:"readWrite"}]})\n'\'' | mongo mongodb://userAdmin:userAdmin123456@some-name-mongos.pitr-physical-backup-source-28911.svc.cluster.local:27017/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.d5lIxUPLf3 ++ mktemp + local LAST_ERR=/tmp/tmp.qAOqm4ukkv + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-cp4ft -- bash -c 'printf '\''db.createUser({user:"myApp",pwd:"myPass",roles:[{db:"myApp",role:"readWrite"}]})\n'\'' | mongo mongodb://userAdmin:userAdmin123456@some-name-mongos.pitr-physical-backup-source-28911.svc.cluster.local:27017/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.d5lIxUPLf3 Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.pitr-physical-backup-source-28911.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("84f2326a-b193-40ca-898c-4291f422c088") } Percona Server for MongoDB server version: v8.0.19-7 WARNING: shell and server versions do not match Successfully added user: { "user" : "myApp", "roles" : [ { "db" : "myApp", "role" : "readWrite" } ] } bye + cat /tmp/tmp.qAOqm4ukkv + rm /tmp/tmp.d5lIxUPLf3 /tmp/tmp.qAOqm4ukkv + return 0 + sleep 2 + write_document + local cmp_postfix= + local sleep_value=0 + log 'write initial data, read from all' + set +o xtrace [2026-03-26T09:35:40+0000] write initial data, read from all + run_mongos 'use myApp\n db.test.insert({ x: 100500 })' myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911 + local 'command=use myApp\n db.test.insert({ x: 100500 })' + local uri=myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911 + local driver=mongodb + local suffix=.svc.cluster.local + local mongo_flag= + local port=27017 + local mongo_bin=mongo ++ echo .svc.cluster.local ++ awk -F: '{print $2}' + suffix_port= + [[ -z '' ]] + suffix=.svc.cluster.local:27017 ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vBB5M8gbWz +++ mktemp ++ local LAST_ERR=/tmp/tmp.YA4UJYNpSX ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.vBB5M8gbWz ++ cat /tmp/tmp.YA4UJYNpSX ++ rm /tmp/tmp.vBB5M8gbWz /tmp/tmp.YA4UJYNpSX ++ return 0 + local client_container=psmdb-client-bb8b97679-cp4ft + kubectl_bin exec psmdb-client-bb8b97679-cp4ft -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911.svc.cluster.local:27017/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.BH7IeGtBzO ++ mktemp + local LAST_ERR=/tmp/tmp.MfgubFLSpn + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-cp4ft -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911.svc.cluster.local:27017/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.BH7IeGtBzO Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.pitr-physical-backup-source-28911.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("c19bc87f-e30f-44d5-99b0-c142bd523ace") } Percona Server for MongoDB server version: v8.0.19-7 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.MfgubFLSpn + rm /tmp/tmp.BH7IeGtBzO /tmp/tmp.MfgubFLSpn + return 0 + sleep 0 + compare_mongos_cmd find myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911 + local command=find + local uri=myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911 + local postfix= + local suffix= + local database=myApp + local collection=test + local port=27017 + local tls=false + [[ false == \t\r\u\e ]] + mongos_command=run_mongos + log 'running db.test.command() in myApp' + set +o xtrace [2026-03-26T09:35:43+0000] running db.test.command() in myApp + run_mongos 'use myApp\n db.test.find()' myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911 mongodb '' '' 27017 + local 'command=use myApp\n db.test.find()' + local uri=myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911 + local driver=mongodb + local suffix=.svc.cluster.local + local mongo_flag= + local port=27017 + local mongo_bin=mongo + grep -E -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:|Started a new thread for the timer service' ++ echo .svc.cluster.local ++ awk -F: '{print $2}' + /usr/sbin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' + suffix_port= + [[ -z '' ]] + suffix=.svc.cluster.local:27017 ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5MyXwPar8U +++ mktemp ++ local LAST_ERR=/tmp/tmp.aq9Jh60fmz ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5MyXwPar8U ++ cat /tmp/tmp.aq9Jh60fmz ++ rm /tmp/tmp.5MyXwPar8U /tmp/tmp.aq9Jh60fmz ++ return 0 + local client_container=psmdb-client-bb8b97679-cp4ft + kubectl_bin exec psmdb-client-bb8b97679-cp4ft -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911.svc.cluster.local:27017/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.YS5R92N0in ++ mktemp + local LAST_ERR=/tmp/tmp.3ANHwaWXGc + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-cp4ft -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911.svc.cluster.local:27017/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.YS5R92N0in + cat /tmp/tmp.3ANHwaWXGc + rm /tmp/tmp.YS5R92N0in /tmp/tmp.3ANHwaWXGc + return 0 + [[ 0 -eq 0 ]] + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2270/e2e-tests/pitr-physical-backup-source/compare/find.json /tmp/tmp.CB9QMdE4Yh/find + wait_backup_agent some-name-rs0-0 + local agent_pod=some-name-rs0-0 + set +o xtrace waiting for pbm-agent to be ready in some-name-rs0-0...2026-03-26T09:34:48.000+0000 I listening for the commands + wait_backup_agent some-name-rs0-1 + local agent_pod=some-name-rs0-1 + set +o xtrace waiting for pbm-agent to be ready in some-name-rs0-1...2026-03-26T09:34:51.000+0000 I listening for the commands + wait_backup_agent some-name-rs0-2 + local agent_pod=some-name-rs0-2 + set +o xtrace waiting for pbm-agent to be ready in some-name-rs0-2...2026-03-26T09:34:57.000+0000 I listening for the commands + wait_backup_agent some-name-rs1-0 + local agent_pod=some-name-rs1-0 + set +o xtrace waiting for pbm-agent to be ready in some-name-rs1-0...2026-03-26T09:34:51.000+0000 I listening for the commands + wait_backup_agent some-name-rs1-1 + local agent_pod=some-name-rs1-1 + set +o xtrace waiting for pbm-agent to be ready in some-name-rs1-1...2026-03-26T09:34:52.000+0000 I listening for the commands + wait_backup_agent some-name-rs1-2 + local agent_pod=some-name-rs1-2 + set +o xtrace waiting for pbm-agent to be ready in some-name-rs1-2...2026-03-26T09:34:59.000+0000 I listening for the commands + wait_backup_agent some-name-rs2-0 + local agent_pod=some-name-rs2-0 + set +o xtrace waiting for pbm-agent to be ready in some-name-rs2-0...2026-03-26T09:34:51.000+0000 I listening for the commands + wait_backup_agent some-name-rs2-1 + local agent_pod=some-name-rs2-1 + set +o xtrace waiting for pbm-agent to be ready in some-name-rs2-1...2026-03-26T09:34:53.000+0000 I listening for the commands + wait_backup_agent some-name-rs2-2 + local agent_pod=some-name-rs2-2 + set +o xtrace waiting for pbm-agent to be ready in some-name-rs2-2...2026-03-26T09:35:02.000+0000 I listening for the commands + echo 'Sleeping for 360 seconds' Sleeping for 360 seconds + sleep 360 + backup_name_minio=backup-minio + desc 'restore pitr type date using backupSource' + set +o xtrace ----------------------------------------------------------------------------------- restore pitr type date using backupSource ----------------------------------------------------------------------------------- + run_backup backup-minio 1 physical + local name=backup-minio + local idx=1 + local type=physical + desc 'run backup backup-minio-1' + set +o xtrace ----------------------------------------------------------------------------------- run backup backup-minio-1 ----------------------------------------------------------------------------------- + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2270/e2e-tests/pitr-physical-backup-source/conf/backup-minio.yml + /usr/sbin/sed -e 's/name:/name: backup-minio-1/' + /usr/sbin/sed -e 's/type:/type: physical/' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.DVjmodsH3q ++ mktemp + local LAST_ERR=/tmp/tmp.W9SZ3Zcvpe + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.DVjmodsH3q perconaservermongodbbackup.psmdb.percona.com/backup-minio-1 created + cat /tmp/tmp.W9SZ3Zcvpe + rm /tmp/tmp.DVjmodsH3q /tmp/tmp.W9SZ3Zcvpe + return 0 + wait_backup backup-minio-1 + local backup_name=backup-minio-1 + local target_state=ready + set +o xtrace waiting for backup-minio-1 to reach ready state..........OK + compare_latest_restorable_time some-name-rs0 backup-minio-1 + local cluster=some-name-rs0 + local backup_name=backup-minio-1 + local latest_restorable_time + local backup_time ++ get_latest_restorable_time some-name-rs0 ++ local cluster=some-name-rs0 ++ local first_timestamp ++ local second_timestamp ++ local retry=0 ++ [[ '' != '' ]] +++ kubectl_bin exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ jq '.backups.pitrChunks.pitrChunks | last | .range.end' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.0PY2zBqKhR ++++ mktemp +++ local LAST_ERR=/tmp/tmp.IrNWP2Vvxu +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.0PY2zBqKhR +++ cat /tmp/tmp.IrNWP2Vvxu +++ rm /tmp/tmp.0PY2zBqKhR /tmp/tmp.IrNWP2Vvxu +++ return 0 ++ first_timestamp=null ++ sleep 5 ++ [[ null != '' ]] ++ [[ null != \n\u\l\l ]] ++ let retry+=1 ++ [[ 1 -gt 30 ]] ++ [[ null != '' ]] ++ [[ null != \n\u\l\l ]] +++ kubectl_bin exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ jq '.backups.pitrChunks.pitrChunks | last | .range.end' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.i29MaQiUhh ++++ mktemp +++ local LAST_ERR=/tmp/tmp.QbX9nsyzjE +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.i29MaQiUhh +++ cat /tmp/tmp.QbX9nsyzjE +++ rm /tmp/tmp.i29MaQiUhh /tmp/tmp.QbX9nsyzjE +++ return 0 ++ first_timestamp=null ++ sleep 5 ++ [[ null != '' ]] ++ [[ null != \n\u\l\l ]] ++ let retry+=1 ++ [[ 2 -gt 30 ]] ++ [[ null != '' ]] ++ [[ null != \n\u\l\l ]] +++ kubectl_bin exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ jq '.backups.pitrChunks.pitrChunks | last | .range.end' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.sIOGUUWq97 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.c22ioj4HQO +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.sIOGUUWq97 +++ cat /tmp/tmp.c22ioj4HQO +++ rm /tmp/tmp.sIOGUUWq97 /tmp/tmp.c22ioj4HQO +++ return 0 ++ first_timestamp=null ++ sleep 5 ++ [[ null != '' ]] ++ [[ null != \n\u\l\l ]] ++ let retry+=1 ++ [[ 3 -gt 30 ]] ++ [[ null != '' ]] ++ [[ null != \n\u\l\l ]] +++ kubectl_bin exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ jq '.backups.pitrChunks.pitrChunks | last | .range.end' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Ipmke8eqNH ++++ mktemp +++ local LAST_ERR=/tmp/tmp.zcVc9EwHjd +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.Ipmke8eqNH +++ cat /tmp/tmp.zcVc9EwHjd +++ rm /tmp/tmp.Ipmke8eqNH /tmp/tmp.zcVc9EwHjd +++ return 0 ++ first_timestamp=null ++ sleep 5 ++ [[ null != '' ]] ++ [[ null != \n\u\l\l ]] ++ let retry+=1 ++ [[ 4 -gt 30 ]] ++ [[ null != '' ]] ++ [[ null != \n\u\l\l ]] +++ kubectl_bin exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ jq '.backups.pitrChunks.pitrChunks | last | .range.end' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Zi64b3xHGQ ++++ mktemp +++ local LAST_ERR=/tmp/tmp.owe68OcPNK +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.Zi64b3xHGQ +++ cat /tmp/tmp.owe68OcPNK +++ rm /tmp/tmp.Zi64b3xHGQ /tmp/tmp.owe68OcPNK +++ return 0 ++ first_timestamp=null ++ sleep 5 ++ [[ null != '' ]] ++ [[ null != \n\u\l\l ]] ++ let retry+=1 ++ [[ 5 -gt 30 ]] ++ [[ null != '' ]] ++ [[ null != \n\u\l\l ]] +++ kubectl_bin exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ jq '.backups.pitrChunks.pitrChunks | last | .range.end' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.wpDAexFDTK ++++ mktemp +++ local LAST_ERR=/tmp/tmp.UuUno8NPVg +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.wpDAexFDTK +++ cat /tmp/tmp.UuUno8NPVg +++ rm /tmp/tmp.wpDAexFDTK /tmp/tmp.UuUno8NPVg +++ return 0 ++ first_timestamp=null ++ sleep 5 ++ [[ null != '' ]] ++ [[ null != \n\u\l\l ]] ++ let retry+=1 ++ [[ 6 -gt 30 ]] ++ [[ null != '' ]] ++ [[ null != \n\u\l\l ]] +++ kubectl_bin exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ jq '.backups.pitrChunks.pitrChunks | last | .range.end' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.lt4lzwSAkO ++++ mktemp +++ local LAST_ERR=/tmp/tmp.F0ROwTnLyC +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.lt4lzwSAkO +++ cat /tmp/tmp.F0ROwTnLyC +++ rm /tmp/tmp.lt4lzwSAkO /tmp/tmp.F0ROwTnLyC +++ return 0 ++ first_timestamp=1774518169 ++ sleep 5 ++ [[ 1774518169 != '' ]] ++ [[ 1774518169 != \n\u\l\l ]] +++ kubectl_bin exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ jq '.backups.pitrChunks.pitrChunks | last | .range.end' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.UkRsVbXj7y ++++ mktemp +++ local LAST_ERR=/tmp/tmp.tqIN1cpDd2 +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.UkRsVbXj7y +++ cat /tmp/tmp.tqIN1cpDd2 +++ rm /tmp/tmp.UkRsVbXj7y /tmp/tmp.tqIN1cpDd2 +++ return 0 ++ second_timestamp=1774518169 ++ let retry+=1 ++ [[ 7 -gt 30 ]] ++ [[ 1774518169 != '' ]] ++ [[ 1774518169 != \n\u\l\l ]] ++ [[ 1774518169 == 1774518169 ]] ++ /usr/sbin/date -u -d @1774518169 +%Y-%m-%dT%H:%M:%SZ + latest_restorable_time=2026-03-26T09:42:49Z ++ get_latest_restorable_time_from_backup_object backup-minio-1 ++ local backup_name=backup-minio-1 ++ local latestRestorableTime ++ local retry=0 ++ [[ '' != '' ]] ++ sleep 5 +++ kubectl_bin get psmdb-backup backup-minio-1 -o 'jsonpath={.status.latestRestorableTime}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.vDVUdAg3ET ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Rz6QQPb0sm +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get psmdb-backup backup-minio-1 -o 'jsonpath={.status.latestRestorableTime}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.vDVUdAg3ET +++ cat /tmp/tmp.Rz6QQPb0sm +++ rm /tmp/tmp.vDVUdAg3ET /tmp/tmp.Rz6QQPb0sm +++ return 0 ++ latestRestorableTime=2026-03-26T09:42:49Z ++ let retry+=1 ++ [[ 1 -gt 30 ]] ++ [[ 2026-03-26T09:42:49Z != '' ]] ++ [[ 2026-03-26T09:42:49Z != \n\u\l\l ]] ++ echo 2026-03-26T09:42:49Z + backup_time=2026-03-26T09:42:49Z + [[ 2026-03-26T09:42:49Z != \2\0\2\6\-\0\3\-\2\6\T\0\9\:\4\2\:\4\9\Z ]] + reset_collection + desc 'reset data' + set +o xtrace ----------------------------------------------------------------------------------- reset data ----------------------------------------------------------------------------------- + run_mongos 'use myApp\n db.test.remove({})' myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911 + local 'command=use myApp\n db.test.remove({})' + local uri=myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911 + local driver=mongodb + local suffix=.svc.cluster.local + local mongo_flag= + local port=27017 + local mongo_bin=mongo ++ echo .svc.cluster.local ++ awk -F: '{print $2}' + suffix_port= + [[ -z '' ]] + suffix=.svc.cluster.local:27017 ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DilTvPsiFp +++ mktemp ++ local LAST_ERR=/tmp/tmp.NRdHpIBcX1 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.DilTvPsiFp ++ cat /tmp/tmp.NRdHpIBcX1 ++ rm /tmp/tmp.DilTvPsiFp /tmp/tmp.NRdHpIBcX1 ++ return 0 + local client_container=psmdb-client-bb8b97679-cp4ft + kubectl_bin exec psmdb-client-bb8b97679-cp4ft -- bash -c 'printf '\''use myApp\n db.test.remove({})\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911.svc.cluster.local:27017/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.q8snjq1dbH ++ mktemp + local LAST_ERR=/tmp/tmp.NXYVMSImlm + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-cp4ft -- bash -c 'printf '\''use myApp\n db.test.remove({})\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911.svc.cluster.local:27017/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.q8snjq1dbH Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.pitr-physical-backup-source-28911.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("a89afd2c-73bf-4edc-8207-83588530962c") } Percona Server for MongoDB server version: v8.0.19-7 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nRemoved" : 1 }) bye + cat /tmp/tmp.NXYVMSImlm + rm /tmp/tmp.q8snjq1dbH /tmp/tmp.NXYVMSImlm + return 0 + sleep 2 + write_document '' 120 + local cmp_postfix= + local sleep_value=120 + log 'write initial data, read from all' + set +o xtrace [2026-03-26T09:43:19+0000] write initial data, read from all + run_mongos 'use myApp\n db.test.insert({ x: 100500 })' myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911 + local 'command=use myApp\n db.test.insert({ x: 100500 })' + local uri=myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911 + local driver=mongodb + local suffix=.svc.cluster.local + local mongo_flag= + local port=27017 + local mongo_bin=mongo ++ echo .svc.cluster.local ++ awk -F: '{print $2}' + suffix_port= + [[ -z '' ]] + suffix=.svc.cluster.local:27017 ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mWNQw2KKYn +++ mktemp ++ local LAST_ERR=/tmp/tmp.FPrMuDLvTu ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.mWNQw2KKYn ++ cat /tmp/tmp.FPrMuDLvTu ++ rm /tmp/tmp.mWNQw2KKYn /tmp/tmp.FPrMuDLvTu ++ return 0 + local client_container=psmdb-client-bb8b97679-cp4ft + kubectl_bin exec psmdb-client-bb8b97679-cp4ft -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911.svc.cluster.local:27017/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.R4CHTyGRzp ++ mktemp + local LAST_ERR=/tmp/tmp.qiCJLisSQO + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-cp4ft -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911.svc.cluster.local:27017/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.R4CHTyGRzp Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.pitr-physical-backup-source-28911.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("202c7835-cbda-493d-b1cf-e2ef3d704a09") } Percona Server for MongoDB server version: v8.0.19-7 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.qiCJLisSQO + rm /tmp/tmp.R4CHTyGRzp /tmp/tmp.qiCJLisSQO + return 0 + sleep 120 + compare_mongos_cmd find myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911 + local command=find + local uri=myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911 + local postfix= + local suffix= + local database=myApp + local collection=test + local port=27017 + local tls=false + [[ false == \t\r\u\e ]] + mongos_command=run_mongos + log 'running db.test.command() in myApp' + set +o xtrace [2026-03-26T09:45:22+0000] running db.test.command() in myApp + run_mongos 'use myApp\n db.test.find()' myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911 mongodb '' '' 27017 + local 'command=use myApp\n db.test.find()' + local uri=myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911 + local driver=mongodb + local suffix=.svc.cluster.local + local mongo_flag= + local port=27017 + local mongo_bin=mongo + grep -E -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:|Started a new thread for the timer service' + /usr/sbin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ echo .svc.cluster.local ++ awk -F: '{print $2}' + suffix_port= + [[ -z '' ]] + suffix=.svc.cluster.local:27017 ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.riot4kSumv +++ mktemp ++ local LAST_ERR=/tmp/tmp.n1GX9MVDQ2 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.riot4kSumv ++ cat /tmp/tmp.n1GX9MVDQ2 ++ rm /tmp/tmp.riot4kSumv /tmp/tmp.n1GX9MVDQ2 ++ return 0 + local client_container=psmdb-client-bb8b97679-cp4ft + kubectl_bin exec psmdb-client-bb8b97679-cp4ft -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911.svc.cluster.local:27017/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.t6oJCe3xKf ++ mktemp + local LAST_ERR=/tmp/tmp.PzxaGQT08G + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-cp4ft -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911.svc.cluster.local:27017/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.t6oJCe3xKf + cat /tmp/tmp.PzxaGQT08G + rm /tmp/tmp.t6oJCe3xKf /tmp/tmp.PzxaGQT08G + return 0 + [[ 0 -eq 0 ]] + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2270/e2e-tests/pitr-physical-backup-source/compare/find.json /tmp/tmp.CB9QMdE4Yh/find ++ run_mongos 'new Date().getTime() / 1000' myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911 mongodb '' --quiet ++ local 'command=new Date().getTime() / 1000' ++ local uri=myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911 ++ local driver=mongodb ++ local suffix=.svc.cluster.local ++ local mongo_flag=--quiet ++ grep -E -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match' ++ local port=27017 ++ local mongo_bin=mongo ++ cut -d. -f1 +++ echo .svc.cluster.local +++ awk -F: '{print $2}' ++ suffix_port= ++ [[ -z '' ]] ++ suffix=.svc.cluster.local:27017 +++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.WQVRzmfByd ++++ mktemp +++ local LAST_ERR=/tmp/tmp.S6bV7btwSB +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.WQVRzmfByd +++ cat /tmp/tmp.S6bV7btwSB +++ rm /tmp/tmp.WQVRzmfByd /tmp/tmp.S6bV7btwSB +++ return 0 ++ local client_container=psmdb-client-bb8b97679-cp4ft ++ kubectl_bin exec psmdb-client-bb8b97679-cp4ft -- bash -c 'printf '\''new Date().getTime() / 1000\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911.svc.cluster.local:27017/admin --quiet' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KY5lPDinl5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.N0dTx6FdGO ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl exec psmdb-client-bb8b97679-cp4ft -- bash -c 'printf '\''new Date().getTime() / 1000\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911.svc.cluster.local:27017/admin --quiet' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.KY5lPDinl5 ++ cat /tmp/tmp.N0dTx6FdGO ++ rm /tmp/tmp.KY5lPDinl5 /tmp/tmp.N0dTx6FdGO ++ return 0 + time_now=1774518326 + check_recovery backup-minio-1 date 1774518326 '' some-name backupSource + local backup_name=backup-minio-1 + local restore_type=date + local restore_date=1774518326 + local cmp_postfix= + local cluster_name=some-name + local backupSource=backupSource ++ get_latest_oplog_chunk_ts some-name ++ local cluster=some-name +++ kubectl_bin exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ jq '.backups.pitrChunks.pitrChunks | last | .range.end' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.BpYjOywepC ++++ mktemp +++ local LAST_ERR=/tmp/tmp.DyAikS9Fmk +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.BpYjOywepC +++ cat /tmp/tmp.DyAikS9Fmk +++ rm /tmp/tmp.BpYjOywepC /tmp/tmp.DyAikS9Fmk +++ return 0 ++ echo 1774518297 + local latest_ts=1774518297 + desc 'write more data before restore by date' + set +o xtrace ----------------------------------------------------------------------------------- write more data before restore by date ----------------------------------------------------------------------------------- + run_mongos 'use myApp\n db.test.insert({ x: 100501 })' myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911 + local 'command=use myApp\n db.test.insert({ x: 100501 })' + local uri=myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911 + local driver=mongodb + local suffix=.svc.cluster.local + local mongo_flag= + local port=27017 + local mongo_bin=mongo ++ echo .svc.cluster.local ++ awk -F: '{print $2}' + suffix_port= + [[ -z '' ]] + suffix=.svc.cluster.local:27017 ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vV160QZGQK +++ mktemp ++ local LAST_ERR=/tmp/tmp.BtjJVNiB0h ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.vV160QZGQK ++ cat /tmp/tmp.BtjJVNiB0h ++ rm /tmp/tmp.vV160QZGQK /tmp/tmp.BtjJVNiB0h ++ return 0 + local client_container=psmdb-client-bb8b97679-cp4ft + kubectl_bin exec psmdb-client-bb8b97679-cp4ft -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911.svc.cluster.local:27017/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.aEnPwpuRJb ++ mktemp + local LAST_ERR=/tmp/tmp.HTpPdOl3jb + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-cp4ft -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.pitr-physical-backup-source-28911.svc.cluster.local:27017/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.aEnPwpuRJb Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.pitr-physical-backup-source-28911.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("5966c0a3-47f2-4ba8-aa21-a52f5a6cec0e") } Percona Server for MongoDB server version: v8.0.19-7 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.HTpPdOl3jb + rm /tmp/tmp.aEnPwpuRJb /tmp/tmp.HTpPdOl3jb + return 0 + [[ -n 1774518326 ]] ++ format_date 1774518326 ++ local timestamp=1774518326 +++ TZ=UTC +++ /usr/sbin/date -d@1774518326 '+%Y-%m-%d %H:%M:%S' ++ echo 2026-03-26 09:45:26 + desc 'Restoring to time 2026-03-26 09:45:26' + set +o xtrace ----------------------------------------------------------------------------------- Restoring to time 2026-03-26 09:45:26 ----------------------------------------------------------------------------------- + retries=0 + [[ 1774518297 -gt 1774518326 ]] + [[ 0 -gt 30 ]] ++ get_latest_oplog_chunk_ts some-name ++ local cluster=some-name +++ kubectl_bin exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ jq '.backups.pitrChunks.pitrChunks | last | .range.end' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.j0qEpT2ceS ++++ mktemp +++ local LAST_ERR=/tmp/tmp.uKTIdzQtwz +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.j0qEpT2ceS +++ cat /tmp/tmp.uKTIdzQtwz +++ rm /tmp/tmp.j0qEpT2ceS /tmp/tmp.uKTIdzQtwz +++ return 0 ++ echo 1774518297 + latest_ts=1774518297 + retries=1 ++ format_date 1774518297 ++ local timestamp=1774518297 +++ TZ=UTC +++ /usr/sbin/date -d@1774518297 '+%Y-%m-%d %H:%M:%S' ++ echo 2026-03-26 09:44:57 ++ format_date 1774518326 ++ local timestamp=1774518326 +++ TZ=UTC +++ /usr/sbin/date -d@1774518326 '+%Y-%m-%d %H:%M:%S' ++ echo 2026-03-26 09:45:26 + echo 'Waiting for last oplog chunk (2026-03-26 09:44:57) to be greater than restore target (2026-03-26 09:45:26)' Waiting for last oplog chunk (2026-03-26 09:44:57) to be greater than restore target (2026-03-26 09:45:26) + sleep 10 + [[ 1774518297 -gt 1774518326 ]] + [[ 1 -gt 30 ]] ++ get_latest_oplog_chunk_ts some-name ++ local cluster=some-name +++ kubectl_bin exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ jq '.backups.pitrChunks.pitrChunks | last | .range.end' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.svJaGiacNg ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Erhv0K5SWs +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.svJaGiacNg +++ cat /tmp/tmp.Erhv0K5SWs +++ rm /tmp/tmp.svJaGiacNg /tmp/tmp.Erhv0K5SWs +++ return 0 ++ echo 1774518297 + latest_ts=1774518297 + retries=2 ++ format_date 1774518297 ++ local timestamp=1774518297 +++ TZ=UTC +++ /usr/sbin/date -d@1774518297 '+%Y-%m-%d %H:%M:%S' ++ echo 2026-03-26 09:44:57 ++ format_date 1774518326 ++ local timestamp=1774518326 +++ TZ=UTC +++ /usr/sbin/date -d@1774518326 '+%Y-%m-%d %H:%M:%S' ++ echo 2026-03-26 09:45:26 + echo 'Waiting for last oplog chunk (2026-03-26 09:44:57) to be greater than restore target (2026-03-26 09:45:26)' Waiting for last oplog chunk (2026-03-26 09:44:57) to be greater than restore target (2026-03-26 09:45:26) + sleep 10 + [[ 1774518297 -gt 1774518326 ]] + [[ 2 -gt 30 ]] ++ get_latest_oplog_chunk_ts some-name ++ local cluster=some-name +++ kubectl_bin exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ jq '.backups.pitrChunks.pitrChunks | last | .range.end' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.W9O89DpxRx ++++ mktemp +++ local LAST_ERR=/tmp/tmp.zwsDfqGf44 +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.W9O89DpxRx +++ cat /tmp/tmp.zwsDfqGf44 +++ rm /tmp/tmp.W9O89DpxRx /tmp/tmp.zwsDfqGf44 +++ return 0 ++ echo 1774518297 + latest_ts=1774518297 + retries=3 ++ format_date 1774518297 ++ local timestamp=1774518297 +++ TZ=UTC +++ /usr/sbin/date -d@1774518297 '+%Y-%m-%d %H:%M:%S' ++ echo 2026-03-26 09:44:57 ++ format_date 1774518326 ++ local timestamp=1774518326 +++ TZ=UTC +++ /usr/sbin/date -d@1774518326 '+%Y-%m-%d %H:%M:%S' ++ echo 2026-03-26 09:45:26 + echo 'Waiting for last oplog chunk (2026-03-26 09:44:57) to be greater than restore target (2026-03-26 09:45:26)' Waiting for last oplog chunk (2026-03-26 09:44:57) to be greater than restore target (2026-03-26 09:45:26) + sleep 10 + [[ 1774518297 -gt 1774518326 ]] + [[ 3 -gt 30 ]] ++ get_latest_oplog_chunk_ts some-name ++ local cluster=some-name +++ kubectl_bin exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ jq '.backups.pitrChunks.pitrChunks | last | .range.end' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.lDJpOpTtyJ ++++ mktemp +++ local LAST_ERR=/tmp/tmp.eZJc0IfZeW +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.lDJpOpTtyJ +++ cat /tmp/tmp.eZJc0IfZeW +++ rm /tmp/tmp.lDJpOpTtyJ /tmp/tmp.eZJc0IfZeW +++ return 0 ++ echo 1774518297 + latest_ts=1774518297 + retries=4 ++ format_date 1774518297 ++ local timestamp=1774518297 +++ TZ=UTC +++ /usr/sbin/date -d@1774518297 '+%Y-%m-%d %H:%M:%S' ++ echo 2026-03-26 09:44:57 ++ format_date 1774518326 ++ local timestamp=1774518326 +++ TZ=UTC +++ /usr/sbin/date -d@1774518326 '+%Y-%m-%d %H:%M:%S' ++ echo 2026-03-26 09:45:26 + echo 'Waiting for last oplog chunk (2026-03-26 09:44:57) to be greater than restore target (2026-03-26 09:45:26)' Waiting for last oplog chunk (2026-03-26 09:44:57) to be greater than restore target (2026-03-26 09:45:26) + sleep 10 + [[ 1774518297 -gt 1774518326 ]] + [[ 4 -gt 30 ]] ++ get_latest_oplog_chunk_ts some-name ++ local cluster=some-name +++ kubectl_bin exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ jq '.backups.pitrChunks.pitrChunks | last | .range.end' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.xnBwxkGeP5 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.zra1WI8Aay +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.xnBwxkGeP5 +++ cat /tmp/tmp.zra1WI8Aay +++ rm /tmp/tmp.xnBwxkGeP5 /tmp/tmp.zra1WI8Aay +++ return 0 ++ echo 1774518297 + latest_ts=1774518297 + retries=5 ++ format_date 1774518297 ++ local timestamp=1774518297 +++ TZ=UTC +++ /usr/sbin/date -d@1774518297 '+%Y-%m-%d %H:%M:%S' ++ echo 2026-03-26 09:44:57 ++ format_date 1774518326 ++ local timestamp=1774518326 +++ TZ=UTC +++ /usr/sbin/date -d@1774518326 '+%Y-%m-%d %H:%M:%S' ++ echo 2026-03-26 09:45:26 + echo 'Waiting for last oplog chunk (2026-03-26 09:44:57) to be greater than restore target (2026-03-26 09:45:26)' Waiting for last oplog chunk (2026-03-26 09:44:57) to be greater than restore target (2026-03-26 09:45:26) + sleep 10 + [[ 1774518297 -gt 1774518326 ]] + [[ 5 -gt 30 ]] ++ get_latest_oplog_chunk_ts some-name ++ local cluster=some-name +++ jq '.backups.pitrChunks.pitrChunks | last | .range.end' +++ kubectl_bin exec some-name-rs0-0 -c backup-agent -- pbm status -o json ++++ mktemp +++ local LAST_OUT=/tmp/tmp.DktqbpyaEi ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ldiAZS3prt +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.DktqbpyaEi +++ cat /tmp/tmp.ldiAZS3prt +++ rm /tmp/tmp.DktqbpyaEi /tmp/tmp.ldiAZS3prt +++ return 0 ++ echo 1774518297 + latest_ts=1774518297 + retries=6 ++ format_date 1774518297 ++ local timestamp=1774518297 +++ TZ=UTC +++ /usr/sbin/date -d@1774518297 '+%Y-%m-%d %H:%M:%S' ++ echo 2026-03-26 09:44:57 ++ format_date 1774518326 ++ local timestamp=1774518326 +++ TZ=UTC +++ /usr/sbin/date -d@1774518326 '+%Y-%m-%d %H:%M:%S' ++ echo 2026-03-26 09:45:26 + echo 'Waiting for last oplog chunk (2026-03-26 09:44:57) to be greater than restore target (2026-03-26 09:45:26)' Waiting for last oplog chunk (2026-03-26 09:44:57) to be greater than restore target (2026-03-26 09:45:26) + sleep 10 + [[ 1774518297 -gt 1774518326 ]] + [[ 6 -gt 30 ]] ++ get_latest_oplog_chunk_ts some-name ++ local cluster=some-name +++ kubectl_bin exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ jq '.backups.pitrChunks.pitrChunks | last | .range.end' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.jMViZveWi5 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.kzPijrgXB9 +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.jMViZveWi5 +++ cat /tmp/tmp.kzPijrgXB9 +++ rm /tmp/tmp.jMViZveWi5 /tmp/tmp.kzPijrgXB9 +++ return 0 ++ echo 1774518297 + latest_ts=1774518297 + retries=7 ++ format_date 1774518297 ++ local timestamp=1774518297 +++ TZ=UTC +++ /usr/sbin/date -d@1774518297 '+%Y-%m-%d %H:%M:%S' ++ echo 2026-03-26 09:44:57 ++ format_date 1774518326 ++ local timestamp=1774518326 +++ TZ=UTC +++ /usr/sbin/date -d@1774518326 '+%Y-%m-%d %H:%M:%S' ++ echo 2026-03-26 09:45:26 + echo 'Waiting for last oplog chunk (2026-03-26 09:44:57) to be greater than restore target (2026-03-26 09:45:26)' Waiting for last oplog chunk (2026-03-26 09:44:57) to be greater than restore target (2026-03-26 09:45:26) + sleep 10 + [[ 1774518297 -gt 1774518326 ]] + [[ 7 -gt 30 ]] ++ get_latest_oplog_chunk_ts some-name ++ local cluster=some-name +++ kubectl_bin exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ jq '.backups.pitrChunks.pitrChunks | last | .range.end' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.8eMdarp4XI ++++ mktemp +++ local LAST_ERR=/tmp/tmp.jecUIbPIHJ +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.8eMdarp4XI +++ cat /tmp/tmp.jecUIbPIHJ +++ rm /tmp/tmp.8eMdarp4XI /tmp/tmp.jecUIbPIHJ +++ return 0 ++ echo 1774518297 + latest_ts=1774518297 + retries=8 ++ format_date 1774518297 ++ local timestamp=1774518297 +++ TZ=UTC +++ /usr/sbin/date -d@1774518297 '+%Y-%m-%d %H:%M:%S' ++ echo 2026-03-26 09:44:57 ++ format_date 1774518326 ++ local timestamp=1774518326 +++ TZ=UTC +++ /usr/sbin/date -d@1774518326 '+%Y-%m-%d %H:%M:%S' ++ echo 2026-03-26 09:45:26 + echo 'Waiting for last oplog chunk (2026-03-26 09:44:57) to be greater than restore target (2026-03-26 09:45:26)' Waiting for last oplog chunk (2026-03-26 09:44:57) to be greater than restore target (2026-03-26 09:45:26) + sleep 10 + [[ 1774518297 -gt 1774518326 ]] + [[ 8 -gt 30 ]] ++ get_latest_oplog_chunk_ts some-name ++ local cluster=some-name +++ kubectl_bin exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ jq '.backups.pitrChunks.pitrChunks | last | .range.end' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.IXFBTM37yU ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ZCENrsMwS9 +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.IXFBTM37yU +++ cat /tmp/tmp.ZCENrsMwS9 +++ rm /tmp/tmp.IXFBTM37yU /tmp/tmp.ZCENrsMwS9 +++ return 0 ++ echo 1774518299 + latest_ts=1774518299 + retries=9 ++ format_date 1774518299 ++ local timestamp=1774518299 +++ TZ=UTC +++ /usr/sbin/date -d@1774518299 '+%Y-%m-%d %H:%M:%S' ++ echo 2026-03-26 09:44:59 ++ format_date 1774518326 ++ local timestamp=1774518326 +++ TZ=UTC +++ /usr/sbin/date -d@1774518326 '+%Y-%m-%d %H:%M:%S' ++ echo 2026-03-26 09:45:26 + echo 'Waiting for last oplog chunk (2026-03-26 09:44:59) to be greater than restore target (2026-03-26 09:45:26)' Waiting for last oplog chunk (2026-03-26 09:44:59) to be greater than restore target (2026-03-26 09:45:26) + sleep 10 + [[ 1774518299 -gt 1774518326 ]] + [[ 9 -gt 30 ]] ++ get_latest_oplog_chunk_ts some-name ++ local cluster=some-name +++ kubectl_bin exec some-name-rs0-0 -c backup-agent -- pbm status -o json ++++ mktemp +++ jq '.backups.pitrChunks.pitrChunks | last | .range.end' +++ local LAST_OUT=/tmp/tmp.mSW3BNHIba ++++ mktemp +++ local LAST_ERR=/tmp/tmp.uLDpcY3gbu +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec some-name-rs0-0 -c backup-agent -- pbm status -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.mSW3BNHIba +++ cat /tmp/tmp.uLDpcY3gbu +++ rm /tmp/tmp.mSW3BNHIba /tmp/tmp.uLDpcY3gbu +++ return 0 ++ echo 1774518417 + latest_ts=1774518417 + retries=10 ++ format_date 1774518417 ++ local timestamp=1774518417 +++ TZ=UTC +++ /usr/sbin/date -d@1774518417 '+%Y-%m-%d %H:%M:%S' ++ echo 2026-03-26 09:46:57 ++ format_date 1774518326 ++ local timestamp=1774518326 +++ TZ=UTC +++ /usr/sbin/date -d@1774518326 '+%Y-%m-%d %H:%M:%S' ++ echo 2026-03-26 09:45:26 + echo 'Waiting for last oplog chunk (2026-03-26 09:46:57) to be greater than restore target (2026-03-26 09:45:26)' Waiting for last oplog chunk (2026-03-26 09:46:57) to be greater than restore target (2026-03-26 09:45:26) + sleep 10 + [[ 1774518417 -gt 1774518326 ]] + '[' -z backupSource ']' + desc 'check restore by date backupSource' + set +o xtrace ----------------------------------------------------------------------------------- check restore by date backupSource ----------------------------------------------------------------------------------- ++ get_backup_dest backup-minio-1 ++ local backup_name=backup-minio-1 ++ kubectl_bin get psmdb-backup backup-minio-1 -o 'jsonpath={.status.destination}' ++ sed -e 's/.json$//' ++ sed 's|s3://||' ++ sed 's|azure://||' +++ mktemp ++ sed 's|gs://||' ++ local LAST_OUT=/tmp/tmp.xrTKY0dZGY +++ mktemp ++ local LAST_ERR=/tmp/tmp.rkCuchwCN2 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb-backup backup-minio-1 -o 'jsonpath={.status.destination}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xrTKY0dZGY ++ cat /tmp/tmp.rkCuchwCN2 ++ rm /tmp/tmp.xrTKY0dZGY /tmp/tmp.rkCuchwCN2 ++ return 0 + backup_dest=operator-testing/2026-03-26T09:42:00Z + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2270/e2e-tests/pitr-physical-backup-source/conf/restore.yml + /usr/sbin/sed -e 's/name:/name: restore-backup-minio-1/' + /usr/sbin/sed -e /backupName/d + /usr/sbin/sed -e 's/pitrType:/type: date/' + '[' -z 1774518326 ']' ++ format_date 1774518326 ++ local timestamp=1774518326 + /usr/sbin/sed -e 's|DESTINATION|operator-testing/2026-03-26T09:42:00Z|' +++ TZ=UTC +++ /usr/sbin/date -d@1774518326 '+%Y-%m-%d %H:%M:%S' + '[' -n '' ']' + yq + kubectl_bin apply -f - ++ get_bucket_name backup-minio-1 ++ local backup_name=backup-minio-1 ++ kubectl_bin get psmdb-backup backup-minio-1 -o 'jsonpath={.status.s3.bucket}' ++ mktemp +++ mktemp + local LAST_OUT=/tmp/tmp.atJbYYQCOE ++ mktemp ++ echo 2026-03-26 09:45:26 + /usr/sbin/sed -e 's/date:/date: 2026-03-26 09:45:26/' ++ local LAST_OUT=/tmp/tmp.JL7sfioqsi + local LAST_ERR=/tmp/tmp.g0lcsKfYCO + local exit_status=0 + local timeout=4 +++ mktemp ++ seq 0 2 ++ local LAST_ERR=/tmp/tmp.H3vQUnKSiA ++ local exit_status=0 ++ local timeout=4 + for i in $(seq 0 2) + set +e + kubectl apply -f - +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb-backup backup-minio-1 -o 'jsonpath={.status.s3.bucket}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.JL7sfioqsi ++ cat /tmp/tmp.H3vQUnKSiA ++ rm /tmp/tmp.JL7sfioqsi /tmp/tmp.H3vQUnKSiA ++ return 0 + /usr/sbin/sed -e 's|BUCKET-NAME|operator-testing|' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.atJbYYQCOE perconaservermongodbrestore.psmdb.percona.com/restore-backup-minio-1 created + cat /tmp/tmp.g0lcsKfYCO + rm /tmp/tmp.atJbYYQCOE /tmp/tmp.g0lcsKfYCO + return 0 + wait_restore backup-minio-1 some-name requested 0 1200 + local backup_name=backup-minio-1 + local cluster_name=some-name + local target_state=requested + local wait_cluster_consistency=0 + local wait_time=1200 + local ok_if_ready=0 + set +o xtrace Waiting for the psmdb-restore/restore-backup-minio-1 object to be created.OK Waiting psmdb-restore/restore-backup-minio-1 to reach state "requested" ........OK after 7 minutes + [[ 0 -eq 1 ]] + echo + wait_restore backup-minio-1 some-name ready 0 1600 + local backup_name=backup-minio-1 + local cluster_name=some-name + local target_state=ready + local wait_cluster_consistency=0 + local wait_time=1600 + local ok_if_ready=0 + set +o xtrace Waiting for the psmdb-restore/restore-backup-minio-1 object to be created.OK Waiting psmdb-restore/restore-backup-minio-1 to reach state "ready" .......................... ----------------------------------------------------------------------------------- pbm-agent logs from some-name-rs0-0 ----------------------------------------------------------------------------------- + kubectl_bin exec some-name-rs0-0 -- cat /tmp/pbm-agent.log ++ mktemp + local LAST_OUT=/tmp/tmp.fFOcUqtMDh ++ mktemp + local LAST_ERR=/tmp/tmp.HH9FOMFcWf + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec some-name-rs0-0 -- cat /tmp/pbm-agent.log + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.fFOcUqtMDh 2026/03/26 09:57:22 [entrypoint] starting `pbm-agent` 2026-03-26T09:57:24.000+0000 I % _____ %%% | __ \ ###%%%%%%%%%%%%* | |__) |__ _ __ ___ ___ _ __ __ _ ### ##%% %%%% | ___/ _ \ '__/ __/ _ \| '_ \ / _` | #### ##% %%%% | | | __/ | | (_| (_) | | | | (_| | ### #### %%% |_| \___|_| \___\___/|_| |_|\__,_| ,((### ### %%% _____ _ (((( (### #### %%%% / ____| | | ((( ((# ###### | (___ __ _ _ _ __ _ __| | (((( (((# #### \___ \ / _` | | | |/ _` |/ _` | /(( ,((( *### ____) | (_| | |_| | (_| | (_| | //// ((( #### |_____/ \__, |\__,_|\__,_|\__,_| /// (((( #### | | /////////////(((((((((((((((((######## |_| Join @ squad.percona.com/mongodb ** Join Percona Squad! ** Participate in monthly SWAG raffles, get early access to new product features, invite-only ”ask me anything” sessions with database performance experts. Interested? Fill in the form at squad.percona.com/mongodb 2026-03-26T09:57:24.000+0000 I log options: log-path=/dev/stderr, log-level:D, log-json:false 2026-03-26T09:57:24.000+0000 I pbm-agent: Version: 2.13.0 Platform: linux/amd64 GitCommit: e9b57e1395fe2602bd9637d5e4ef19b1df2bb3b7 GitBranch: release-2.13.0 BuildTime: 2026-02-26_12:41_UTC GoVersion: go1.25.7 2026-03-26T09:57:24.000+0000 I starting PITR routine 2026-03-26T09:57:24.000+0000 I node: rs0/some-name-rs0-0.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017 2026-03-26T09:57:24.000+0000 I conn level ReadConcern: majority; WriteConcern: majority 2026-03-26T09:57:24.000+0000 I listening for the commands + cat /tmp/tmp.HH9FOMFcWf Defaulted container "mongod" out of: mongod, mongo-init (init), pbm-init (init) + rm /tmp/tmp.fFOcUqtMDh /tmp/tmp.HH9FOMFcWf + return 0 + for pod in $(get_mongod_pods ${cluster}) + desc 'pbm-agent logs from some-name-rs0-1' + set +o xtrace ----------------------------------------------------------------------------------- pbm-agent logs from some-name-rs0-1 ----------------------------------------------------------------------------------- + kubectl_bin exec some-name-rs0-1 -- cat /tmp/pbm-agent.log ++ mktemp + local LAST_OUT=/tmp/tmp.g0X9J1G3WN ++ mktemp + local LAST_ERR=/tmp/tmp.4grvbCHuio + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec some-name-rs0-1 -- cat /tmp/pbm-agent.log + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.g0X9J1G3WN 2026/03/26 09:50:42 [entrypoint] starting `pbm-agent` 2026-03-26T09:50:44.000+0000 I % _____ %%% | __ \ ###%%%%%%%%%%%%* | |__) |__ _ __ ___ ___ _ __ __ _ ### ##%% %%%% | ___/ _ \ '__/ __/ _ \| '_ \ / _` | #### ##% %%%% | | | __/ | | (_| (_) | | | | (_| | ### #### %%% |_| \___|_| \___\___/|_| |_|\__,_| ,((### ### %%% _____ _ (((( (### #### %%%% / ____| | | ((( ((# ###### | (___ __ _ _ _ __ _ __| | (((( (((# #### \___ \ / _` | | | |/ _` |/ _` | /(( ,((( *### ____) | (_| | |_| | (_| | (_| | //// ((( #### |_____/ \__, |\__,_|\__,_|\__,_| /// (((( #### | | /////////////(((((((((((((((((######## |_| Join @ squad.percona.com/mongodb ** Join Percona Squad! ** Participate in monthly SWAG raffles, get early access to new product features, invite-only ”ask me anything” sessions with database performance experts. Interested? Fill in the form at squad.percona.com/mongodb 2026-03-26T09:50:44.000+0000 I log options: log-path=/dev/stderr, log-level:D, log-json:false 2026-03-26T09:50:44.000+0000 I pbm-agent: Version: 2.13.0 Platform: linux/amd64 GitCommit: e9b57e1395fe2602bd9637d5e4ef19b1df2bb3b7 GitBranch: release-2.13.0 BuildTime: 2026-02-26_12:41_UTC GoVersion: go1.25.7 2026-03-26T09:50:44.000+0000 I starting PITR routine 2026-03-26T09:50:44.000+0000 I node: rs0/some-name-rs0-1.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017 2026-03-26T09:50:44.000+0000 I conn level ReadConcern: majority; WriteConcern: majority 2026-03-26T09:50:44.000+0000 I listening for the commands 2026-03-26T09:55:42.000+0000 I got command restore [name: 2026-03-26T09:55:42.160775713Z, snapshot: 2026-03-26T09:42:00Z point-in-time: <1774518326,0>] , opid: 69c5029e01503e098e91130e 2026-03-26T09:55:42.000+0000 I got epoch {1774518473 3} 2026-03-26T09:55:42.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] to time: 2026-03-26T09:45:26Z 2026-03-26T09:55:42.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] backup: 2026-03-26T09:42:00Z 2026-03-26T09:55:42.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] recovery started 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] port: 27595 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-1.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] download max buf 1073741824 (arena 268435456, span 33554432, concurrency 4) 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] mongod binary: mongod, version: v8.0.19-7 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] restore opts: fallbackEnabled: false; allowPartlyDone: true 2026-03-26T09:55:42.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] moving to state starting 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-1.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.starting" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:42.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] waiting for cluster 2026-03-26T09:57:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-1.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:57:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:57:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:59:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-1.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:59:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:59:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:01:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-1.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:01:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:01:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:03:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-1.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:03:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:03:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:05:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-1.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:05:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:05:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:07:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-1.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:07:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:07:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:09:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-1.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:09:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:09:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:11:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-1.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:11:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:11:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:13:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-1.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:13:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:13:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:15:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-1.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:15:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:15:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:17:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-1.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:17:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:17:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:19:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-1.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:19:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:19:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:21:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-1.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:21:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:21:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] + cat /tmp/tmp.4grvbCHuio Defaulted container "mongod" out of: mongod, mongo-init (init), pbm-init (init) + rm /tmp/tmp.g0X9J1G3WN /tmp/tmp.4grvbCHuio + return 0 + for pod in $(get_mongod_pods ${cluster}) + desc 'pbm-agent logs from some-name-rs0-2' + set +o xtrace ----------------------------------------------------------------------------------- pbm-agent logs from some-name-rs0-2 ----------------------------------------------------------------------------------- + kubectl_bin exec some-name-rs0-2 -- cat /tmp/pbm-agent.log ++ mktemp + local LAST_OUT=/tmp/tmp.w25MSncelX ++ mktemp + local LAST_ERR=/tmp/tmp.4FRSjJjqJr + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec some-name-rs0-2 -- cat /tmp/pbm-agent.log + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.w25MSncelX 2026/03/26 09:50:02 [entrypoint] starting `pbm-agent` 2026-03-26T09:50:04.000+0000 I % _____ %%% | __ \ ###%%%%%%%%%%%%* | |__) |__ _ __ ___ ___ _ __ __ _ ### ##%% %%%% | ___/ _ \ '__/ __/ _ \| '_ \ / _` | #### ##% %%%% | | | __/ | | (_| (_) | | | | (_| | ### #### %%% |_| \___|_| \___\___/|_| |_|\__,_| ,((### ### %%% _____ _ (((( (### #### %%%% / ____| | | ((( ((# ###### | (___ __ _ _ _ __ _ __| | (((( (((# #### \___ \ / _` | | | |/ _` |/ _` | /(( ,((( *### ____) | (_| | |_| | (_| | (_| | //// ((( #### |_____/ \__, |\__,_|\__,_|\__,_| /// (((( #### | | /////////////(((((((((((((((((######## |_| Join @ squad.percona.com/mongodb ** Join Percona Squad! ** Participate in monthly SWAG raffles, get early access to new product features, invite-only ”ask me anything” sessions with database performance experts. Interested? Fill in the form at squad.percona.com/mongodb 2026-03-26T09:50:04.000+0000 I log options: log-path=/dev/stderr, log-level:D, log-json:false 2026-03-26T09:50:04.000+0000 I pbm-agent: Version: 2.13.0 Platform: linux/amd64 GitCommit: e9b57e1395fe2602bd9637d5e4ef19b1df2bb3b7 GitBranch: release-2.13.0 BuildTime: 2026-02-26_12:41_UTC GoVersion: go1.25.7 2026-03-26T09:50:04.000+0000 I starting PITR routine 2026-03-26T09:50:04.000+0000 I node: rs0/some-name-rs0-2.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017 2026-03-26T09:50:04.000+0000 I conn level ReadConcern: majority; WriteConcern: majority 2026-03-26T09:50:04.000+0000 I listening for the commands 2026-03-26T09:55:42.000+0000 I got command restore [name: 2026-03-26T09:55:42.160775713Z, snapshot: 2026-03-26T09:42:00Z point-in-time: <1774518326,0>] , opid: 69c5029e01503e098e91130e 2026-03-26T09:55:42.000+0000 I got epoch {1774518942 83} 2026-03-26T09:55:42.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] to time: 2026-03-26T09:45:26Z 2026-03-26T09:55:42.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] backup: 2026-03-26T09:42:00Z 2026-03-26T09:55:42.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] recovery started 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] port: 27519 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-2.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] download max buf 1073741824 (arena 268435456, span 33554432, concurrency 4) 2026-03-26T09:55:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] mongod binary: mongod, version: v8.0.19-7 2026-03-26T09:55:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] restore opts: fallbackEnabled: false; allowPartlyDone: true 2026-03-26T09:55:43.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] moving to state starting 2026-03-26T09:55:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-2.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.starting" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:43.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] waiting for cluster 2026-03-26T09:57:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-2.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:57:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:57:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:59:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-2.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:59:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:59:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:01:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-2.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:01:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:01:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:03:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-2.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:03:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:03:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:05:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-2.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:05:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:05:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:07:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-2.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:07:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:07:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:09:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-2.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:09:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:09:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:11:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-2.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:11:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:11:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:13:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-2.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:13:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:13:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:15:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-2.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:15:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:15:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:17:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-2.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:17:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:17:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:19:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-2.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:19:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:19:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:21:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/node.some-name-rs0-2.some-name-rs0.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:21:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs0/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:21:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] + cat /tmp/tmp.4FRSjJjqJr Defaulted container "mongod" out of: mongod, mongo-init (init), pbm-init (init) + rm /tmp/tmp.w25MSncelX /tmp/tmp.4FRSjJjqJr + return 0 + for pod in $(get_mongod_pods ${cluster}) + desc 'pbm-agent logs from some-name-rs1-0' + set +o xtrace ----------------------------------------------------------------------------------- pbm-agent logs from some-name-rs1-0 ----------------------------------------------------------------------------------- + kubectl_bin exec some-name-rs1-0 -- cat /tmp/pbm-agent.log ++ mktemp + local LAST_OUT=/tmp/tmp.dKJqpFhR0w ++ mktemp + local LAST_ERR=/tmp/tmp.GOTWNNhjTD + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec some-name-rs1-0 -- cat /tmp/pbm-agent.log + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.dKJqpFhR0w 2026/03/26 09:53:05 [entrypoint] starting `pbm-agent` 2026-03-26T09:53:09.000+0000 I % _____ %%% | __ \ ###%%%%%%%%%%%%* | |__) |__ _ __ ___ ___ _ __ __ _ ### ##%% %%%% | ___/ _ \ '__/ __/ _ \| '_ \ / _` | #### ##% %%%% | | | __/ | | (_| (_) | | | | (_| | ### #### %%% |_| \___|_| \___\___/|_| |_|\__,_| ,((### ### %%% _____ _ (((( (### #### %%%% / ____| | | ((( ((# ###### | (___ __ _ _ _ __ _ __| | (((( (((# #### \___ \ / _` | | | |/ _` |/ _` | /(( ,((( *### ____) | (_| | |_| | (_| | (_| | //// ((( #### |_____/ \__, |\__,_|\__,_|\__,_| /// (((( #### | | /////////////(((((((((((((((((######## |_| Join @ squad.percona.com/mongodb ** Join Percona Squad! ** Participate in monthly SWAG raffles, get early access to new product features, invite-only ”ask me anything” sessions with database performance experts. Interested? Fill in the form at squad.percona.com/mongodb 2026-03-26T09:53:09.000+0000 I log options: log-path=/dev/stderr, log-level:D, log-json:false 2026-03-26T09:53:09.000+0000 I pbm-agent: Version: 2.13.0 Platform: linux/amd64 GitCommit: e9b57e1395fe2602bd9637d5e4ef19b1df2bb3b7 GitBranch: release-2.13.0 BuildTime: 2026-02-26_12:41_UTC GoVersion: go1.25.7 2026-03-26T09:53:09.000+0000 I starting PITR routine 2026-03-26T09:53:09.000+0000 I node: rs1/some-name-rs1-0.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017 2026-03-26T09:53:09.000+0000 I conn level ReadConcern: majority; WriteConcern: majority 2026-03-26T09:53:09.000+0000 I listening for the commands 2026-03-26T09:55:43.000+0000 I got command restore [name: 2026-03-26T09:55:42.160775713Z, snapshot: 2026-03-26T09:42:00Z point-in-time: <1774518326,0>] , opid: 69c5029e01503e098e91130e 2026-03-26T09:55:43.000+0000 I got epoch {1774518942 114} 2026-03-26T09:55:43.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] to time: 2026-03-26T09:45:26Z 2026-03-26T09:55:43.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] oplog slicer disabled 2026-03-26T09:55:43.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] backup: 2026-03-26T09:42:00Z 2026-03-26T09:55:43.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] recovery started 2026-03-26T09:55:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] port: 27556 2026-03-26T09:55:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-0.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] download max buf 536870912 (arena 268435456, span 33554432, concurrency 2) 2026-03-26T09:55:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] mongod binary: mongod, version: v8.0.19-7 2026-03-26T09:55:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] restore opts: fallbackEnabled: false; allowPartlyDone: true 2026-03-26T09:55:43.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] moving to state starting 2026-03-26T09:55:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-0.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.starting" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:43.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] waiting for `starting` status in rs map[.pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-0.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017:{} .pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-1.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017:{} .pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-2.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017:{}] 2026-03-26T09:55:48.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.starting" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:48.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] waiting for cluster 2026-03-26T09:57:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-0.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:57:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:57:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:59:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-0.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:59:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:59:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:01:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-0.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:01:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:01:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:03:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-0.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:03:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:03:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:05:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-0.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:05:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:05:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:07:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-0.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:07:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:07:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:09:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-0.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:09:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:09:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:11:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-0.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:11:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:11:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:13:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-0.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:13:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:13:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:15:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-0.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:15:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:15:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:17:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-0.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:17:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:17:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:19:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-0.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:19:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:19:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:21:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-0.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:21:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:21:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] + cat /tmp/tmp.GOTWNNhjTD Defaulted container "mongod" out of: mongod, mongo-init (init), pbm-init (init) + rm /tmp/tmp.dKJqpFhR0w /tmp/tmp.GOTWNNhjTD + return 0 + for pod in $(get_mongod_pods ${cluster}) + desc 'pbm-agent logs from some-name-rs1-1' + set +o xtrace ----------------------------------------------------------------------------------- pbm-agent logs from some-name-rs1-1 ----------------------------------------------------------------------------------- + kubectl_bin exec some-name-rs1-1 -- cat /tmp/pbm-agent.log ++ mktemp + local LAST_OUT=/tmp/tmp.bbqqS5oYxr ++ mktemp + local LAST_ERR=/tmp/tmp.ztyqcdD0qM + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec some-name-rs1-1 -- cat /tmp/pbm-agent.log + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.bbqqS5oYxr 2026/03/26 09:57:22 [entrypoint] starting `pbm-agent` 2026-03-26T09:57:26.000+0000 I % _____ %%% | __ \ ###%%%%%%%%%%%%* | |__) |__ _ __ ___ ___ _ __ __ _ ### ##%% %%%% | ___/ _ \ '__/ __/ _ \| '_ \ / _` | #### ##% %%%% | | | __/ | | (_| (_) | | | | (_| | ### #### %%% |_| \___|_| \___\___/|_| |_|\__,_| ,((### ### %%% _____ _ (((( (### #### %%%% / ____| | | ((( ((# ###### | (___ __ _ _ _ __ _ __| | (((( (((# #### \___ \ / _` | | | |/ _` |/ _` | /(( ,((( *### ____) | (_| | |_| | (_| | (_| | //// ((( #### |_____/ \__, |\__,_|\__,_|\__,_| /// (((( #### | | /////////////(((((((((((((((((######## |_| Join @ squad.percona.com/mongodb ** Join Percona Squad! ** Participate in monthly SWAG raffles, get early access to new product features, invite-only ”ask me anything” sessions with database performance experts. Interested? Fill in the form at squad.percona.com/mongodb 2026-03-26T09:57:26.000+0000 I log options: log-path=/dev/stderr, log-level:D, log-json:false 2026-03-26T09:57:26.000+0000 I pbm-agent: Version: 2.13.0 Platform: linux/amd64 GitCommit: e9b57e1395fe2602bd9637d5e4ef19b1df2bb3b7 GitBranch: release-2.13.0 BuildTime: 2026-02-26_12:41_UTC GoVersion: go1.25.7 2026-03-26T09:57:26.000+0000 I starting PITR routine 2026-03-26T09:57:26.000+0000 I node: rs1/some-name-rs1-1.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017 2026-03-26T09:57:26.000+0000 I conn level ReadConcern: majority; WriteConcern: majority 2026-03-26T09:57:26.000+0000 I listening for the commands + cat /tmp/tmp.ztyqcdD0qM Defaulted container "mongod" out of: mongod, mongo-init (init), pbm-init (init) + rm /tmp/tmp.bbqqS5oYxr /tmp/tmp.ztyqcdD0qM + return 0 + for pod in $(get_mongod_pods ${cluster}) + desc 'pbm-agent logs from some-name-rs1-2' + set +o xtrace ----------------------------------------------------------------------------------- pbm-agent logs from some-name-rs1-2 ----------------------------------------------------------------------------------- + kubectl_bin exec some-name-rs1-2 -- cat /tmp/pbm-agent.log ++ mktemp + local LAST_OUT=/tmp/tmp.2RpoFOrzss ++ mktemp + local LAST_ERR=/tmp/tmp.Tw2V3jpvzx + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec some-name-rs1-2 -- cat /tmp/pbm-agent.log + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.2RpoFOrzss 2026/03/26 09:51:54 [entrypoint] starting `pbm-agent` 2026-03-26T09:51:58.000+0000 I % _____ %%% | __ \ ###%%%%%%%%%%%%* | |__) |__ _ __ ___ ___ _ __ __ _ ### ##%% %%%% | ___/ _ \ '__/ __/ _ \| '_ \ / _` | #### ##% %%%% | | | __/ | | (_| (_) | | | | (_| | ### #### %%% |_| \___|_| \___\___/|_| |_|\__,_| ,((### ### %%% _____ _ (((( (### #### %%%% / ____| | | ((( ((# ###### | (___ __ _ _ _ __ _ __| | (((( (((# #### \___ \ / _` | | | |/ _` |/ _` | /(( ,((( *### ____) | (_| | |_| | (_| | (_| | //// ((( #### |_____/ \__, |\__,_|\__,_|\__,_| /// (((( #### | | /////////////(((((((((((((((((######## |_| Join @ squad.percona.com/mongodb ** Join Percona Squad! ** Participate in monthly SWAG raffles, get early access to new product features, invite-only ”ask me anything” sessions with database performance experts. Interested? Fill in the form at squad.percona.com/mongodb 2026-03-26T09:51:58.000+0000 I log options: log-path=/dev/stderr, log-level:D, log-json:false 2026-03-26T09:51:58.000+0000 I pbm-agent: Version: 2.13.0 Platform: linux/amd64 GitCommit: e9b57e1395fe2602bd9637d5e4ef19b1df2bb3b7 GitBranch: release-2.13.0 BuildTime: 2026-02-26_12:41_UTC GoVersion: go1.25.7 2026-03-26T09:51:58.000+0000 I starting PITR routine 2026-03-26T09:51:58.000+0000 I node: rs1/some-name-rs1-2.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017 2026-03-26T09:51:58.000+0000 I conn level ReadConcern: majority; WriteConcern: majority 2026-03-26T09:51:58.000+0000 I listening for the commands 2026-03-26T09:55:42.000+0000 I got command restore [name: 2026-03-26T09:55:42.160775713Z, snapshot: 2026-03-26T09:42:00Z point-in-time: <1774518326,0>] , opid: 69c5029e01503e098e91130e 2026-03-26T09:55:42.000+0000 I got epoch {1774518473 3} 2026-03-26T09:55:42.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] to time: 2026-03-26T09:45:26Z 2026-03-26T09:55:42.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] backup: 2026-03-26T09:42:00Z 2026-03-26T09:55:42.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] recovery started 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] port: 27861 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-2.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] download max buf 536870912 (arena 268435456, span 33554432, concurrency 2) 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] mongod binary: mongod, version: v8.0.19-7 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] restore opts: fallbackEnabled: false; allowPartlyDone: true 2026-03-26T09:55:42.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] moving to state starting 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-2.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.starting" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:42.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] waiting for cluster 2026-03-26T09:57:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-2.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:57:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:57:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:59:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-2.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:59:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:59:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:01:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-2.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:01:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:01:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:03:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-2.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:03:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:03:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:05:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-2.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:05:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:05:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:07:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-2.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:07:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:07:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:09:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-2.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:09:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:09:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:11:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-2.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:11:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:11:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:13:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-2.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:13:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:13:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:15:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-2.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:15:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:15:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:17:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-2.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:17:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:17:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:19:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-2.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:19:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:19:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:21:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/node.some-name-rs1-2.some-name-rs1.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:21:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs1/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:21:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] + cat /tmp/tmp.Tw2V3jpvzx Defaulted container "mongod" out of: mongod, mongo-init (init), pbm-init (init) + rm /tmp/tmp.2RpoFOrzss /tmp/tmp.Tw2V3jpvzx + return 0 + for pod in $(get_mongod_pods ${cluster}) + desc 'pbm-agent logs from some-name-rs2-0' + set +o xtrace ----------------------------------------------------------------------------------- pbm-agent logs from some-name-rs2-0 ----------------------------------------------------------------------------------- + kubectl_bin exec some-name-rs2-0 -- cat /tmp/pbm-agent.log ++ mktemp + local LAST_OUT=/tmp/tmp.nol4gz861g ++ mktemp + local LAST_ERR=/tmp/tmp.niClDDCVEB + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec some-name-rs2-0 -- cat /tmp/pbm-agent.log + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.nol4gz861g 2026/03/26 09:54:47 [entrypoint] starting `pbm-agent` 2026-03-26T09:54:51.000+0000 I % _____ %%% | __ \ ###%%%%%%%%%%%%* | |__) |__ _ __ ___ ___ _ __ __ _ ### ##%% %%%% | ___/ _ \ '__/ __/ _ \| '_ \ / _` | #### ##% %%%% | | | __/ | | (_| (_) | | | | (_| | ### #### %%% |_| \___|_| \___\___/|_| |_|\__,_| ,((### ### %%% _____ _ (((( (### #### %%%% / ____| | | ((( ((# ###### | (___ __ _ _ _ __ _ __| | (((( (((# #### \___ \ / _` | | | |/ _` |/ _` | /(( ,((( *### ____) | (_| | |_| | (_| | (_| | //// ((( #### |_____/ \__, |\__,_|\__,_|\__,_| /// (((( #### | | /////////////(((((((((((((((((######## |_| Join @ squad.percona.com/mongodb ** Join Percona Squad! ** Participate in monthly SWAG raffles, get early access to new product features, invite-only ”ask me anything” sessions with database performance experts. Interested? Fill in the form at squad.percona.com/mongodb 2026-03-26T09:54:51.000+0000 I log options: log-path=/dev/stderr, log-level:D, log-json:false 2026-03-26T09:54:51.000+0000 I pbm-agent: Version: 2.13.0 Platform: linux/amd64 GitCommit: e9b57e1395fe2602bd9637d5e4ef19b1df2bb3b7 GitBranch: release-2.13.0 BuildTime: 2026-02-26_12:41_UTC GoVersion: go1.25.7 2026-03-26T09:54:51.000+0000 I starting PITR routine 2026-03-26T09:54:51.000+0000 I node: rs2/some-name-rs2-0.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017 2026-03-26T09:54:51.000+0000 I conn level ReadConcern: majority; WriteConcern: majority 2026-03-26T09:54:51.000+0000 I listening for the commands 2026-03-26T09:55:42.000+0000 I got command restore [name: 2026-03-26T09:55:42.160775713Z, snapshot: 2026-03-26T09:42:00Z point-in-time: <1774518326,0>] , opid: 69c5029e01503e098e91130e 2026-03-26T09:55:42.000+0000 I got epoch {1774518942 83} 2026-03-26T09:55:42.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] to time: 2026-03-26T09:45:26Z 2026-03-26T09:55:42.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] oplog slicer disabled 2026-03-26T09:55:42.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] backup: 2026-03-26T09:42:00Z 2026-03-26T09:55:42.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] recovery started 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] port: 27395 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-0.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] download max buf 536870912 (arena 268435456, span 33554432, concurrency 2) 2026-03-26T09:55:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] mongod binary: mongod, version: v8.0.19-7 2026-03-26T09:55:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] restore opts: fallbackEnabled: false; allowPartlyDone: true 2026-03-26T09:55:43.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] moving to state starting 2026-03-26T09:55:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-0.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.starting" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:43.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] waiting for `starting` status in rs map[.pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-0.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017:{} .pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-1.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017:{} .pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-2.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017:{}] 2026-03-26T09:55:48.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.starting" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:48.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] waiting for cluster 2026-03-26T09:57:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-0.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:57:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:57:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:59:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-0.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:59:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:59:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:01:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-0.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:01:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:01:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:03:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-0.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:03:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:03:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:05:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-0.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:05:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:05:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:07:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-0.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:07:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:07:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:09:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-0.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:09:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:09:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:11:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-0.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:11:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:11:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:13:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-0.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:13:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:13:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:15:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-0.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:15:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:15:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:17:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-0.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:17:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:17:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:19:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-0.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:19:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:19:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:21:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-0.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:21:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:21:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] + cat /tmp/tmp.niClDDCVEB Defaulted container "mongod" out of: mongod, mongo-init (init), pbm-init (init) + rm /tmp/tmp.nol4gz861g /tmp/tmp.niClDDCVEB + return 0 + for pod in $(get_mongod_pods ${cluster}) + desc 'pbm-agent logs from some-name-rs2-1' + set +o xtrace ----------------------------------------------------------------------------------- pbm-agent logs from some-name-rs2-1 ----------------------------------------------------------------------------------- + kubectl_bin exec some-name-rs2-1 -- cat /tmp/pbm-agent.log ++ mktemp + local LAST_OUT=/tmp/tmp.uLhOBlYW8U ++ mktemp + local LAST_ERR=/tmp/tmp.cLT95yHkyk + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec some-name-rs2-1 -- cat /tmp/pbm-agent.log + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.uLhOBlYW8U 2026/03/26 09:56:45 [entrypoint] starting `pbm-agent` 2026-03-26T09:57:14.000+0000 I % _____ %%% | __ \ ###%%%%%%%%%%%%* | |__) |__ _ __ ___ ___ _ __ __ _ ### ##%% %%%% | ___/ _ \ '__/ __/ _ \| '_ \ / _` | #### ##% %%%% | | | __/ | | (_| (_) | | | | (_| | ### #### %%% |_| \___|_| \___\___/|_| |_|\__,_| ,((### ### %%% _____ _ (((( (### #### %%%% / ____| | | ((( ((# ###### | (___ __ _ _ _ __ _ __| | (((( (((# #### \___ \ / _` | | | |/ _` |/ _` | /(( ,((( *### ____) | (_| | |_| | (_| | (_| | //// ((( #### |_____/ \__, |\__,_|\__,_|\__,_| /// (((( #### | | /////////////(((((((((((((((((######## |_| Join @ squad.percona.com/mongodb ** Join Percona Squad! ** Participate in monthly SWAG raffles, get early access to new product features, invite-only ”ask me anything” sessions with database performance experts. Interested? Fill in the form at squad.percona.com/mongodb 2026-03-26T09:57:14.000+0000 I log options: log-path=/dev/stderr, log-level:D, log-json:false 2026-03-26T09:57:14.000+0000 I pbm-agent: Version: 2.13.0 Platform: linux/amd64 GitCommit: e9b57e1395fe2602bd9637d5e4ef19b1df2bb3b7 GitBranch: release-2.13.0 BuildTime: 2026-02-26_12:41_UTC GoVersion: go1.25.7 2026-03-26T09:57:14.000+0000 I starting PITR routine 2026-03-26T09:57:14.000+0000 I node: rs2/some-name-rs2-1.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017 2026-03-26T09:57:14.000+0000 I conn level ReadConcern: majority; WriteConcern: majority 2026-03-26T09:57:14.000+0000 I listening for the commands + cat /tmp/tmp.cLT95yHkyk Defaulted container "mongod" out of: mongod, mongo-init (init), pbm-init (init) + rm /tmp/tmp.uLhOBlYW8U /tmp/tmp.cLT95yHkyk + return 0 + for pod in $(get_mongod_pods ${cluster}) + desc 'pbm-agent logs from some-name-rs2-2' + set +o xtrace ----------------------------------------------------------------------------------- pbm-agent logs from some-name-rs2-2 ----------------------------------------------------------------------------------- + kubectl_bin exec some-name-rs2-2 -- cat /tmp/pbm-agent.log ++ mktemp + local LAST_OUT=/tmp/tmp.O5TCacMmzy ++ mktemp + local LAST_ERR=/tmp/tmp.TjjnUqag34 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec some-name-rs2-2 -- cat /tmp/pbm-agent.log + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.O5TCacMmzy 2026/03/26 09:53:40 [entrypoint] starting `pbm-agent` 2026-03-26T09:53:44.000+0000 I % _____ %%% | __ \ ###%%%%%%%%%%%%* | |__) |__ _ __ ___ ___ _ __ __ _ ### ##%% %%%% | ___/ _ \ '__/ __/ _ \| '_ \ / _` | #### ##% %%%% | | | __/ | | (_| (_) | | | | (_| | ### #### %%% |_| \___|_| \___\___/|_| |_|\__,_| ,((### ### %%% _____ _ (((( (### #### %%%% / ____| | | ((( ((# ###### | (___ __ _ _ _ __ _ __| | (((( (((# #### \___ \ / _` | | | |/ _` |/ _` | /(( ,((( *### ____) | (_| | |_| | (_| | (_| | //// ((( #### |_____/ \__, |\__,_|\__,_|\__,_| /// (((( #### | | /////////////(((((((((((((((((######## |_| Join @ squad.percona.com/mongodb ** Join Percona Squad! ** Participate in monthly SWAG raffles, get early access to new product features, invite-only ”ask me anything” sessions with database performance experts. Interested? Fill in the form at squad.percona.com/mongodb 2026-03-26T09:53:44.000+0000 I log options: log-path=/dev/stderr, log-level:D, log-json:false 2026-03-26T09:53:44.000+0000 I pbm-agent: Version: 2.13.0 Platform: linux/amd64 GitCommit: e9b57e1395fe2602bd9637d5e4ef19b1df2bb3b7 GitBranch: release-2.13.0 BuildTime: 2026-02-26_12:41_UTC GoVersion: go1.25.7 2026-03-26T09:53:44.000+0000 I starting PITR routine 2026-03-26T09:53:44.000+0000 I node: rs2/some-name-rs2-2.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017 2026-03-26T09:53:44.000+0000 I conn level ReadConcern: majority; WriteConcern: majority 2026-03-26T09:53:44.000+0000 I listening for the commands 2026-03-26T09:55:42.000+0000 I got command restore [name: 2026-03-26T09:55:42.160775713Z, snapshot: 2026-03-26T09:42:00Z point-in-time: <1774518326,0>] , opid: 69c5029e01503e098e91130e 2026-03-26T09:55:42.000+0000 I got epoch {1774518942 114} 2026-03-26T09:55:42.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] to time: 2026-03-26T09:45:26Z 2026-03-26T09:55:42.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] backup: 2026-03-26T09:42:00Z 2026-03-26T09:55:42.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] recovery started 2026-03-26T09:55:42.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] port: 27715 2026-03-26T09:55:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-2.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] download max buf 536870912 (arena 268435456, span 33554432, concurrency 2) 2026-03-26T09:55:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] mongod binary: mongod, version: v8.0.19-7 2026-03-26T09:55:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] restore opts: fallbackEnabled: false; allowPartlyDone: true 2026-03-26T09:55:43.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] moving to state starting 2026-03-26T09:55:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-2.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.starting" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:55:43.000+0000 I [restore/2026-03-26T09:55:42.160775713Z] waiting for cluster 2026-03-26T09:57:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-2.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:57:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:57:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:59:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-2.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:59:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T09:59:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:01:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-2.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:01:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:01:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:03:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-2.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:03:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:03:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:05:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-2.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:05:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:05:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:07:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-2.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:07:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:07:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:09:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-2.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:09:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:09:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:11:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-2.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:11:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:11:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:13:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-2.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:13:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:13:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:15:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-2.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:15:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:15:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:17:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-2.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:17:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:17:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:19:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-2.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:19:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:19:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:21:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/node.some-name-rs2-2.some-name-rs2.pitr-physical-backup-source-28911.svc.cluster.local:27017.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:21:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/rs.rs2/rs.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] 2026-03-26T10:21:43.000+0000 D [restore/2026-03-26T09:55:42.160775713Z] uploading ".pbm.restore/2026-03-26T09:55:42.160775713Z/cluster.hb" [size hint: 10 (10.00B); part size: 10485760 (10.00MB)] + cat /tmp/tmp.TjjnUqag34 Defaulted container "mongod" out of: mongod, mongo-init (init), pbm-init (init) + rm /tmp/tmp.O5TCacMmzy /tmp/tmp.TjjnUqag34 + return 0 + grep restore-backup-minio-1 + tail -100 ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.z8gUDhEmpa +++ mktemp ++ local LAST_ERR=/tmp/tmp.hFs8UVwLPE ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.z8gUDhEmpa ++ cat /tmp/tmp.hFs8UVwLPE ++ rm /tmp/tmp.z8gUDhEmpa /tmp/tmp.hFs8UVwLPE ++ return 0 + kubectl_bin logs -n psmdb-operator percona-server-mongodb-operator-dfd77fbfc-cdfb4 ++ mktemp + local LAST_OUT=/tmp/tmp.vDo6ko44lY ++ mktemp + local LAST_ERR=/tmp/tmp.M8GXdGV1uD + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl logs -n psmdb-operator percona-server-mongodb-operator-dfd77fbfc-cdfb4 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.vDo6ko44lY + cat /tmp/tmp.M8GXdGV1uD + rm /tmp/tmp.vDo6ko44lY /tmp/tmp.M8GXdGV1uD + return 0 2026-03-26T10:21:51.916Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "31d5df23-2957-40b7-aa98-c3b41ff83036", "sts": "some-name-rs2", "replset": "rs2"} 2026-03-26T10:21:51.916Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "31d5df23-2957-40b7-aa98-c3b41ff83036", "sts": "some-name-cfg", "replset": "cfg"} 2026-03-26T10:21:51.916Z DEBUG Check restore status {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "31d5df23-2957-40b7-aa98-c3b41ff83036", "command": ["/opt/percona/pbm", "describe-restore", "2026-03-26T09:55:42.160775713Z", "--config", "/etc/pbm/pbm_config.yaml", "--out", "json"], "pod": "some-name-rs0-0"} 2026-03-26T10:21:52.418Z DEBUG PBM restore status {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "31d5df23-2957-40b7-aa98-c3b41ff83036", "status": {"type":"physical","opid":"","name":"2026-03-26T09:55:42.160775713Z","replsets":[{"name":"rs2","start_ts":0,"status":"starting","size":0,"size_uncompressed":0,"last_transition_ts":1774518948,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null},{"name":"rs0","start_ts":0,"status":"","size":0,"size_uncompressed":0,"last_transition_ts":0,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null},{"name":"cfg","start_ts":0,"status":"starting","size":0,"size_uncompressed":0,"last_transition_ts":1774518948,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null},{"name":"rs1","start_ts":0,"status":"starting","size":0,"size_uncompressed":0,"last_transition_ts":1774518948,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null}],"compression":"","store":{"type":""},"size":0,"size_uncompressed":0,"mongodb_version":"","fcv":"","start_ts":0,"last_transition_ts":0,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"hb":{"T":0,"I":0},"status":"","conditions":null,"n":null,"pbm_version":"","balancer":""}} 2026-03-26T10:21:57.418Z DEBUG checking if restore is allowed {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "23f216cf-56e5-46a1-b2c5-ca99b1bd8b2e", "cluster": "some-name", "namespace": "pitr-physical-backup-source-28911"} 2026-03-26T10:21:57.420Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "23f216cf-56e5-46a1-b2c5-ca99b1bd8b2e", "pod": "some-name-rs0-0"} 2026-03-26T10:21:57.420Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "23f216cf-56e5-46a1-b2c5-ca99b1bd8b2e", "pod": "some-name-rs0-2"} 2026-03-26T10:21:57.420Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "23f216cf-56e5-46a1-b2c5-ca99b1bd8b2e", "pod": "some-name-rs0-1"} 2026-03-26T10:21:57.420Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "23f216cf-56e5-46a1-b2c5-ca99b1bd8b2e", "sts": "some-name-rs0", "replset": "rs0"} 2026-03-26T10:21:57.421Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "23f216cf-56e5-46a1-b2c5-ca99b1bd8b2e", "pod": "some-name-rs1-0"} 2026-03-26T10:21:57.421Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "23f216cf-56e5-46a1-b2c5-ca99b1bd8b2e", "pod": "some-name-rs1-1"} 2026-03-26T10:21:57.421Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "23f216cf-56e5-46a1-b2c5-ca99b1bd8b2e", "pod": "some-name-rs1-2"} 2026-03-26T10:21:57.421Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "23f216cf-56e5-46a1-b2c5-ca99b1bd8b2e", "sts": "some-name-rs1", "replset": "rs1"} 2026-03-26T10:21:57.421Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "23f216cf-56e5-46a1-b2c5-ca99b1bd8b2e", "pod": "some-name-rs2-2"} 2026-03-26T10:21:57.421Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "23f216cf-56e5-46a1-b2c5-ca99b1bd8b2e", "pod": "some-name-rs2-0"} 2026-03-26T10:21:57.421Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "23f216cf-56e5-46a1-b2c5-ca99b1bd8b2e", "pod": "some-name-rs2-1"} 2026-03-26T10:21:57.421Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "23f216cf-56e5-46a1-b2c5-ca99b1bd8b2e", "sts": "some-name-rs2", "replset": "rs2"} 2026-03-26T10:21:57.421Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "23f216cf-56e5-46a1-b2c5-ca99b1bd8b2e", "sts": "some-name-cfg", "replset": "cfg"} 2026-03-26T10:21:57.421Z DEBUG Check restore status {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "23f216cf-56e5-46a1-b2c5-ca99b1bd8b2e", "command": ["/opt/percona/pbm", "describe-restore", "2026-03-26T09:55:42.160775713Z", "--config", "/etc/pbm/pbm_config.yaml", "--out", "json"], "pod": "some-name-rs0-0"} 2026-03-26T10:21:58.158Z DEBUG PBM restore status {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "23f216cf-56e5-46a1-b2c5-ca99b1bd8b2e", "status": {"type":"physical","opid":"","name":"2026-03-26T09:55:42.160775713Z","replsets":[{"name":"cfg","start_ts":0,"status":"starting","size":0,"size_uncompressed":0,"last_transition_ts":1774518948,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null},{"name":"rs0","start_ts":0,"status":"","size":0,"size_uncompressed":0,"last_transition_ts":0,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null},{"name":"rs1","start_ts":0,"status":"starting","size":0,"size_uncompressed":0,"last_transition_ts":1774518948,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null},{"name":"rs2","start_ts":0,"status":"starting","size":0,"size_uncompressed":0,"last_transition_ts":1774518948,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null}],"compression":"","store":{"type":""},"size":0,"size_uncompressed":0,"mongodb_version":"","fcv":"","start_ts":0,"last_transition_ts":0,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"hb":{"T":0,"I":0},"status":"","conditions":null,"n":null,"pbm_version":"","balancer":""}} 2026-03-26T10:22:03.159Z DEBUG checking if restore is allowed {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "18e0b783-0b4b-4352-800a-71c0f3a6d086", "cluster": "some-name", "namespace": "pitr-physical-backup-source-28911"} 2026-03-26T10:22:03.161Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "18e0b783-0b4b-4352-800a-71c0f3a6d086", "pod": "some-name-rs0-0"} 2026-03-26T10:22:03.161Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "18e0b783-0b4b-4352-800a-71c0f3a6d086", "pod": "some-name-rs0-2"} 2026-03-26T10:22:03.161Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "18e0b783-0b4b-4352-800a-71c0f3a6d086", "pod": "some-name-rs0-1"} 2026-03-26T10:22:03.161Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "18e0b783-0b4b-4352-800a-71c0f3a6d086", "sts": "some-name-rs0", "replset": "rs0"} 2026-03-26T10:22:03.161Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "18e0b783-0b4b-4352-800a-71c0f3a6d086", "pod": "some-name-rs1-2"} 2026-03-26T10:22:03.161Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "18e0b783-0b4b-4352-800a-71c0f3a6d086", "pod": "some-name-rs1-0"} 2026-03-26T10:22:03.161Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "18e0b783-0b4b-4352-800a-71c0f3a6d086", "pod": "some-name-rs1-1"} 2026-03-26T10:22:03.161Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "18e0b783-0b4b-4352-800a-71c0f3a6d086", "sts": "some-name-rs1", "replset": "rs1"} 2026-03-26T10:22:03.161Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "18e0b783-0b4b-4352-800a-71c0f3a6d086", "pod": "some-name-rs2-2"} 2026-03-26T10:22:03.161Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "18e0b783-0b4b-4352-800a-71c0f3a6d086", "pod": "some-name-rs2-0"} 2026-03-26T10:22:03.161Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "18e0b783-0b4b-4352-800a-71c0f3a6d086", "pod": "some-name-rs2-1"} 2026-03-26T10:22:03.161Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "18e0b783-0b4b-4352-800a-71c0f3a6d086", "sts": "some-name-rs2", "replset": "rs2"} 2026-03-26T10:22:03.161Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "18e0b783-0b4b-4352-800a-71c0f3a6d086", "sts": "some-name-cfg", "replset": "cfg"} 2026-03-26T10:22:03.161Z DEBUG Check restore status {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "18e0b783-0b4b-4352-800a-71c0f3a6d086", "command": ["/opt/percona/pbm", "describe-restore", "2026-03-26T09:55:42.160775713Z", "--config", "/etc/pbm/pbm_config.yaml", "--out", "json"], "pod": "some-name-rs0-0"} 2026-03-26T10:22:03.945Z DEBUG PBM restore status {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "18e0b783-0b4b-4352-800a-71c0f3a6d086", "status": {"type":"physical","opid":"","name":"2026-03-26T09:55:42.160775713Z","replsets":[{"name":"rs0","start_ts":0,"status":"","size":0,"size_uncompressed":0,"last_transition_ts":0,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null},{"name":"rs1","start_ts":0,"status":"starting","size":0,"size_uncompressed":0,"last_transition_ts":1774518948,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null},{"name":"rs2","start_ts":0,"status":"starting","size":0,"size_uncompressed":0,"last_transition_ts":1774518948,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null},{"name":"cfg","start_ts":0,"status":"starting","size":0,"size_uncompressed":0,"last_transition_ts":1774518948,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null}],"compression":"","store":{"type":""},"size":0,"size_uncompressed":0,"mongodb_version":"","fcv":"","start_ts":0,"last_transition_ts":0,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"hb":{"T":0,"I":0},"status":"","conditions":null,"n":null,"pbm_version":"","balancer":""}} 2026-03-26T10:22:08.947Z DEBUG checking if restore is allowed {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "bc5017eb-4e40-47bc-8290-35d7d567936a", "cluster": "some-name", "namespace": "pitr-physical-backup-source-28911"} 2026-03-26T10:22:08.949Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "bc5017eb-4e40-47bc-8290-35d7d567936a", "pod": "some-name-rs0-1"} 2026-03-26T10:22:08.949Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "bc5017eb-4e40-47bc-8290-35d7d567936a", "pod": "some-name-rs0-0"} 2026-03-26T10:22:08.949Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "bc5017eb-4e40-47bc-8290-35d7d567936a", "pod": "some-name-rs0-2"} 2026-03-26T10:22:08.949Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "bc5017eb-4e40-47bc-8290-35d7d567936a", "sts": "some-name-rs0", "replset": "rs0"} 2026-03-26T10:22:08.949Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "bc5017eb-4e40-47bc-8290-35d7d567936a", "pod": "some-name-rs1-2"} 2026-03-26T10:22:08.949Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "bc5017eb-4e40-47bc-8290-35d7d567936a", "pod": "some-name-rs1-0"} 2026-03-26T10:22:08.949Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "bc5017eb-4e40-47bc-8290-35d7d567936a", "pod": "some-name-rs1-1"} 2026-03-26T10:22:08.949Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "bc5017eb-4e40-47bc-8290-35d7d567936a", "sts": "some-name-rs1", "replset": "rs1"} 2026-03-26T10:22:08.950Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "bc5017eb-4e40-47bc-8290-35d7d567936a", "pod": "some-name-rs2-2"} 2026-03-26T10:22:08.950Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "bc5017eb-4e40-47bc-8290-35d7d567936a", "pod": "some-name-rs2-0"} 2026-03-26T10:22:08.950Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "bc5017eb-4e40-47bc-8290-35d7d567936a", "pod": "some-name-rs2-1"} 2026-03-26T10:22:08.950Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "bc5017eb-4e40-47bc-8290-35d7d567936a", "sts": "some-name-rs2", "replset": "rs2"} 2026-03-26T10:22:08.950Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "bc5017eb-4e40-47bc-8290-35d7d567936a", "sts": "some-name-cfg", "replset": "cfg"} 2026-03-26T10:22:08.950Z DEBUG Check restore status {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "bc5017eb-4e40-47bc-8290-35d7d567936a", "command": ["/opt/percona/pbm", "describe-restore", "2026-03-26T09:55:42.160775713Z", "--config", "/etc/pbm/pbm_config.yaml", "--out", "json"], "pod": "some-name-rs0-0"} 2026-03-26T10:22:09.450Z DEBUG PBM restore status {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "bc5017eb-4e40-47bc-8290-35d7d567936a", "status": {"type":"physical","opid":"","name":"2026-03-26T09:55:42.160775713Z","replsets":[{"name":"cfg","start_ts":0,"status":"starting","size":0,"size_uncompressed":0,"last_transition_ts":1774518948,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null},{"name":"rs0","start_ts":0,"status":"","size":0,"size_uncompressed":0,"last_transition_ts":0,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null},{"name":"rs1","start_ts":0,"status":"starting","size":0,"size_uncompressed":0,"last_transition_ts":1774518948,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null},{"name":"rs2","start_ts":0,"status":"starting","size":0,"size_uncompressed":0,"last_transition_ts":1774518948,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null}],"compression":"","store":{"type":""},"size":0,"size_uncompressed":0,"mongodb_version":"","fcv":"","start_ts":0,"last_transition_ts":0,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"hb":{"T":0,"I":0},"status":"","conditions":null,"n":null,"pbm_version":"","balancer":""}} 2026-03-26T10:22:14.451Z DEBUG checking if restore is allowed {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8b0c978-f2a0-40a8-b43c-09372371aaef", "cluster": "some-name", "namespace": "pitr-physical-backup-source-28911"} 2026-03-26T10:22:14.453Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8b0c978-f2a0-40a8-b43c-09372371aaef", "pod": "some-name-rs0-1"} 2026-03-26T10:22:14.453Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8b0c978-f2a0-40a8-b43c-09372371aaef", "pod": "some-name-rs0-0"} 2026-03-26T10:22:14.453Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8b0c978-f2a0-40a8-b43c-09372371aaef", "pod": "some-name-rs0-2"} 2026-03-26T10:22:14.453Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8b0c978-f2a0-40a8-b43c-09372371aaef", "sts": "some-name-rs0", "replset": "rs0"} 2026-03-26T10:22:14.454Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8b0c978-f2a0-40a8-b43c-09372371aaef", "pod": "some-name-rs1-0"} 2026-03-26T10:22:14.454Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8b0c978-f2a0-40a8-b43c-09372371aaef", "pod": "some-name-rs1-1"} 2026-03-26T10:22:14.454Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8b0c978-f2a0-40a8-b43c-09372371aaef", "pod": "some-name-rs1-2"} 2026-03-26T10:22:14.454Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8b0c978-f2a0-40a8-b43c-09372371aaef", "sts": "some-name-rs1", "replset": "rs1"} 2026-03-26T10:22:14.454Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8b0c978-f2a0-40a8-b43c-09372371aaef", "pod": "some-name-rs2-2"} 2026-03-26T10:22:14.454Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8b0c978-f2a0-40a8-b43c-09372371aaef", "pod": "some-name-rs2-0"} 2026-03-26T10:22:14.454Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8b0c978-f2a0-40a8-b43c-09372371aaef", "pod": "some-name-rs2-1"} 2026-03-26T10:22:14.454Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8b0c978-f2a0-40a8-b43c-09372371aaef", "sts": "some-name-rs2", "replset": "rs2"} 2026-03-26T10:22:14.454Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8b0c978-f2a0-40a8-b43c-09372371aaef", "sts": "some-name-cfg", "replset": "cfg"} 2026-03-26T10:22:14.454Z DEBUG Check restore status {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8b0c978-f2a0-40a8-b43c-09372371aaef", "command": ["/opt/percona/pbm", "describe-restore", "2026-03-26T09:55:42.160775713Z", "--config", "/etc/pbm/pbm_config.yaml", "--out", "json"], "pod": "some-name-rs0-0"} 2026-03-26T10:22:14.943Z DEBUG PBM restore status {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8b0c978-f2a0-40a8-b43c-09372371aaef", "status": {"type":"physical","opid":"","name":"2026-03-26T09:55:42.160775713Z","replsets":[{"name":"rs2","start_ts":0,"status":"starting","size":0,"size_uncompressed":0,"last_transition_ts":1774518948,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null},{"name":"cfg","start_ts":0,"status":"starting","size":0,"size_uncompressed":0,"last_transition_ts":1774518948,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null},{"name":"rs0","start_ts":0,"status":"","size":0,"size_uncompressed":0,"last_transition_ts":0,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null},{"name":"rs1","start_ts":0,"status":"starting","size":0,"size_uncompressed":0,"last_transition_ts":1774518948,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null}],"compression":"","store":{"type":""},"size":0,"size_uncompressed":0,"mongodb_version":"","fcv":"","start_ts":0,"last_transition_ts":0,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"hb":{"T":0,"I":0},"status":"","conditions":null,"n":null,"pbm_version":"","balancer":""}} 2026-03-26T10:22:19.945Z DEBUG checking if restore is allowed {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "ef3787a9-8f86-4b14-94bb-bd50d9f5fda2", "cluster": "some-name", "namespace": "pitr-physical-backup-source-28911"} 2026-03-26T10:22:19.946Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "ef3787a9-8f86-4b14-94bb-bd50d9f5fda2", "pod": "some-name-rs0-2"} 2026-03-26T10:22:19.946Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "ef3787a9-8f86-4b14-94bb-bd50d9f5fda2", "pod": "some-name-rs0-1"} 2026-03-26T10:22:19.946Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "ef3787a9-8f86-4b14-94bb-bd50d9f5fda2", "pod": "some-name-rs0-0"} 2026-03-26T10:22:19.946Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "ef3787a9-8f86-4b14-94bb-bd50d9f5fda2", "sts": "some-name-rs0", "replset": "rs0"} 2026-03-26T10:22:19.946Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "ef3787a9-8f86-4b14-94bb-bd50d9f5fda2", "pod": "some-name-rs1-0"} 2026-03-26T10:22:19.947Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "ef3787a9-8f86-4b14-94bb-bd50d9f5fda2", "pod": "some-name-rs1-1"} 2026-03-26T10:22:19.947Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "ef3787a9-8f86-4b14-94bb-bd50d9f5fda2", "pod": "some-name-rs1-2"} 2026-03-26T10:22:19.947Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "ef3787a9-8f86-4b14-94bb-bd50d9f5fda2", "sts": "some-name-rs1", "replset": "rs1"} 2026-03-26T10:22:19.947Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "ef3787a9-8f86-4b14-94bb-bd50d9f5fda2", "pod": "some-name-rs2-2"} 2026-03-26T10:22:19.947Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "ef3787a9-8f86-4b14-94bb-bd50d9f5fda2", "pod": "some-name-rs2-0"} 2026-03-26T10:22:19.947Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "ef3787a9-8f86-4b14-94bb-bd50d9f5fda2", "pod": "some-name-rs2-1"} 2026-03-26T10:22:19.947Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "ef3787a9-8f86-4b14-94bb-bd50d9f5fda2", "sts": "some-name-rs2", "replset": "rs2"} 2026-03-26T10:22:19.947Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "ef3787a9-8f86-4b14-94bb-bd50d9f5fda2", "sts": "some-name-cfg", "replset": "cfg"} 2026-03-26T10:22:19.947Z DEBUG Check restore status {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "ef3787a9-8f86-4b14-94bb-bd50d9f5fda2", "command": ["/opt/percona/pbm", "describe-restore", "2026-03-26T09:55:42.160775713Z", "--config", "/etc/pbm/pbm_config.yaml", "--out", "json"], "pod": "some-name-rs0-0"} 2026-03-26T10:22:20.434Z DEBUG PBM restore status {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "ef3787a9-8f86-4b14-94bb-bd50d9f5fda2", "status": {"type":"physical","opid":"","name":"2026-03-26T09:55:42.160775713Z","replsets":[{"name":"rs1","start_ts":0,"status":"starting","size":0,"size_uncompressed":0,"last_transition_ts":1774518948,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null},{"name":"rs0","start_ts":0,"status":"","size":0,"size_uncompressed":0,"last_transition_ts":0,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null},{"name":"cfg","start_ts":0,"status":"starting","size":0,"size_uncompressed":0,"last_transition_ts":1774518948,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null},{"name":"rs2","start_ts":0,"status":"starting","size":0,"size_uncompressed":0,"last_transition_ts":1774518948,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null}],"compression":"","store":{"type":""},"size":0,"size_uncompressed":0,"mongodb_version":"","fcv":"","start_ts":0,"last_transition_ts":0,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"hb":{"T":0,"I":0},"status":"","conditions":null,"n":null,"pbm_version":"","balancer":""}} 2026-03-26T10:22:25.435Z DEBUG checking if restore is allowed {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8a916a7-4f13-45dc-8673-fabd331ab0db", "cluster": "some-name", "namespace": "pitr-physical-backup-source-28911"} 2026-03-26T10:22:25.437Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8a916a7-4f13-45dc-8673-fabd331ab0db", "pod": "some-name-rs0-0"} 2026-03-26T10:22:25.437Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8a916a7-4f13-45dc-8673-fabd331ab0db", "pod": "some-name-rs0-2"} 2026-03-26T10:22:25.437Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8a916a7-4f13-45dc-8673-fabd331ab0db", "pod": "some-name-rs0-1"} 2026-03-26T10:22:25.437Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8a916a7-4f13-45dc-8673-fabd331ab0db", "sts": "some-name-rs0", "replset": "rs0"} 2026-03-26T10:22:25.437Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8a916a7-4f13-45dc-8673-fabd331ab0db", "pod": "some-name-rs1-0"} 2026-03-26T10:22:25.437Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8a916a7-4f13-45dc-8673-fabd331ab0db", "pod": "some-name-rs1-1"} 2026-03-26T10:22:25.437Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8a916a7-4f13-45dc-8673-fabd331ab0db", "pod": "some-name-rs1-2"} 2026-03-26T10:22:25.437Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8a916a7-4f13-45dc-8673-fabd331ab0db", "sts": "some-name-rs1", "replset": "rs1"} 2026-03-26T10:22:25.437Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8a916a7-4f13-45dc-8673-fabd331ab0db", "pod": "some-name-rs2-2"} 2026-03-26T10:22:25.437Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8a916a7-4f13-45dc-8673-fabd331ab0db", "pod": "some-name-rs2-0"} 2026-03-26T10:22:25.437Z DEBUG Pod is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8a916a7-4f13-45dc-8673-fabd331ab0db", "pod": "some-name-rs2-1"} 2026-03-26T10:22:25.437Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8a916a7-4f13-45dc-8673-fabd331ab0db", "sts": "some-name-rs2", "replset": "rs2"} 2026-03-26T10:22:25.437Z DEBUG Statefulset is ready for physical restore {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8a916a7-4f13-45dc-8673-fabd331ab0db", "sts": "some-name-cfg", "replset": "cfg"} 2026-03-26T10:22:25.437Z DEBUG Check restore status {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8a916a7-4f13-45dc-8673-fabd331ab0db", "command": ["/opt/percona/pbm", "describe-restore", "2026-03-26T09:55:42.160775713Z", "--config", "/etc/pbm/pbm_config.yaml", "--out", "json"], "pod": "some-name-rs0-0"} 2026-03-26T10:22:26.225Z DEBUG PBM restore status {"controller": "psmdbrestore-controller", "controllerGroup": "psmdb.percona.com", "controllerKind": "PerconaServerMongoDBRestore", "PerconaServerMongoDBRestore": {"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"}, "namespace": "pitr-physical-backup-source-28911", "name": "restore-backup-minio-1", "reconcileID": "f8a916a7-4f13-45dc-8673-fabd331ab0db", "status": {"type":"physical","opid":"","name":"2026-03-26T09:55:42.160775713Z","replsets":[{"name":"rs0","start_ts":0,"status":"","size":0,"size_uncompressed":0,"last_transition_ts":0,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null},{"name":"rs1","start_ts":0,"status":"starting","size":0,"size_uncompressed":0,"last_transition_ts":1774518948,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null},{"name":"rs2","start_ts":0,"status":"starting","size":0,"size_uncompressed":0,"last_transition_ts":1774518948,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null},{"name":"cfg","start_ts":0,"status":"starting","size":0,"size_uncompressed":0,"last_transition_ts":1774518948,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"node":"","conditions":null}],"compression":"","store":{"type":""},"size":0,"size_uncompressed":0,"mongodb_version":"","fcv":"","start_ts":0,"last_transition_ts":0,"first_write_ts":{"T":0,"I":0},"last_write_ts":{"T":0,"I":0},"hb":{"T":0,"I":0},"status":"","conditions":null,"n":null,"pbm_version":"","balancer":""}} + kubectl_bin get psmdb-restore restore-backup-minio-1 -o yaml ++ mktemp + local LAST_OUT=/tmp/tmp.Jc2N0GE74K ++ mktemp + local LAST_ERR=/tmp/tmp.xlb1fYv4X7 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get psmdb-restore restore-backup-minio-1 -o yaml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Jc2N0GE74K apiVersion: psmdb.percona.com/v1 kind: PerconaServerMongoDBRestore metadata: annotations: kubectl.kubernetes.io/last-applied-configuration: | {"apiVersion":"psmdb.percona.com/v1","kind":"PerconaServerMongoDBRestore","metadata":{"annotations":{},"name":"restore-backup-minio-1","namespace":"pitr-physical-backup-source-28911"},"spec":{"backupSource":{"destination":"s3://operator-testing/2026-03-26T09:42:00Z","s3":{"bucket":"operator-testing","credentialsSecret":"minio-secret","endpointUrl":"http://minio-service:9000/","insecureSkipTLSVerify":false,"region":"us-east-1"},"type":"physical"},"clusterName":"some-name","pitr":{"date":"2026-03-26 09:45:26","type":"date"}}} creationTimestamp: "2026-03-26T09:47:35Z" generation: 1 name: restore-backup-minio-1 namespace: pitr-physical-backup-source-28911 resourceVersion: "1774518942200047001" uid: 6c0b3c82-f3fe-4fc3-85ab-972e68635289 spec: backupSource: destination: s3://operator-testing/2026-03-26T09:42:00Z s3: bucket: operator-testing credentialsSecret: minio-secret endpointUrl: http://minio-service:9000/ insecureSkipTLSVerify: false region: us-east-1 type: physical clusterName: some-name pitr: date: "2026-03-26 09:45:26" type: date status: pbmName: "2026-03-26T09:55:42.160775713Z" pitrTarget: 2026-03-26T09:45:26 state: requested + cat /tmp/tmp.xlb1fYv4X7 + rm /tmp/tmp.Jc2N0GE74K /tmp/tmp.xlb1fYv4X7 + return 0 + kubectl_bin describe psmdb-restore restore-backup-minio-1 ++ mktemp + local LAST_OUT=/tmp/tmp.QQnO5mzL77 ++ mktemp + local LAST_ERR=/tmp/tmp.5NZ8ZBQJhn + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl describe psmdb-restore restore-backup-minio-1 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.QQnO5mzL77 Name: restore-backup-minio-1 Namespace: pitr-physical-backup-source-28911 Labels: Annotations: API Version: psmdb.percona.com/v1 Kind: PerconaServerMongoDBRestore Metadata: Creation Timestamp: 2026-03-26T09:47:35Z Generation: 1 Resource Version: 1774518942200047001 UID: 6c0b3c82-f3fe-4fc3-85ab-972e68635289 Spec: Backup Source: Destination: s3://operator-testing/2026-03-26T09:42:00Z s3: Bucket: operator-testing Credentials Secret: minio-secret Endpoint URL: http://minio-service:9000/ Insecure Skip Tls Verify: false Region: us-east-1 Type: physical Cluster Name: some-name Pitr: Date: 2026-03-26 09:45:26 Type: date Status: Pbm Name: 2026-03-26T09:55:42.160775713Z Pitr Target: 2026-03-26T09:45:26 State: requested Events: + cat /tmp/tmp.5NZ8ZBQJhn + rm /tmp/tmp.QQnO5mzL77 /tmp/tmp.5NZ8ZBQJhn + return 0 + echo 'Restore object restore-backup-minio-1 is in requested state after 26 minutes.' Restore object restore-backup-minio-1 is in requested state after 26 minutes. + echo something went wrong with operator or kubernetes cluster something went wrong with operator or kubernetes cluster + exit 1