Log: /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/e2e-tests/logs/demand-backup-if-unhealthy.log grep: warning: stray \ before - Warning: version difference between client (1.35) and server (1.31) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.35) and server (1.31) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.35) and server (1.31) exceeds the supported minor version skew of +/-1 + create_infra demand-backup-if-unhealthy-13462 + local ns=demand-backup-if-unhealthy-13462 + [[ 1 == 1 ]] + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.M2OVoYfu8J ++ mktemp + local LAST_ERR=/tmp/tmp.Rw8I53wiI0 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.M2OVoYfu8J customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.Rw8I53wiI0 + rm /tmp/tmp.M2OVoYfu8J /tmp/tmp.Rw8I53wiI0 + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/deploy/crd.yaml ++ grep -v '\-\-\-' grep: warning: stray \ before - grep: warning: stray \ before - + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' No resources found + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: resource(s) were provided, but no name was specified + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.9qJLTeEn2V ++ mktemp + local LAST_ERR=/tmp/tmp.nfDovNz20t + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.9qJLTeEn2V + cat /tmp/tmp.nfDovNz20t + rm /tmp/tmp.9qJLTeEn2V /tmp/tmp.nfDovNz20t + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.GeZfrFqjZM ++ mktemp + local LAST_ERR=/tmp/tmp.RzHzBbh0tL + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.GeZfrFqjZM + cat /tmp/tmp.RzHzBbh0tL + rm /tmp/tmp.GeZfrFqjZM /tmp/tmp.RzHzBbh0tL + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbs" + kubectl patch perconaservermongodbs.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbs" + : + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.jHvvVOUuDc ++ mktemp + local LAST_ERR=/tmp/tmp.kFBO83uwqz + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.jHvvVOUuDc + cat /tmp/tmp.kFBO83uwqz + rm /tmp/tmp.jHvvVOUuDc /tmp/tmp.kFBO83uwqz + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.njMmONc4yA ++ mktemp + local LAST_ERR=/tmp/tmp.11cLkuEqcQ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.njMmONc4yA clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.11cLkuEqcQ + rm /tmp/tmp.njMmONc4yA /tmp/tmp.11cLkuEqcQ + return 0 + check_crd_for_deletion PR-2159-33bd82f1 + local git_tag=PR-2159-33bd82f1 ++ curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/PR-2159-33bd82f1/deploy/crd.yaml ++ yq eval .metadata.name ++ /usr/sbin/sed s/---//g ++ /usr/sbin/sed ':a;N;$!ba;s/\n/ /g' + for crd_name in $(curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/${git_tag}/deploy/crd.yaml | yq eval '.metadata.name' | $sed 's/---//g' | $sed ':a;N;$!ba;s/\n/ /g') ++ kubectl_bin get crd/null -o 'jsonpath={.status.conditions[-1].type}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pMFQG0hBxH +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZkPqLFe3CR ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.pMFQG0hBxH ++ cat /tmp/tmp.ZkPqLFe3CR Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 0 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.pMFQG0hBxH ++ cat /tmp/tmp.ZkPqLFe3CR Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 4 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.pMFQG0hBxH ++ cat /tmp/tmp.ZkPqLFe3CR Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 8 ++ cat /tmp/tmp.pMFQG0hBxH ++ cat /tmp/tmp.ZkPqLFe3CR Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ rm /tmp/tmp.pMFQG0hBxH /tmp/tmp.ZkPqLFe3CR ++ return 1 + [[ '' == \T\e\r\m\i\n\a\t\i\n\g ]] + '[' -n psmdb-operator ']' + create_namespace psmdb-operator + local namespace=psmdb-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep validate-auth ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + egrep -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' + awk '{print$1}' + '[' -n '' ']' + desc 'cleaned up old namespaces psmdb-operator' ++ mktemp + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace psmdb-operator --ignore-not-found + xargs kubectl delete ns ++ mktemp egrep: warning: egrep is obsolescent; using grep -E + local LAST_OUT=/tmp/tmp.O0zStOcvfL + local LAST_OUT=/tmp/tmp.MDQsJoZoQj ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.irXaqThwBh + local exit_status=0 + local timeout=4 + local LAST_ERR=/tmp/tmp.6AEBZbScee + local exit_status=0 + local timeout=4 ++ seq 0 2 ++ seq 0 2 + for i in $(seq 0 2) + for i in $(seq 0 2) + set +e + set +e + kubectl get ns + kubectl delete namespace psmdb-operator --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.O0zStOcvfL + cat /tmp/tmp.6AEBZbScee + rm /tmp/tmp.O0zStOcvfL /tmp/tmp.6AEBZbScee + return 0 namespace "demand-backup-if-unhealthy-2259" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.MDQsJoZoQj namespace "psmdb-operator" deleted + cat /tmp/tmp.irXaqThwBh + rm /tmp/tmp.MDQsJoZoQj /tmp/tmp.irXaqThwBh + return 0 + kubectl_bin wait --for=delete namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.XyRcTdo01E ++ mktemp + local LAST_ERR=/tmp/tmp.oTA1zw6cpQ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.XyRcTdo01E + cat /tmp/tmp.oTA1zw6cpQ + rm /tmp/tmp.XyRcTdo01E /tmp/tmp.oTA1zw6cpQ + return 0 + desc 'create namespace psmdb-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.dFyyLHwMPm ++ mktemp + local LAST_ERR=/tmp/tmp.z9vjHvydM2 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.dFyyLHwMPm namespace/psmdb-operator created + cat /tmp/tmp.z9vjHvydM2 + rm /tmp/tmp.dFyyLHwMPm /tmp/tmp.z9vjHvydM2 + return 0 + set_kube_ctx psmdb-operator + local namespace=psmdb-operator ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.hD467RKIuG +++ mktemp ++ local LAST_ERR=/tmp/tmp.coRQYc2GOn ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.hD467RKIuG ++ cat /tmp/tmp.coRQYc2GOn ++ rm /tmp/tmp.hD467RKIuG /tmp/tmp.coRQYc2GOn ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2159-33bd82f1-1-cluster4 --namespace=psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.2qlHgEn3KD ++ mktemp + local LAST_ERR=/tmp/tmp.XTCbHAbF9W + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2159-33bd82f1-1-cluster4 --namespace=psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.2qlHgEn3KD Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2159-33bd82f1-1-cluster4" modified. + cat /tmp/tmp.XTCbHAbF9W + rm /tmp/tmp.2qlHgEn3KD /tmp/tmp.XTCbHAbF9W + return 0 + deploy_operator + desc 'start PSMDB operator: perconalab/percona-server-mongodb-operator:PR-2159-33bd82f1' + set +o xtrace ----------------------------------------------------------------------------------- start PSMDB operator: perconalab/percona-server-mongodb-operator:PR-2159-33bd82f1 ----------------------------------------------------------------------------------- + local cr_file + '[' -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/e2e-tests/demand-backup-if-unhealthy/conf/crd.yaml ']' + cr_file=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/deploy/crd.yaml + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.rYkTqaBEKZ ++ mktemp + local LAST_ERR=/tmp/tmp.tZKw8Ozjq7 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.rYkTqaBEKZ customresourcedefinition.apiextensions.k8s.io/perconaservermongodbbackups.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbrestores.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com serverside-applied + cat /tmp/tmp.tZKw8Ozjq7 + rm /tmp/tmp.rYkTqaBEKZ /tmp/tmp.tZKw8Ozjq7 + return 0 + '[' -n psmdb-operator ']' + apply_rbac cw-rbac + local operator_namespace=psmdb-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: psmdb-operator^' + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.sWXpOk692q ++ mktemp + local LAST_ERR=/tmp/tmp.pVkzyDwiu4 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.sWXpOk692q clusterrole.rbac.authorization.k8s.io/percona-server-mongodb-operator created serviceaccount/percona-server-mongodb-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-server-mongodb-operator created + cat /tmp/tmp.pVkzyDwiu4 + rm /tmp/tmp.sWXpOk692q /tmp/tmp.pVkzyDwiu4 + return 0 + yq eval ' (.spec.template.spec.containers[].image = "perconalab/percona-server-mongodb-operator:PR-2159-33bd82f1") | ((.. | select(.[] == "DISABLE_TELEMETRY")) |= .value="true") | ((.. | select(.[] == "LOG_LEVEL")) |= .value="DEBUG")' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/deploy/cw-operator.yaml + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.ffGxmAdh1y ++ mktemp + local LAST_ERR=/tmp/tmp.PjLi6ZVa0K + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ffGxmAdh1y deployment.apps/percona-server-mongodb-operator created + cat /tmp/tmp.PjLi6ZVa0K + rm /tmp/tmp.ffGxmAdh1y /tmp/tmp.PjLi6ZVa0K + return 0 + sleep 20 ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.X4igjRW2wB +++ mktemp ++ local LAST_ERR=/tmp/tmp.e2UHwT4Vi4 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.X4igjRW2wB ++ cat /tmp/tmp.e2UHwT4Vi4 ++ rm /tmp/tmp.X4igjRW2wB /tmp/tmp.e2UHwT4Vi4 ++ return 0 + wait_operator_pod percona-server-mongodb-operator-6fd5955c87-wcvbf + local pod=percona-server-mongodb-operator-6fd5955c87-wcvbf + set +o xtrace waiting for pod/percona-server-mongodb-operator-6fd5955c87-wcvbf to be ready.OK + echo 'Print operator info from log' Print operator info from log + grep 'Manager starting up' ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.ATBKrkCLJX +++ mktemp ++ local LAST_ERR=/tmp/tmp.swJZkSMwd9 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ATBKrkCLJX ++ cat /tmp/tmp.swJZkSMwd9 ++ rm /tmp/tmp.ATBKrkCLJX /tmp/tmp.swJZkSMwd9 ++ return 0 + kubectl_bin logs -n psmdb-operator percona-server-mongodb-operator-6fd5955c87-wcvbf ++ mktemp + local LAST_OUT=/tmp/tmp.2Uzn6aSBY0 ++ mktemp + local LAST_ERR=/tmp/tmp.n7xjNWvH4l + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl logs -n psmdb-operator percona-server-mongodb-operator-6fd5955c87-wcvbf + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.2Uzn6aSBY0 + cat /tmp/tmp.n7xjNWvH4l + rm /tmp/tmp.2Uzn6aSBY0 /tmp/tmp.n7xjNWvH4l + return 0 2025-12-22T07:42:02.518Z INFO setup Manager starting up {"gitCommit": "33bd82f11d1f1477ae05ded547e1741b683ae31e", "gitBranch": "PR-2159-33bd82f1", "buildTime": "", "goVersion": "go1.25.5", "os": "linux", "arch": "amd64"} + create_namespace demand-backup-if-unhealthy-13462 + local namespace=demand-backup-if-unhealthy-13462 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + egrep -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' + '[' -n '' ']' + desc 'cleaned up old namespaces demand-backup-if-unhealthy-13462' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces demand-backup-if-unhealthy-13462 ----------------------------------------------------------------------------------- + awk '{print$1}' + kubectl_bin delete namespace demand-backup-if-unhealthy-13462 --ignore-not-found + xargs kubectl delete ns ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.ddNbBSDleC ++ mktemp + local LAST_OUT=/tmp/tmp.vNxucY1anx egrep: warning: egrep is obsolescent; using grep -E + local LAST_ERR=/tmp/tmp.qa1xzLGMlf + local exit_status=0 + local timeout=4 ++ mktemp ++ seq 0 2 + local LAST_ERR=/tmp/tmp.REMoWnGnOB + local exit_status=0 + local timeout=4 + for i in $(seq 0 2) + set +e + kubectl get ns ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete namespace demand-backup-if-unhealthy-13462 --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ddNbBSDleC + cat /tmp/tmp.qa1xzLGMlf + rm /tmp/tmp.ddNbBSDleC /tmp/tmp.qa1xzLGMlf + return 0 error: resource(s) were provided, but no name was specified + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.vNxucY1anx + cat /tmp/tmp.REMoWnGnOB + rm /tmp/tmp.vNxucY1anx /tmp/tmp.REMoWnGnOB + return 0 + kubectl_bin wait --for=delete namespace demand-backup-if-unhealthy-13462 ++ mktemp + local LAST_OUT=/tmp/tmp.j5h7m2dIS7 ++ mktemp + local LAST_ERR=/tmp/tmp.6Fw2QaIMSx + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete namespace demand-backup-if-unhealthy-13462 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.j5h7m2dIS7 + cat /tmp/tmp.6Fw2QaIMSx + rm /tmp/tmp.j5h7m2dIS7 /tmp/tmp.6Fw2QaIMSx + return 0 + desc 'create namespace demand-backup-if-unhealthy-13462' + set +o xtrace ----------------------------------------------------------------------------------- create namespace demand-backup-if-unhealthy-13462 ----------------------------------------------------------------------------------- + kubectl_bin create namespace demand-backup-if-unhealthy-13462 ++ mktemp + local LAST_OUT=/tmp/tmp.d1OFYVerH9 ++ mktemp + local LAST_ERR=/tmp/tmp.evRG5lOCnP + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace demand-backup-if-unhealthy-13462 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.d1OFYVerH9 namespace/demand-backup-if-unhealthy-13462 created + cat /tmp/tmp.evRG5lOCnP + rm /tmp/tmp.d1OFYVerH9 /tmp/tmp.evRG5lOCnP + return 0 + set_kube_ctx demand-backup-if-unhealthy-13462 + local namespace=demand-backup-if-unhealthy-13462 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.4K99BUQwAg +++ mktemp ++ local LAST_ERR=/tmp/tmp.jt3XQHZAM7 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4K99BUQwAg ++ cat /tmp/tmp.jt3XQHZAM7 ++ rm /tmp/tmp.4K99BUQwAg /tmp/tmp.jt3XQHZAM7 ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2159-33bd82f1-1-cluster4 --namespace=demand-backup-if-unhealthy-13462 ++ mktemp + local LAST_OUT=/tmp/tmp.gxfHhJnwG4 ++ mktemp + local LAST_ERR=/tmp/tmp.VoQYgZOzNs + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2159-33bd82f1-1-cluster4 --namespace=demand-backup-if-unhealthy-13462 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.gxfHhJnwG4 Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2159-33bd82f1-1-cluster4" modified. + cat /tmp/tmp.VoQYgZOzNs + rm /tmp/tmp.gxfHhJnwG4 /tmp/tmp.VoQYgZOzNs + return 0 + deploy_minio + desc 'install Minio' + set +o xtrace ----------------------------------------------------------------------------------- install Minio ----------------------------------------------------------------------------------- + helm uninstall minio-service Error: uninstall: Release not loaded: minio-service: release: not found + : + helm repo remove minio "minio" has been removed from your repositories + helm repo add minio https://charts.min.io/ "minio" has been added to your repositories + retry 10 60 helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/.minio/ --set persistence.size=2G --set securityContext.enabled=false minio/minio + local max=10 + local delay=60 + shift 2 + local n=1 + helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/.minio/ --set persistence.size=2G --set securityContext.enabled=false minio/minio NAME: minio-service LAST DEPLOYED: Mon Dec 22 07:42:40 2025 NAMESPACE: demand-backup-if-unhealthy-13462 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: MinIO can be accessed via port 9000 on the following DNS name from within your cluster: minio-service.demand-backup-if-unhealthy-13462.cluster.local To access MinIO from localhost, run the below commands: 1. export POD_NAME=$(kubectl get pods --namespace demand-backup-if-unhealthy-13462 -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}") 2. kubectl port-forward $POD_NAME 9000 --namespace demand-backup-if-unhealthy-13462 Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/ You can now access MinIO server on http://localhost:9000. Follow the below steps to connect to MinIO server with mc client: 1. Download the MinIO mc client - https://min.io/docs/minio/linux/reference/minio-mc.html#quickstart 2. export MC_HOST_minio-service-local=http://$(kubectl get secret --namespace demand-backup-if-unhealthy-13462 minio-service -o jsonpath="{.data.rootUser}" | base64 --decode):$(kubectl get secret --namespace demand-backup-if-unhealthy-13462 minio-service -o jsonpath="{.data.rootPassword}" | base64 --decode)@localhost:9000 3. mc ls minio-service-local ++ kubectl_bin get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9VVk1NzaBq +++ mktemp ++ local LAST_ERR=/tmp/tmp.u3BQ2Dtn3L ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9VVk1NzaBq ++ cat /tmp/tmp.u3BQ2Dtn3L ++ rm /tmp/tmp.9VVk1NzaBq /tmp/tmp.u3BQ2Dtn3L ++ return 0 + MINIO_POD=minio-service-d9589b474-rdxm9 + wait_pod minio-service-d9589b474-rdxm9 + local pod=minio-service-d9589b474-rdxm9 + set +o xtrace waiting for pod/minio-service-d9589b474-rdxm9 to be ready.OK + '[' -n psmdb-operator ']' + kubectl_bin create svc -n psmdb-operator externalname minio-service --external-name=minio-service.demand-backup-if-unhealthy-13462.svc.cluster.local --tcp=9000 ++ mktemp + local LAST_OUT=/tmp/tmp.XCIznP66RN ++ mktemp + local LAST_ERR=/tmp/tmp.fDrlSt4ItB + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create svc -n psmdb-operator externalname minio-service --external-name=minio-service.demand-backup-if-unhealthy-13462.svc.cluster.local --tcp=9000 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.XCIznP66RN service/minio-service created + cat /tmp/tmp.fDrlSt4ItB + rm /tmp/tmp.XCIznP66RN /tmp/tmp.fDrlSt4ItB + return 0 + create_minio_bucket operator-testing + local bucket=operator-testing + kubectl_bin run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- bash -c 'AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url http://minio-service:9000 s3 mb s3://operator-testing' ++ mktemp + local LAST_OUT=/tmp/tmp.I8Ar17M1aV ++ mktemp + local LAST_ERR=/tmp/tmp.rjTplA2ftP + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- bash -c 'AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url http://minio-service:9000 s3 mb s3://operator-testing' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.I8Ar17M1aV make_bucket: operator-testing pod "aws-cli" deleted from demand-backup-if-unhealthy-13462 namespace + cat /tmp/tmp.rjTplA2ftP + rm /tmp/tmp.I8Ar17M1aV /tmp/tmp.rjTplA2ftP + return 0 + create_secrets + echo 'Creating secrets and start client' Creating secrets and start client + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/e2e-tests/demand-backup-if-unhealthy/conf/secrets.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.TmImbRK6ka ++ mktemp + local LAST_ERR=/tmp/tmp.tOMR3rYGGC + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/e2e-tests/demand-backup-if-unhealthy/conf/secrets.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/e2e-tests/conf/client.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.TmImbRK6ka secret/some-users created deployment.apps/psmdb-client created + cat /tmp/tmp.tOMR3rYGGC + rm /tmp/tmp.TmImbRK6ka /tmp/tmp.tOMR3rYGGC + return 0 + apply_s3_storage_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.YuF5VKcwBu ++ mktemp + local LAST_ERR=/tmp/tmp.n62P3aernK + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.YuF5VKcwBu secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created secret/gcp-cs-sa-key-secret created + cat /tmp/tmp.n62P3aernK + rm /tmp/tmp.YuF5VKcwBu /tmp/tmp.n62P3aernK + return 0 + cluster=some-name + desc 'create PSMDB cluster some-name' + set +o xtrace ----------------------------------------------------------------------------------- create PSMDB cluster some-name ----------------------------------------------------------------------------------- + apply_cluster /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/e2e-tests/demand-backup-if-unhealthy/conf/some-name.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/e2e-tests/demand-backup-if-unhealthy/conf/some-name.yml + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/e2e-tests/demand-backup-if-unhealthy/conf/some-name.yml ++ mktemp + yq eval '(.spec | select(.image == null)).image = "perconalab/percona-server-mongodb-operator:main-mongod8.0"' + yq eval '(.spec | select(has("pmm"))).pmm.image = "percona/pmm-client:2.44.1-1"' + local LAST_OUT=/tmp/tmp.HvYoCi3Zud ++ mktemp + yq eval '(.spec | select(has("backup"))).backup.image = "perconalab/percona-server-mongodb-operator:main-backup"' + /usr/sbin/sed -e s/NAME_SPACE/demand-backup-if-unhealthy-13462/g + local LAST_ERR=/tmp/tmp.7UCqwxM9ib + local exit_status=0 + local timeout=4 + yq eval '(.spec | select(has("initImage"))).initImage = "perconalab/percona-server-mongodb-operator:PR-2159-33bd82f1"' + yq eval '.spec.upgradeOptions.apply="Never"' ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.HvYoCi3Zud perconaservermongodb.psmdb.percona.com/some-name created + cat /tmp/tmp.7UCqwxM9ib + rm /tmp/tmp.HvYoCi3Zud /tmp/tmp.7UCqwxM9ib + return 0 + wait_for_cluster_state some-name ready + local cluster_name=some-name + local target_state=ready + echo -n 'Waiting for psmdb/some-name to reach ready state' Waiting for psmdb/some-name to reach ready state+ local timeout=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PfP413OqFu +++ mktemp ++ local LAST_ERR=/tmp/tmp.NYRBrRGBFV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.PfP413OqFu ++ cat /tmp/tmp.NYRBrRGBFV ++ rm /tmp/tmp.PfP413OqFu /tmp/tmp.NYRBrRGBFV ++ return 0 + [[ '' =~ ready ]] + sleep 1 + timeout=1 + echo -n . .+ [[ 1 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.owq9O07BPd +++ mktemp ++ local LAST_ERR=/tmp/tmp.aqJm4kSR3P ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.owq9O07BPd ++ cat /tmp/tmp.aqJm4kSR3P ++ rm /tmp/tmp.owq9O07BPd /tmp/tmp.aqJm4kSR3P ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=2 + echo -n . .+ [[ 2 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wCrWrOi3J0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.oAK71sXeMk ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.wCrWrOi3J0 ++ cat /tmp/tmp.oAK71sXeMk ++ rm /tmp/tmp.wCrWrOi3J0 /tmp/tmp.oAK71sXeMk ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=3 + echo -n . .+ [[ 3 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9yJ97VHr3n +++ mktemp ++ local LAST_ERR=/tmp/tmp.yxa7kMRJ9C ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9yJ97VHr3n ++ cat /tmp/tmp.yxa7kMRJ9C ++ rm /tmp/tmp.9yJ97VHr3n /tmp/tmp.yxa7kMRJ9C ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=4 + echo -n . .+ [[ 4 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KMg5OtEcjM +++ mktemp ++ local LAST_ERR=/tmp/tmp.2GQarY3pzu ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.KMg5OtEcjM ++ cat /tmp/tmp.2GQarY3pzu ++ rm /tmp/tmp.KMg5OtEcjM /tmp/tmp.2GQarY3pzu ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=5 + echo -n . .+ [[ 5 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.64ANvavT4g +++ mktemp ++ local LAST_ERR=/tmp/tmp.5y9OCB1r29 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.64ANvavT4g ++ cat /tmp/tmp.5y9OCB1r29 ++ rm /tmp/tmp.64ANvavT4g /tmp/tmp.5y9OCB1r29 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=6 + echo -n . .+ [[ 6 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jRPbUNzMR0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.1XKHkCLggV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jRPbUNzMR0 ++ cat /tmp/tmp.1XKHkCLggV ++ rm /tmp/tmp.jRPbUNzMR0 /tmp/tmp.1XKHkCLggV ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=7 + echo -n . .+ [[ 7 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SU0kA9qmDG +++ mktemp ++ local LAST_ERR=/tmp/tmp.ah9qj676TG ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.SU0kA9qmDG ++ cat /tmp/tmp.ah9qj676TG ++ rm /tmp/tmp.SU0kA9qmDG /tmp/tmp.ah9qj676TG ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=8 + echo -n . .+ [[ 8 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.d56x2bfhaj +++ mktemp ++ local LAST_ERR=/tmp/tmp.apEVmib4FL ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.d56x2bfhaj ++ cat /tmp/tmp.apEVmib4FL ++ rm /tmp/tmp.d56x2bfhaj /tmp/tmp.apEVmib4FL ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=9 + echo -n . .+ [[ 9 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FVBW0qKbVC +++ mktemp ++ local LAST_ERR=/tmp/tmp.uSGVnyUoEM ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.FVBW0qKbVC ++ cat /tmp/tmp.uSGVnyUoEM ++ rm /tmp/tmp.FVBW0qKbVC /tmp/tmp.uSGVnyUoEM ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=10 + echo -n . .+ [[ 10 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YuXfnXPr6y +++ mktemp ++ local LAST_ERR=/tmp/tmp.144iGIO2hS ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.YuXfnXPr6y ++ cat /tmp/tmp.144iGIO2hS ++ rm /tmp/tmp.YuXfnXPr6y /tmp/tmp.144iGIO2hS ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=11 + echo -n . .+ [[ 11 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HcP2cyVrwH +++ mktemp ++ local LAST_ERR=/tmp/tmp.hFbtJHkeW1 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.HcP2cyVrwH ++ cat /tmp/tmp.hFbtJHkeW1 ++ rm /tmp/tmp.HcP2cyVrwH /tmp/tmp.hFbtJHkeW1 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=12 + echo -n . .+ [[ 12 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tbaVeNZZfY +++ mktemp ++ local LAST_ERR=/tmp/tmp.bHe2qqP23x ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.tbaVeNZZfY ++ cat /tmp/tmp.bHe2qqP23x ++ rm /tmp/tmp.tbaVeNZZfY /tmp/tmp.bHe2qqP23x ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=13 + echo -n . .+ [[ 13 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oUskn6kD1A +++ mktemp ++ local LAST_ERR=/tmp/tmp.LvEKtVWrE9 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.oUskn6kD1A ++ cat /tmp/tmp.LvEKtVWrE9 ++ rm /tmp/tmp.oUskn6kD1A /tmp/tmp.LvEKtVWrE9 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=14 + echo -n . .+ [[ 14 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6va0l2P0eO +++ mktemp ++ local LAST_ERR=/tmp/tmp.9qwGspAuly ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.6va0l2P0eO ++ cat /tmp/tmp.9qwGspAuly ++ rm /tmp/tmp.6va0l2P0eO /tmp/tmp.9qwGspAuly ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=15 + echo -n . .+ [[ 15 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VoZtuUREbz +++ mktemp ++ local LAST_ERR=/tmp/tmp.wgV0UvapbP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.VoZtuUREbz ++ cat /tmp/tmp.wgV0UvapbP ++ rm /tmp/tmp.VoZtuUREbz /tmp/tmp.wgV0UvapbP ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=16 + echo -n . .+ [[ 16 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RJTHq0Y8gi +++ mktemp ++ local LAST_ERR=/tmp/tmp.ouFqJMASCX ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.RJTHq0Y8gi ++ cat /tmp/tmp.ouFqJMASCX ++ rm /tmp/tmp.RJTHq0Y8gi /tmp/tmp.ouFqJMASCX ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=17 + echo -n . .+ [[ 17 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.F79g8htiD6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.XdMdGw6hJB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.F79g8htiD6 ++ cat /tmp/tmp.XdMdGw6hJB ++ rm /tmp/tmp.F79g8htiD6 /tmp/tmp.XdMdGw6hJB ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=18 + echo -n . .+ [[ 18 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cKBqDCuJo1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.EwMKftluJs ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.cKBqDCuJo1 ++ cat /tmp/tmp.EwMKftluJs ++ rm /tmp/tmp.cKBqDCuJo1 /tmp/tmp.EwMKftluJs ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=19 + echo -n . .+ [[ 19 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hIFDz3yUsR +++ mktemp ++ local LAST_ERR=/tmp/tmp.AE97VHggBp ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.hIFDz3yUsR ++ cat /tmp/tmp.AE97VHggBp ++ rm /tmp/tmp.hIFDz3yUsR /tmp/tmp.AE97VHggBp ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=20 + echo -n . .+ [[ 20 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NQOKRG0N6r +++ mktemp ++ local LAST_ERR=/tmp/tmp.cQekqTB4h8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.NQOKRG0N6r ++ cat /tmp/tmp.cQekqTB4h8 ++ rm /tmp/tmp.NQOKRG0N6r /tmp/tmp.cQekqTB4h8 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=21 + echo -n . .+ [[ 21 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.w0EIY1E1vW +++ mktemp ++ local LAST_ERR=/tmp/tmp.oKA4EcO0cm ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.w0EIY1E1vW ++ cat /tmp/tmp.oKA4EcO0cm ++ rm /tmp/tmp.w0EIY1E1vW /tmp/tmp.oKA4EcO0cm ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=22 + echo -n . .+ [[ 22 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xXM208RdCF +++ mktemp ++ local LAST_ERR=/tmp/tmp.z9wNEy9TPW ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xXM208RdCF ++ cat /tmp/tmp.z9wNEy9TPW ++ rm /tmp/tmp.xXM208RdCF /tmp/tmp.z9wNEy9TPW ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=23 + echo -n . .+ [[ 23 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ycyojgcy0z +++ mktemp ++ local LAST_ERR=/tmp/tmp.waGh6Hyrjf ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Ycyojgcy0z ++ cat /tmp/tmp.waGh6Hyrjf ++ rm /tmp/tmp.Ycyojgcy0z /tmp/tmp.waGh6Hyrjf ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=24 + echo -n . .+ [[ 24 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lxvGrHGYHS +++ mktemp ++ local LAST_ERR=/tmp/tmp.D4DB1otszX ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.lxvGrHGYHS ++ cat /tmp/tmp.D4DB1otszX ++ rm /tmp/tmp.lxvGrHGYHS /tmp/tmp.D4DB1otszX ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=25 + echo -n . .+ [[ 25 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SzCTx2bg6k +++ mktemp ++ local LAST_ERR=/tmp/tmp.u9X62rgbb6 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.SzCTx2bg6k ++ cat /tmp/tmp.u9X62rgbb6 ++ rm /tmp/tmp.SzCTx2bg6k /tmp/tmp.u9X62rgbb6 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=26 + echo -n . .+ [[ 26 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZsGeE11zA1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.kBqxF8IS0H ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ZsGeE11zA1 ++ cat /tmp/tmp.kBqxF8IS0H ++ rm /tmp/tmp.ZsGeE11zA1 /tmp/tmp.kBqxF8IS0H ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=27 + echo -n . .+ [[ 27 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sEvXFqs6ZA +++ mktemp ++ local LAST_ERR=/tmp/tmp.uGMxuiZ402 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.sEvXFqs6ZA ++ cat /tmp/tmp.uGMxuiZ402 ++ rm /tmp/tmp.sEvXFqs6ZA /tmp/tmp.uGMxuiZ402 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=28 + echo -n . .+ [[ 28 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Z90kDDtJyg +++ mktemp ++ local LAST_ERR=/tmp/tmp.NrPau1CXTR ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Z90kDDtJyg ++ cat /tmp/tmp.NrPau1CXTR ++ rm /tmp/tmp.Z90kDDtJyg /tmp/tmp.NrPau1CXTR ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=29 + echo -n . .+ [[ 29 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PFYmAiAqbG +++ mktemp ++ local LAST_ERR=/tmp/tmp.93vhlKrFEO ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.PFYmAiAqbG ++ cat /tmp/tmp.93vhlKrFEO ++ rm /tmp/tmp.PFYmAiAqbG /tmp/tmp.93vhlKrFEO ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=30 + echo -n . .+ [[ 30 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cj1o5TQEpV +++ mktemp ++ local LAST_ERR=/tmp/tmp.jUIrIaqoP2 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.cj1o5TQEpV ++ cat /tmp/tmp.jUIrIaqoP2 ++ rm /tmp/tmp.cj1o5TQEpV /tmp/tmp.jUIrIaqoP2 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=31 + echo -n . .+ [[ 31 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LZ000dS18C +++ mktemp ++ local LAST_ERR=/tmp/tmp.jbWtaSBAHx ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.LZ000dS18C ++ cat /tmp/tmp.jbWtaSBAHx ++ rm /tmp/tmp.LZ000dS18C /tmp/tmp.jbWtaSBAHx ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=32 + echo -n . .+ [[ 32 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PaXlcmvjxz +++ mktemp ++ local LAST_ERR=/tmp/tmp.xczIKlBxmy ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.PaXlcmvjxz ++ cat /tmp/tmp.xczIKlBxmy ++ rm /tmp/tmp.PaXlcmvjxz /tmp/tmp.xczIKlBxmy ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=33 + echo -n . .+ [[ 33 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PXtqHHYeeV +++ mktemp ++ local LAST_ERR=/tmp/tmp.SHDYSwU3gt ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.PXtqHHYeeV ++ cat /tmp/tmp.SHDYSwU3gt ++ rm /tmp/tmp.PXtqHHYeeV /tmp/tmp.SHDYSwU3gt ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=34 + echo -n . .+ [[ 34 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jnlVspz3ka +++ mktemp ++ local LAST_ERR=/tmp/tmp.PSuuwQ7d3t ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jnlVspz3ka ++ cat /tmp/tmp.PSuuwQ7d3t ++ rm /tmp/tmp.jnlVspz3ka /tmp/tmp.PSuuwQ7d3t ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=35 + echo -n . .+ [[ 35 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XHZx1orBrH +++ mktemp ++ local LAST_ERR=/tmp/tmp.BlwqqziNfs ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.XHZx1orBrH ++ cat /tmp/tmp.BlwqqziNfs ++ rm /tmp/tmp.XHZx1orBrH /tmp/tmp.BlwqqziNfs ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=36 + echo -n . .+ [[ 36 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.P4xoQZOFGi +++ mktemp ++ local LAST_ERR=/tmp/tmp.0ko0AzKxKb ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.P4xoQZOFGi ++ cat /tmp/tmp.0ko0AzKxKb ++ rm /tmp/tmp.P4xoQZOFGi /tmp/tmp.0ko0AzKxKb ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=37 + echo -n . .+ [[ 37 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qs7Exut40t +++ mktemp ++ local LAST_ERR=/tmp/tmp.7FeNn31VDx ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.qs7Exut40t ++ cat /tmp/tmp.7FeNn31VDx ++ rm /tmp/tmp.qs7Exut40t /tmp/tmp.7FeNn31VDx ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=38 + echo -n . .+ [[ 38 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eVfUZDRxPj +++ mktemp ++ local LAST_ERR=/tmp/tmp.wVvNk3YfUY ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.eVfUZDRxPj ++ cat /tmp/tmp.wVvNk3YfUY ++ rm /tmp/tmp.eVfUZDRxPj /tmp/tmp.wVvNk3YfUY ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=39 + echo -n . .+ [[ 39 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8zuV3ee8Zj +++ mktemp ++ local LAST_ERR=/tmp/tmp.bU8rJs6HpS ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.8zuV3ee8Zj ++ cat /tmp/tmp.bU8rJs6HpS ++ rm /tmp/tmp.8zuV3ee8Zj /tmp/tmp.bU8rJs6HpS ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=40 + echo -n . .+ [[ 40 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gjiLShMG9Y +++ mktemp ++ local LAST_ERR=/tmp/tmp.toLvEOcrTN ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.gjiLShMG9Y ++ cat /tmp/tmp.toLvEOcrTN ++ rm /tmp/tmp.gjiLShMG9Y /tmp/tmp.toLvEOcrTN ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=41 + echo -n . .+ [[ 41 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pE3jRmyI0r +++ mktemp ++ local LAST_ERR=/tmp/tmp.BahO7mBnLm ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.pE3jRmyI0r ++ cat /tmp/tmp.BahO7mBnLm ++ rm /tmp/tmp.pE3jRmyI0r /tmp/tmp.BahO7mBnLm ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=42 + echo -n . .+ [[ 42 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TTIb9G7jYm +++ mktemp ++ local LAST_ERR=/tmp/tmp.XOgXWv1rHE ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.TTIb9G7jYm ++ cat /tmp/tmp.XOgXWv1rHE ++ rm /tmp/tmp.TTIb9G7jYm /tmp/tmp.XOgXWv1rHE ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=43 + echo -n . .+ [[ 43 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RbkzmGJrgI +++ mktemp ++ local LAST_ERR=/tmp/tmp.jcBlBgUYN6 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.RbkzmGJrgI ++ cat /tmp/tmp.jcBlBgUYN6 ++ rm /tmp/tmp.RbkzmGJrgI /tmp/tmp.jcBlBgUYN6 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=44 + echo -n . .+ [[ 44 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EJHIz0DdaS +++ mktemp ++ local LAST_ERR=/tmp/tmp.34hYTnEqhT ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.EJHIz0DdaS ++ cat /tmp/tmp.34hYTnEqhT ++ rm /tmp/tmp.EJHIz0DdaS /tmp/tmp.34hYTnEqhT ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=45 + echo -n . .+ [[ 45 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GWSy81aME0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.j46ZXL1l60 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.GWSy81aME0 ++ cat /tmp/tmp.j46ZXL1l60 ++ rm /tmp/tmp.GWSy81aME0 /tmp/tmp.j46ZXL1l60 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=46 + echo -n . .+ [[ 46 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4FUqlvqmDE +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ud4KrEAQFu ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4FUqlvqmDE ++ cat /tmp/tmp.Ud4KrEAQFu ++ rm /tmp/tmp.4FUqlvqmDE /tmp/tmp.Ud4KrEAQFu ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=47 + echo -n . .+ [[ 47 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uGmzPLtoBn +++ mktemp ++ local LAST_ERR=/tmp/tmp.3ueHN6aI0Z ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.uGmzPLtoBn ++ cat /tmp/tmp.3ueHN6aI0Z ++ rm /tmp/tmp.uGmzPLtoBn /tmp/tmp.3ueHN6aI0Z ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=48 + echo -n . .+ [[ 48 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HAJcIjURkj +++ mktemp ++ local LAST_ERR=/tmp/tmp.QulGgzE7z8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.HAJcIjURkj ++ cat /tmp/tmp.QulGgzE7z8 ++ rm /tmp/tmp.HAJcIjURkj /tmp/tmp.QulGgzE7z8 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=49 + echo -n . .+ [[ 49 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LDM1J1hAA2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.k4Qvfb9ciL ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.LDM1J1hAA2 ++ cat /tmp/tmp.k4Qvfb9ciL ++ rm /tmp/tmp.LDM1J1hAA2 /tmp/tmp.k4Qvfb9ciL ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=50 + echo -n . .+ [[ 50 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MbqPCoHRrh +++ mktemp ++ local LAST_ERR=/tmp/tmp.gOC8kVCkCW ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.MbqPCoHRrh ++ cat /tmp/tmp.gOC8kVCkCW ++ rm /tmp/tmp.MbqPCoHRrh /tmp/tmp.gOC8kVCkCW ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=51 + echo -n . .+ [[ 51 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7HYBaouxFq +++ mktemp ++ local LAST_ERR=/tmp/tmp.sb1aIx7DI9 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.7HYBaouxFq ++ cat /tmp/tmp.sb1aIx7DI9 ++ rm /tmp/tmp.7HYBaouxFq /tmp/tmp.sb1aIx7DI9 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=52 + echo -n . .+ [[ 52 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.y9Y20qtZ13 +++ mktemp ++ local LAST_ERR=/tmp/tmp.OfUIFWU6WK ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.y9Y20qtZ13 ++ cat /tmp/tmp.OfUIFWU6WK ++ rm /tmp/tmp.y9Y20qtZ13 /tmp/tmp.OfUIFWU6WK ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=53 + echo -n . .+ [[ 53 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bqeSxQFFcZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.HRf6OHNgIR ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.bqeSxQFFcZ ++ cat /tmp/tmp.HRf6OHNgIR ++ rm /tmp/tmp.bqeSxQFFcZ /tmp/tmp.HRf6OHNgIR ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=54 + echo -n . .+ [[ 54 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PDfcZnLw0F +++ mktemp ++ local LAST_ERR=/tmp/tmp.G6FjJlknaa ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.PDfcZnLw0F ++ cat /tmp/tmp.G6FjJlknaa ++ rm /tmp/tmp.PDfcZnLw0F /tmp/tmp.G6FjJlknaa ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=55 + echo -n . .+ [[ 55 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RJmDr3OBVt +++ mktemp ++ local LAST_ERR=/tmp/tmp.YSbQTg4NZM ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.RJmDr3OBVt ++ cat /tmp/tmp.YSbQTg4NZM ++ rm /tmp/tmp.RJmDr3OBVt /tmp/tmp.YSbQTg4NZM ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=56 + echo -n . .+ [[ 56 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KLy13UwsQV +++ mktemp ++ local LAST_ERR=/tmp/tmp.Wc8lBTGUu5 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.KLy13UwsQV ++ cat /tmp/tmp.Wc8lBTGUu5 ++ rm /tmp/tmp.KLy13UwsQV /tmp/tmp.Wc8lBTGUu5 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=57 + echo -n . .+ [[ 57 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yelBIPfwC9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.vFv5M0T5Og ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.yelBIPfwC9 ++ cat /tmp/tmp.vFv5M0T5Og ++ rm /tmp/tmp.yelBIPfwC9 /tmp/tmp.vFv5M0T5Og ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=58 + echo -n . .+ [[ 58 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4LAeAzPCuS +++ mktemp ++ local LAST_ERR=/tmp/tmp.A6bZLN1aO1 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4LAeAzPCuS ++ cat /tmp/tmp.A6bZLN1aO1 ++ rm /tmp/tmp.4LAeAzPCuS /tmp/tmp.A6bZLN1aO1 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=59 + echo -n . .+ [[ 59 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DgKpip3zpM +++ mktemp ++ local LAST_ERR=/tmp/tmp.hL2wTEys8z ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.DgKpip3zpM ++ cat /tmp/tmp.hL2wTEys8z ++ rm /tmp/tmp.DgKpip3zpM /tmp/tmp.hL2wTEys8z ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=60 + echo -n . .+ [[ 60 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.A1nPnxkt0U +++ mktemp ++ local LAST_ERR=/tmp/tmp.xEIXLUXcAm ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.A1nPnxkt0U ++ cat /tmp/tmp.xEIXLUXcAm ++ rm /tmp/tmp.A1nPnxkt0U /tmp/tmp.xEIXLUXcAm ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=61 + echo -n . .+ [[ 61 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.n8byWaIlPb +++ mktemp ++ local LAST_ERR=/tmp/tmp.d8PkK0KZHA ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.n8byWaIlPb ++ cat /tmp/tmp.d8PkK0KZHA ++ rm /tmp/tmp.n8byWaIlPb /tmp/tmp.d8PkK0KZHA ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=62 + echo -n . .+ [[ 62 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LNMfuhnQrP +++ mktemp ++ local LAST_ERR=/tmp/tmp.Eo1y4lxl43 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.LNMfuhnQrP ++ cat /tmp/tmp.Eo1y4lxl43 ++ rm /tmp/tmp.LNMfuhnQrP /tmp/tmp.Eo1y4lxl43 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=63 + echo -n . .+ [[ 63 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.u0IcHfLCJA +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZzpE2B00oW ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.u0IcHfLCJA ++ cat /tmp/tmp.ZzpE2B00oW ++ rm /tmp/tmp.u0IcHfLCJA /tmp/tmp.ZzpE2B00oW ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=64 + echo -n . .+ [[ 64 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dswya0OVul +++ mktemp ++ local LAST_ERR=/tmp/tmp.yi2egE742M ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.dswya0OVul ++ cat /tmp/tmp.yi2egE742M ++ rm /tmp/tmp.dswya0OVul /tmp/tmp.yi2egE742M ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=65 + echo -n . .+ [[ 65 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.szx43hbjNG +++ mktemp ++ local LAST_ERR=/tmp/tmp.mOM0qI721q ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.szx43hbjNG ++ cat /tmp/tmp.mOM0qI721q ++ rm /tmp/tmp.szx43hbjNG /tmp/tmp.mOM0qI721q ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=66 + echo -n . .+ [[ 66 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zQgzMNgBI7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Q4M8xX0eMF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.zQgzMNgBI7 ++ cat /tmp/tmp.Q4M8xX0eMF ++ rm /tmp/tmp.zQgzMNgBI7 /tmp/tmp.Q4M8xX0eMF ++ return 0 + [[ ready =~ ready ]] + echo + log 'psmdb/some-name is ready: OK' + set +o xtrace [2025-12-22T07:45:44+0000] psmdb/some-name is ready: OK + desc 'Adding new user' + set +o xtrace ----------------------------------------------------------------------------------- Adding new user ----------------------------------------------------------------------------------- + run_mongo 'db.createUser({user:"myApp",pwd:"myPass",roles:[{db:"myApp",role:"readWrite"}]})' userAdmin:userAdmin123456@some-name-rs0.demand-backup-if-unhealthy-13462 + local 'command=db.createUser({user:"myApp",pwd:"myPass",roles:[{db:"myApp",role:"readWrite"}]})' + local uri=userAdmin:userAdmin123456@some-name-rs0.demand-backup-if-unhealthy-13462 + local driver=mongodb+srv + local suffix=.svc.cluster.local + local mongo_flag= + local replica_set=rs0 + [[ userAdmin:userAdmin123456@some-name-rs0.demand-backup-if-unhealthy-13462 == *cfg* ]] ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jijGq1hSYq +++ mktemp ++ local LAST_ERR=/tmp/tmp.BDev0y3Rlu ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jijGq1hSYq ++ cat /tmp/tmp.BDev0y3Rlu ++ rm /tmp/tmp.jijGq1hSYq /tmp/tmp.BDev0y3Rlu ++ return 0 + local client_container=psmdb-client-696897d69b-sc75l + kubectl_bin exec psmdb-client-696897d69b-sc75l -- bash -c 'printf '\''db.createUser({user:"myApp",pwd:"myPass",roles:[{db:"myApp",role:"readWrite"}]})\n'\'' | mongo mongodb+srv://userAdmin:userAdmin123456@some-name-rs0.demand-backup-if-unhealthy-13462.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.ulKS0sOEZq ++ mktemp + local LAST_ERR=/tmp/tmp.phzrwVddsd + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-696897d69b-sc75l -- bash -c 'printf '\''db.createUser({user:"myApp",pwd:"myPass",roles:[{db:"myApp",role:"readWrite"}]})\n'\'' | mongo mongodb+srv://userAdmin:userAdmin123456@some-name-rs0.demand-backup-if-unhealthy-13462.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ulKS0sOEZq Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-rs0-2.some-name-rs0.demand-backup-if-unhealthy-13462.svc.cluster.local:27017,some-name-rs0-1.some-name-rs0.demand-backup-if-unhealthy-13462.svc.cluster.local:27017,some-name-rs0-0.some-name-rs0.demand-backup-if-unhealthy-13462.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb&replicaSet=rs0&ssl=false Implicit session: session { "id" : UUID("54d4d49b-77bd-49e7-9640-fb16ae838078") } Percona Server for MongoDB server version: v8.0.16-5 WARNING: shell and server versions do not match Successfully added user: { "user" : "myApp", "roles" : [ { "db" : "myApp", "role" : "readWrite" } ] } bye + cat /tmp/tmp.phzrwVddsd + rm /tmp/tmp.ulKS0sOEZq /tmp/tmp.phzrwVddsd + return 0 + echo -n 'Verifying user creation...' Verifying user creation...+ until_with_timeout 10 check_user + local timeout=10 + shift + local elapsed=0 + check_user + run_mongo 'db.getUser("myApp")' userAdmin:userAdmin123456@some-name-rs0.demand-backup-if-unhealthy-13462 + local 'command=db.getUser("myApp")' + grep -q '"user" : "myApp"' + local uri=userAdmin:userAdmin123456@some-name-rs0.demand-backup-if-unhealthy-13462 + local driver=mongodb+srv + local suffix=.svc.cluster.local + local mongo_flag= + local replica_set=rs0 + [[ userAdmin:userAdmin123456@some-name-rs0.demand-backup-if-unhealthy-13462 == *cfg* ]] ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.modEKBL6LU +++ mktemp ++ local LAST_ERR=/tmp/tmp.82gUBRFyFG ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.modEKBL6LU ++ cat /tmp/tmp.82gUBRFyFG ++ rm /tmp/tmp.modEKBL6LU /tmp/tmp.82gUBRFyFG ++ return 0 + local client_container=psmdb-client-696897d69b-sc75l + kubectl_bin exec psmdb-client-696897d69b-sc75l -- bash -c 'printf '\''db.getUser("myApp")\n'\'' | mongo mongodb+srv://userAdmin:userAdmin123456@some-name-rs0.demand-backup-if-unhealthy-13462.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.rSfFhKPWni ++ mktemp + local LAST_ERR=/tmp/tmp.s3Bm6X2txj + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-696897d69b-sc75l -- bash -c 'printf '\''db.getUser("myApp")\n'\'' | mongo mongodb+srv://userAdmin:userAdmin123456@some-name-rs0.demand-backup-if-unhealthy-13462.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.rSfFhKPWni + cat /tmp/tmp.s3Bm6X2txj + rm /tmp/tmp.rSfFhKPWni /tmp/tmp.s3Bm6X2txj + return 0 + return 0 + desc 'Testing logical backup with an unhealthy container' + set +o xtrace ----------------------------------------------------------------------------------- Testing logical backup with an unhealthy container ----------------------------------------------------------------------------------- + type=logical + write_data 'This is some data for backup if unhealthy' + local 'data=This is some data for backup if unhealthy' + local find_prefix= + run_mongo 'use myApp\n db.test.insert({ x: "This is some data for backup if unhealthy" })' myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-13462 + local 'command=use myApp\n db.test.insert({ x: "This is some data for backup if unhealthy" })' + local uri=myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-13462 + local driver=mongodb+srv + local suffix=.svc.cluster.local + local mongo_flag= + local replica_set=rs0 + [[ myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-13462 == *cfg* ]] ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ke88weG02Y +++ mktemp ++ local LAST_ERR=/tmp/tmp.a8GSZUXwLu ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ke88weG02Y ++ cat /tmp/tmp.a8GSZUXwLu ++ rm /tmp/tmp.ke88weG02Y /tmp/tmp.a8GSZUXwLu ++ return 0 + local client_container=psmdb-client-696897d69b-sc75l + kubectl_bin exec psmdb-client-696897d69b-sc75l -- bash -c 'printf '\''use myApp\n db.test.insert({ x: "This is some data for backup if unhealthy" })\n'\'' | mongo mongodb+srv://myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-13462.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.tjk0yMXJMh ++ mktemp + local LAST_ERR=/tmp/tmp.P2pp9jQNNK + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-696897d69b-sc75l -- bash -c 'printf '\''use myApp\n db.test.insert({ x: "This is some data for backup if unhealthy" })\n'\'' | mongo mongodb+srv://myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-13462.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.tjk0yMXJMh Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-rs0-2.some-name-rs0.demand-backup-if-unhealthy-13462.svc.cluster.local:27017,some-name-rs0-1.some-name-rs0.demand-backup-if-unhealthy-13462.svc.cluster.local:27017,some-name-rs0-0.some-name-rs0.demand-backup-if-unhealthy-13462.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb&replicaSet=rs0&ssl=false Implicit session: session { "id" : UUID("9b61fd70-a1ce-40e1-bad6-8f2fc906d2e0") } Percona Server for MongoDB server version: v8.0.16-5 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.P2pp9jQNNK + rm /tmp/tmp.tjk0yMXJMh /tmp/tmp.P2pp9jQNNK + return 0 + compare_mongo_cmd find myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-13462 '' .svc.cluster.local myApp test + local command=find + local uri=myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-13462 + local postfix= + local suffix=.svc.cluster.local + local database=myApp + local collection=test + local sort= + local tls=false + local replicaset= + local 'full_command=db.test.find()' + [[ -n '' ]] + log 'running db.test.find() in myApp' + set +o xtrace [2025-12-22T07:45:55+0000] running db.test.find() in myApp + [[ false == \t\r\u\e ]] + mongo_command=run_mongo + run_mongo 'use myApp\n db.test.find()' myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-13462 mongodb .svc.cluster.local '' + local 'command=use myApp\n db.test.find()' + local uri=myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-13462 + local driver=mongodb + local suffix=.svc.cluster.local + local mongo_flag= + local replica_set=rs0 + [[ myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-13462 == *cfg* ]] + egrep -v 'I NETWORK|W NETWORK|F NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:|Started a new thread for the timer service' + /usr/sbin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp egrep: warning: egrep is obsolescent; using grep -E ++ local LAST_OUT=/tmp/tmp.VMa5SuZg2G +++ mktemp ++ local LAST_ERR=/tmp/tmp.dxrJYku4Le ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.VMa5SuZg2G ++ cat /tmp/tmp.dxrJYku4Le ++ rm /tmp/tmp.VMa5SuZg2G /tmp/tmp.dxrJYku4Le ++ return 0 + local client_container=psmdb-client-696897d69b-sc75l + kubectl_bin exec psmdb-client-696897d69b-sc75l -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-13462.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.FrPsJ4U41Y ++ mktemp + local LAST_ERR=/tmp/tmp.ulRq6dLhnk + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-696897d69b-sc75l -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-13462.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.FrPsJ4U41Y + cat /tmp/tmp.ulRq6dLhnk + rm /tmp/tmp.FrPsJ4U41Y /tmp/tmp.ulRq6dLhnk + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/e2e-tests/demand-backup-if-unhealthy/compare/find.json /tmp/tmp.5ZphUZDmGh/find + wait_sts_to_become_not_ready + local sts_name=some-name-rs0 + local timeout=60 + local pod_name=some-name-rs0-1 + local interval=2 + local elapsed=0 + echo 'Updating cluster with invalid image...' Updating cluster with invalid image... + update_with_invalid_db_image + local cluster_name=some-name + yq /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/e2e-tests/demand-backup-if-unhealthy/conf/some-name.yml + yq '.spec.image="invalid-image:invalid-tag"' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.B1iEQXtmv7 ++ mktemp + local LAST_ERR=/tmp/tmp.NUKpbN051U + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.B1iEQXtmv7 perconaservermongodb.psmdb.percona.com/some-name configured + cat /tmp/tmp.NUKpbN051U + rm /tmp/tmp.B1iEQXtmv7 /tmp/tmp.NUKpbN051U + return 0 + echo -n 'Wait for statefulset some-name-rs0 to become not ready...' Wait for statefulset some-name-rs0 to become not ready...+ is_sts_ready some-name-rs0 + local sts_name=some-name-rs0 + local replicas ready_replicas + kubectl_bin get sts some-name-rs0 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YyVsIeqN0J +++ mktemp ++ local LAST_ERR=/tmp/tmp.bStdJsD4qu ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.YyVsIeqN0J ++ cat /tmp/tmp.bStdJsD4qu ++ rm /tmp/tmp.YyVsIeqN0J /tmp/tmp.bStdJsD4qu ++ return 0 + replicas=3 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cm6iV4u254 +++ mktemp ++ local LAST_ERR=/tmp/tmp.dP4KqRv9tr ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.cm6iV4u254 ++ cat /tmp/tmp.dP4KqRv9tr ++ rm /tmp/tmp.cm6iV4u254 /tmp/tmp.dP4KqRv9tr ++ return 0 + ready_replicas=3 + [[ 3 -eq 3 ]] + (( elapsed >= timeout )) + sleep 2 + (( elapsed += interval )) + echo -n . .+ is_sts_ready some-name-rs0 + local sts_name=some-name-rs0 + local replicas ready_replicas + kubectl_bin get sts some-name-rs0 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZL5OLtwjs0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.PXbccDrMtf ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ZL5OLtwjs0 ++ cat /tmp/tmp.PXbccDrMtf ++ rm /tmp/tmp.ZL5OLtwjs0 /tmp/tmp.PXbccDrMtf ++ return 0 + replicas=3 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AqRsjXgZ7U +++ mktemp ++ local LAST_ERR=/tmp/tmp.AEWXgrPVdg ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.AqRsjXgZ7U ++ cat /tmp/tmp.AEWXgrPVdg ++ rm /tmp/tmp.AqRsjXgZ7U /tmp/tmp.AEWXgrPVdg ++ return 0 + ready_replicas=3 + [[ 3 -eq 3 ]] + (( elapsed >= timeout )) + sleep 2 + (( elapsed += interval )) + echo -n . .+ is_sts_ready some-name-rs0 + local sts_name=some-name-rs0 + local replicas ready_replicas + kubectl_bin get sts some-name-rs0 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.K8zJPaVHAq +++ mktemp ++ local LAST_ERR=/tmp/tmp.HBHo3ZFTMq ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.K8zJPaVHAq ++ cat /tmp/tmp.HBHo3ZFTMq ++ rm /tmp/tmp.K8zJPaVHAq /tmp/tmp.HBHo3ZFTMq ++ return 0 + replicas=3 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YnjA3p8MHe +++ mktemp ++ local LAST_ERR=/tmp/tmp.p02QHPxcW2 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.YnjA3p8MHe ++ cat /tmp/tmp.p02QHPxcW2 ++ rm /tmp/tmp.YnjA3p8MHe /tmp/tmp.p02QHPxcW2 ++ return 0 + ready_replicas=3 + [[ 3 -eq 3 ]] + (( elapsed >= timeout )) + sleep 2 + (( elapsed += interval )) + echo -n . .+ is_sts_ready some-name-rs0 + local sts_name=some-name-rs0 + local replicas ready_replicas + kubectl_bin get sts some-name-rs0 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SA3R5Ydicr +++ mktemp ++ local LAST_ERR=/tmp/tmp.N8nqGUWjx2 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.SA3R5Ydicr ++ cat /tmp/tmp.N8nqGUWjx2 ++ rm /tmp/tmp.SA3R5Ydicr /tmp/tmp.N8nqGUWjx2 ++ return 0 + replicas=3 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Dbip5QjqxM +++ mktemp ++ local LAST_ERR=/tmp/tmp.xT5k34M6bO ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Dbip5QjqxM ++ cat /tmp/tmp.xT5k34M6bO ++ rm /tmp/tmp.Dbip5QjqxM /tmp/tmp.xT5k34M6bO ++ return 0 + ready_replicas=3 + [[ 3 -eq 3 ]] + (( elapsed >= timeout )) + sleep 2 + (( elapsed += interval )) + echo -n . .+ is_sts_ready some-name-rs0 + local sts_name=some-name-rs0 + local replicas ready_replicas + kubectl_bin get sts some-name-rs0 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PBgFEHSfTq +++ mktemp ++ local LAST_ERR=/tmp/tmp.CHHRNrMKNV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.PBgFEHSfTq ++ cat /tmp/tmp.CHHRNrMKNV ++ rm /tmp/tmp.PBgFEHSfTq /tmp/tmp.CHHRNrMKNV ++ return 0 + replicas=3 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4r2ZRpVpL7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.rYfaSJxTMy ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4r2ZRpVpL7 ++ cat /tmp/tmp.rYfaSJxTMy ++ rm /tmp/tmp.4r2ZRpVpL7 /tmp/tmp.rYfaSJxTMy ++ return 0 + ready_replicas=2 + [[ 3 -eq 2 ]] + wait_for_cluster_state some-name 'initializing|error' + local cluster_name=some-name + local 'target_state=initializing|error' + echo -n 'Waiting for psmdb/some-name to reach initializing|error state' Waiting for psmdb/some-name to reach initializing|error state+ local timeout=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rzxL2jeQs1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.tBdPdLFJDx ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.rzxL2jeQs1 ++ cat /tmp/tmp.tBdPdLFJDx ++ rm /tmp/tmp.rzxL2jeQs1 /tmp/tmp.tBdPdLFJDx ++ return 0 + [[ initializing =~ initializing|error ]] + echo + log 'psmdb/some-name is initializing|error: OK' + set +o xtrace [2025-12-22T07:46:24+0000] psmdb/some-name is initializing|error: OK + backup_name_minio=backup-minio-logical + run_backup minio backup-minio-logical logical + local storage=minio + local backup_name=backup-minio-logical + local type=logical + log 'running backup backup-minio-logical' + set +o xtrace [2025-12-22T07:46:24+0000] running backup backup-minio-logical + yq eval '.metadata.name = "backup-minio-logical" | .spec.storageName = "minio" | .spec.type = "logical"' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/e2e-tests/demand-backup-if-unhealthy/conf/backup-minio.yml + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.MvzM1wtJDF ++ mktemp + local LAST_ERR=/tmp/tmp.GF94T4Wre9 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.MvzM1wtJDF perconaservermongodbbackup.psmdb.percona.com/backup-minio-logical created + cat /tmp/tmp.GF94T4Wre9 + rm /tmp/tmp.MvzM1wtJDF /tmp/tmp.GF94T4Wre9 + return 0 + verify_sts_not_ready + local sts_name=some-name-rs0 + is_sts_ready some-name-rs0 + local sts_name=some-name-rs0 + local replicas ready_replicas + kubectl_bin get sts some-name-rs0 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dVKuYaDQEE +++ mktemp ++ local LAST_ERR=/tmp/tmp.eBgid3KbRC ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.dVKuYaDQEE ++ cat /tmp/tmp.eBgid3KbRC ++ rm /tmp/tmp.dVKuYaDQEE /tmp/tmp.eBgid3KbRC ++ return 0 + replicas=3 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ghgRSuQ9nw +++ mktemp ++ local LAST_ERR=/tmp/tmp.46JCdp285D ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ghgRSuQ9nw ++ cat /tmp/tmp.46JCdp285D ++ rm /tmp/tmp.ghgRSuQ9nw /tmp/tmp.46JCdp285D ++ return 0 + ready_replicas=2 + [[ 3 -eq 2 ]] + wait_for_cluster_state some-name 'initializing|error' + local cluster_name=some-name + local 'target_state=initializing|error' + echo -n 'Waiting for psmdb/some-name to reach initializing|error state' Waiting for psmdb/some-name to reach initializing|error state+ local timeout=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4LjdJ9Cg1X +++ mktemp ++ local LAST_ERR=/tmp/tmp.tCpeNSvQaQ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4LjdJ9Cg1X ++ cat /tmp/tmp.tCpeNSvQaQ ++ rm /tmp/tmp.4LjdJ9Cg1X /tmp/tmp.tCpeNSvQaQ ++ return 0 + [[ initializing =~ initializing|error ]] + echo + log 'psmdb/some-name is initializing|error: OK' + set +o xtrace [2025-12-22T07:46:30+0000] psmdb/some-name is initializing|error: OK + desc 'Verifying backup-minio-logical and update cluster back to ready state' + set +o xtrace ----------------------------------------------------------------------------------- Verifying backup-minio-logical and update cluster back to ready state ----------------------------------------------------------------------------------- + wait_backup backup-minio-logical ready + local backup_name=backup-minio-logical + local target_state=ready + set +o xtrace waiting for backup-minio-logical to reach ready state........OK + verify_sts_not_ready + local sts_name=some-name-rs0 + is_sts_ready some-name-rs0 + local sts_name=some-name-rs0 + local replicas ready_replicas + kubectl_bin get sts some-name-rs0 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vVwP8TwBqi +++ mktemp ++ local LAST_ERR=/tmp/tmp.L0r3LMzmwa ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.vVwP8TwBqi ++ cat /tmp/tmp.L0r3LMzmwa ++ rm /tmp/tmp.vVwP8TwBqi /tmp/tmp.L0r3LMzmwa ++ return 0 + replicas=3 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wGUKgIQmEc +++ mktemp ++ local LAST_ERR=/tmp/tmp.UbtAXr1sh3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.wGUKgIQmEc ++ cat /tmp/tmp.UbtAXr1sh3 ++ rm /tmp/tmp.wGUKgIQmEc /tmp/tmp.UbtAXr1sh3 ++ return 0 + ready_replicas=2 + [[ 3 -eq 2 ]] + wait_for_cluster_state some-name 'initializing|error' + local cluster_name=some-name + local 'target_state=initializing|error' + echo -n 'Waiting for psmdb/some-name to reach initializing|error state' Waiting for psmdb/some-name to reach initializing|error state+ local timeout=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MDPC0zKwZQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.7u2aNUy0mz ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.MDPC0zKwZQ ++ cat /tmp/tmp.7u2aNUy0mz ++ rm /tmp/tmp.MDPC0zKwZQ /tmp/tmp.7u2aNUy0mz ++ return 0 + [[ initializing =~ initializing|error ]] + echo + log 'psmdb/some-name is initializing|error: OK' + set +o xtrace [2025-12-22T07:46:47+0000] psmdb/some-name is initializing|error: OK + rollback_invalid_image + local cluster_name=some-name + local timeout= + apply_cluster /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/e2e-tests/demand-backup-if-unhealthy/conf/some-name.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/e2e-tests/demand-backup-if-unhealthy/conf/some-name.yml + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/e2e-tests/demand-backup-if-unhealthy/conf/some-name.yml + yq eval '(.spec | select(.image == null)).image = "perconalab/percona-server-mongodb-operator:main-mongod8.0"' ++ mktemp + yq eval '(.spec | select(has("pmm"))).pmm.image = "percona/pmm-client:2.44.1-1"' + yq eval '(.spec | select(has("initImage"))).initImage = "perconalab/percona-server-mongodb-operator:PR-2159-33bd82f1"' + yq eval '(.spec | select(has("backup"))).backup.image = "perconalab/percona-server-mongodb-operator:main-backup"' + yq eval '.spec.upgradeOptions.apply="Never"' + local LAST_OUT=/tmp/tmp.RnkOp0md0c + /usr/sbin/sed -e s/NAME_SPACE/demand-backup-if-unhealthy-13462/g ++ mktemp + local LAST_ERR=/tmp/tmp.7Ij9Qowkev + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.RnkOp0md0c perconaservermongodb.psmdb.percona.com/some-name configured + cat /tmp/tmp.7Ij9Qowkev + rm /tmp/tmp.RnkOp0md0c /tmp/tmp.7Ij9Qowkev + return 0 + echo -n 'Waiting new statefulset version...' Waiting new statefulset version...+ kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.HR3QbywLL3 ++ mktemp + local LAST_ERR=/tmp/tmp.C5pN5fxSsL + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.HR3QbywLL3 + cat /tmp/tmp.C5pN5fxSsL + rm /tmp/tmp.HR3QbywLL3 /tmp/tmp.C5pN5fxSsL + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.B1R2fLgrXB ++ mktemp + local LAST_ERR=/tmp/tmp.YRzLS6USgK + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.B1R2fLgrXB + cat /tmp/tmp.YRzLS6USgK + rm /tmp/tmp.B1R2fLgrXB /tmp/tmp.YRzLS6USgK + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.e0zb22sCuT ++ mktemp + local LAST_ERR=/tmp/tmp.7hkvrAIFiN + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.e0zb22sCuT + cat /tmp/tmp.7hkvrAIFiN + rm /tmp/tmp.e0zb22sCuT /tmp/tmp.7hkvrAIFiN + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.efPL76JeHK ++ mktemp + local LAST_ERR=/tmp/tmp.eRCthmAoIc + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.efPL76JeHK + cat /tmp/tmp.eRCthmAoIc + rm /tmp/tmp.efPL76JeHK /tmp/tmp.eRCthmAoIc + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.TNP992xf3F ++ mktemp + local LAST_ERR=/tmp/tmp.ec51F773OL + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.TNP992xf3F + cat /tmp/tmp.ec51F773OL + rm /tmp/tmp.TNP992xf3F /tmp/tmp.ec51F773OL + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.CxXvpBzLn2 ++ mktemp + local LAST_ERR=/tmp/tmp.P5D8gHt1Cb + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.CxXvpBzLn2 + cat /tmp/tmp.P5D8gHt1Cb + rm /tmp/tmp.CxXvpBzLn2 /tmp/tmp.P5D8gHt1Cb + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.M5ySL7hskm ++ mktemp + local LAST_ERR=/tmp/tmp.EYdjAX9kgh + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.M5ySL7hskm + cat /tmp/tmp.EYdjAX9kgh + rm /tmp/tmp.M5ySL7hskm /tmp/tmp.EYdjAX9kgh + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.ig4EIMe6nB ++ mktemp + local LAST_ERR=/tmp/tmp.Ca9S5U1e3c + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ig4EIMe6nB + cat /tmp/tmp.Ca9S5U1e3c + rm /tmp/tmp.ig4EIMe6nB /tmp/tmp.Ca9S5U1e3c + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.E2EO2eeGHO ++ mktemp + local LAST_ERR=/tmp/tmp.wiCoMXqhkM + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.E2EO2eeGHO + cat /tmp/tmp.wiCoMXqhkM + rm /tmp/tmp.E2EO2eeGHO /tmp/tmp.wiCoMXqhkM + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.FcibBhjk4X ++ mktemp + local LAST_ERR=/tmp/tmp.vvOUem8eqm + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.FcibBhjk4X + cat /tmp/tmp.vvOUem8eqm + rm /tmp/tmp.FcibBhjk4X /tmp/tmp.vvOUem8eqm + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.e8HantqZA3 ++ mktemp + local LAST_ERR=/tmp/tmp.X5ArwWRiki + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.e8HantqZA3 + cat /tmp/tmp.X5ArwWRiki + rm /tmp/tmp.e8HantqZA3 /tmp/tmp.X5ArwWRiki + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.cXrYc4D50L ++ mktemp + local LAST_ERR=/tmp/tmp.QxBWTkbSt0 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.cXrYc4D50L + cat /tmp/tmp.QxBWTkbSt0 + rm /tmp/tmp.cXrYc4D50L /tmp/tmp.QxBWTkbSt0 + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.AD4xokjCFe ++ mktemp + local LAST_ERR=/tmp/tmp.35zdWBvtet + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.AD4xokjCFe + cat /tmp/tmp.35zdWBvtet + rm /tmp/tmp.AD4xokjCFe /tmp/tmp.35zdWBvtet + return 0 + echo + echo 'Deleting crashing pod some-name-rs0-2' Deleting crashing pod some-name-rs0-2 + kubectl_bin delete pod some-name-rs0-2 ++ mktemp + local LAST_OUT=/tmp/tmp.MqjWydZKOZ ++ mktemp + local LAST_ERR=/tmp/tmp.7BKbl78Q1j + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete pod some-name-rs0-2 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.MqjWydZKOZ pod "some-name-rs0-2" deleted from demand-backup-if-unhealthy-13462 namespace + cat /tmp/tmp.7BKbl78Q1j + rm /tmp/tmp.MqjWydZKOZ /tmp/tmp.7BKbl78Q1j + return 0 + wait_for_cluster_state some-name ready + local cluster_name=some-name + local target_state=ready + echo -n 'Waiting for psmdb/some-name to reach ready state' Waiting for psmdb/some-name to reach ready state+ local timeout=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0RPKoTKn8e +++ mktemp ++ local LAST_ERR=/tmp/tmp.52F2tjPrwN ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.0RPKoTKn8e ++ cat /tmp/tmp.52F2tjPrwN ++ rm /tmp/tmp.0RPKoTKn8e /tmp/tmp.52F2tjPrwN ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=1 + echo -n . .+ [[ 1 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HC4v7F4lMW +++ mktemp ++ local LAST_ERR=/tmp/tmp.nBh3xs0wtm ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.HC4v7F4lMW ++ cat /tmp/tmp.nBh3xs0wtm ++ rm /tmp/tmp.HC4v7F4lMW /tmp/tmp.nBh3xs0wtm ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=2 + echo -n . .+ [[ 2 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hFoXBrB7jC +++ mktemp ++ local LAST_ERR=/tmp/tmp.pbBlLt33JK ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.hFoXBrB7jC ++ cat /tmp/tmp.pbBlLt33JK ++ rm /tmp/tmp.hFoXBrB7jC /tmp/tmp.pbBlLt33JK ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=3 + echo -n . .+ [[ 3 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1X4jH5iSAH +++ mktemp ++ local LAST_ERR=/tmp/tmp.p2BlOzy2C3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.1X4jH5iSAH ++ cat /tmp/tmp.p2BlOzy2C3 ++ rm /tmp/tmp.1X4jH5iSAH /tmp/tmp.p2BlOzy2C3 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=4 + echo -n . .+ [[ 4 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wiWWaZ8Toa +++ mktemp ++ local LAST_ERR=/tmp/tmp.wozloa7k9n ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.wiWWaZ8Toa ++ cat /tmp/tmp.wozloa7k9n ++ rm /tmp/tmp.wiWWaZ8Toa /tmp/tmp.wozloa7k9n ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=5 + echo -n . .+ [[ 5 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9LdguRrFET +++ mktemp ++ local LAST_ERR=/tmp/tmp.GuA0iyp9E3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9LdguRrFET ++ cat /tmp/tmp.GuA0iyp9E3 ++ rm /tmp/tmp.9LdguRrFET /tmp/tmp.GuA0iyp9E3 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=6 + echo -n . .+ [[ 6 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2uKTOonVAl +++ mktemp ++ local LAST_ERR=/tmp/tmp.2OK74CiLh4 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.2uKTOonVAl ++ cat /tmp/tmp.2OK74CiLh4 ++ rm /tmp/tmp.2uKTOonVAl /tmp/tmp.2OK74CiLh4 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=7 + echo -n . .+ [[ 7 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Gp90X33Yqw +++ mktemp ++ local LAST_ERR=/tmp/tmp.IeH4t6ksHu ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Gp90X33Yqw ++ cat /tmp/tmp.IeH4t6ksHu ++ rm /tmp/tmp.Gp90X33Yqw /tmp/tmp.IeH4t6ksHu ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=8 + echo -n . .+ [[ 8 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Zsfvfq2Nq9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.yYaQbxjhcl ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Zsfvfq2Nq9 ++ cat /tmp/tmp.yYaQbxjhcl ++ rm /tmp/tmp.Zsfvfq2Nq9 /tmp/tmp.yYaQbxjhcl ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=9 + echo -n . .+ [[ 9 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sbWkWbTxvl +++ mktemp ++ local LAST_ERR=/tmp/tmp.VvDwm8b8Tr ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.sbWkWbTxvl ++ cat /tmp/tmp.VvDwm8b8Tr ++ rm /tmp/tmp.sbWkWbTxvl /tmp/tmp.VvDwm8b8Tr ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=10 + echo -n . .+ [[ 10 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LBV1PaR24G +++ mktemp ++ local LAST_ERR=/tmp/tmp.T3anCgMtnB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.LBV1PaR24G ++ cat /tmp/tmp.T3anCgMtnB ++ rm /tmp/tmp.LBV1PaR24G /tmp/tmp.T3anCgMtnB ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=11 + echo -n . .+ [[ 11 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.10l8hlTIWQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.h7Z3fKXZsm ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.10l8hlTIWQ ++ cat /tmp/tmp.h7Z3fKXZsm ++ rm /tmp/tmp.10l8hlTIWQ /tmp/tmp.h7Z3fKXZsm ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=12 + echo -n . .+ [[ 12 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zVRpYIEmX2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.HHMrKizSB7 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.zVRpYIEmX2 ++ cat /tmp/tmp.HHMrKizSB7 ++ rm /tmp/tmp.zVRpYIEmX2 /tmp/tmp.HHMrKizSB7 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=13 + echo -n . .+ [[ 13 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cRUh5rtdhe +++ mktemp ++ local LAST_ERR=/tmp/tmp.sdUB0gm3LH ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.cRUh5rtdhe ++ cat /tmp/tmp.sdUB0gm3LH ++ rm /tmp/tmp.cRUh5rtdhe /tmp/tmp.sdUB0gm3LH ++ return 0 + [[ ready =~ ready ]] + echo + log 'psmdb/some-name is ready: OK' + set +o xtrace [2025-12-22T07:47:47+0000] psmdb/some-name is ready: OK + desc 'Delete data and run restore of backup-minio-logical' + set +o xtrace ----------------------------------------------------------------------------------- Delete data and run restore of backup-minio-logical ----------------------------------------------------------------------------------- + delete_data 'This is some data for backup if unhealthy' + local 'data=This is some data for backup if unhealthy' + run_mongo 'use myApp\n db.test.deleteOne({ x: "This is some data for backup if unhealthy" })' myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-13462 + local 'command=use myApp\n db.test.deleteOne({ x: "This is some data for backup if unhealthy" })' + local uri=myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-13462 + local driver=mongodb+srv + local suffix=.svc.cluster.local + local mongo_flag= + local replica_set=rs0 + [[ myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-13462 == *cfg* ]] ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mTNiVUoHtR +++ mktemp ++ local LAST_ERR=/tmp/tmp.hD3lDattzR ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.mTNiVUoHtR ++ cat /tmp/tmp.hD3lDattzR ++ rm /tmp/tmp.mTNiVUoHtR /tmp/tmp.hD3lDattzR ++ return 0 + local client_container=psmdb-client-696897d69b-sc75l + kubectl_bin exec psmdb-client-696897d69b-sc75l -- bash -c 'printf '\''use myApp\n db.test.deleteOne({ x: "This is some data for backup if unhealthy" })\n'\'' | mongo mongodb+srv://myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-13462.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.ZE8cFfuWxj ++ mktemp + local LAST_ERR=/tmp/tmp.ptoEBZMCIW + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-696897d69b-sc75l -- bash -c 'printf '\''use myApp\n db.test.deleteOne({ x: "This is some data for backup if unhealthy" })\n'\'' | mongo mongodb+srv://myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-13462.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ZE8cFfuWxj Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-rs0-1.some-name-rs0.demand-backup-if-unhealthy-13462.svc.cluster.local:27017,some-name-rs0-0.some-name-rs0.demand-backup-if-unhealthy-13462.svc.cluster.local:27017,some-name-rs0-2.some-name-rs0.demand-backup-if-unhealthy-13462.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb&replicaSet=rs0&ssl=false Implicit session: session { "id" : UUID("fe70aa52-f7a5-4960-8668-59fcb7551618") } Percona Server for MongoDB server version: v8.0.16-5 WARNING: shell and server versions do not match switched to db myApp { "acknowledged" : true, "deletedCount" : 1 } bye + cat /tmp/tmp.ptoEBZMCIW + rm /tmp/tmp.ZE8cFfuWxj /tmp/tmp.ptoEBZMCIW + return 0 + run_restore backup-minio-logical + local backup_name=backup-minio-logical + log 'running restore restore-backup-minio-logical' + set +o xtrace [2025-12-22T07:47:49+0000] running restore restore-backup-minio-logical + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/e2e-tests/demand-backup-if-unhealthy/conf/restore.yml + /usr/sbin/sed -e 's/name:/name: restore-backup-minio-logical/' + /usr/sbin/sed -e 's/backupName:/backupName: backup-minio-logical/' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.5G1Vy9MypF ++ mktemp + local LAST_ERR=/tmp/tmp.AXnHNz9mmn + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.5G1Vy9MypF perconaservermongodbrestore.psmdb.percona.com/restore-backup-minio-logical created + cat /tmp/tmp.AXnHNz9mmn + rm /tmp/tmp.5G1Vy9MypF /tmp/tmp.AXnHNz9mmn + return 0 + wait_restore backup-minio-logical some-name + local backup_name=backup-minio-logical + local cluster_name=some-name + local target_state=ready + local wait_cluster_consistency=1 + local wait_time=1780 + local ok_if_ready=0 + set +o xtrace Waiting for the psmdb-restore/restore-backup-minio-logical object to be created.OK Waiting psmdb-restore/restore-backup-minio-logical to reach state "ready" .OK after 0 minutes + [[ 1 -eq 1 ]] + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BooBX3EgW4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.EGPXrKwnca ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.BooBX3EgW4 ++ cat /tmp/tmp.EGPXrKwnca ++ rm /tmp/tmp.BooBX3EgW4 /tmp/tmp.EGPXrKwnca ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo .OK .OK + wait_for_cluster_state some-name ready + local cluster_name=some-name + local target_state=ready + echo -n 'Waiting for psmdb/some-name to reach ready state' Waiting for psmdb/some-name to reach ready state+ local timeout=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AgArgeVvDI +++ mktemp ++ local LAST_ERR=/tmp/tmp.XVjCTFOlMJ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.AgArgeVvDI ++ cat /tmp/tmp.XVjCTFOlMJ ++ rm /tmp/tmp.AgArgeVvDI /tmp/tmp.XVjCTFOlMJ ++ return 0 + [[ ready =~ ready ]] + echo + log 'psmdb/some-name is ready: OK' + set +o xtrace [2025-12-22T07:48:16+0000] psmdb/some-name is ready: OK + desc 'Verify data is back to the cluster' + set +o xtrace ----------------------------------------------------------------------------------- Verify data is back to the cluster ----------------------------------------------------------------------------------- + simple_data_check some-name 3 0 -rs0 + local cluster_name=some-name + let last_pod=3-1 + local isSharded=0 + local cluster_pfx=-rs0 + '[' 0 -eq 1 ']' ++ seq 0 2 + for i in $(seq 0 $last_pod) + compare_mongo_cmd find myApp:myPass@some-name-rs0-0.some-name-rs0.demand-backup-if-unhealthy-13462 + local command=find + local uri=myApp:myPass@some-name-rs0-0.some-name-rs0.demand-backup-if-unhealthy-13462 + local postfix= + local suffix= + local database=myApp + local collection=test + local sort= + local tls=false + local replicaset= + local 'full_command=db.test.find()' + [[ -n '' ]] + log 'running db.test.find() in myApp' + set +o xtrace [2025-12-22T07:48:16+0000] running db.test.find() in myApp + [[ false == \t\r\u\e ]] + mongo_command=run_mongo + run_mongo 'use myApp\n db.test.find()' myApp:myPass@some-name-rs0-0.some-name-rs0.demand-backup-if-unhealthy-13462 mongodb '' '' + local 'command=use myApp\n db.test.find()' + local uri=myApp:myPass@some-name-rs0-0.some-name-rs0.demand-backup-if-unhealthy-13462 + local driver=mongodb + local suffix=.svc.cluster.local + local mongo_flag= + local replica_set=rs0 + [[ myApp:myPass@some-name-rs0-0.some-name-rs0.demand-backup-if-unhealthy-13462 == *cfg* ]] + egrep -v 'I NETWORK|W NETWORK|F NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:|Started a new thread for the timer service' + /usr/sbin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp egrep: warning: egrep is obsolescent; using grep -E ++ local LAST_OUT=/tmp/tmp.rW1o67M9qG +++ mktemp ++ local LAST_ERR=/tmp/tmp.dmC2pPu5jq ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.rW1o67M9qG ++ cat /tmp/tmp.dmC2pPu5jq ++ rm /tmp/tmp.rW1o67M9qG /tmp/tmp.dmC2pPu5jq ++ return 0 + local client_container=psmdb-client-696897d69b-sc75l + kubectl_bin exec psmdb-client-696897d69b-sc75l -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-rs0-0.some-name-rs0.demand-backup-if-unhealthy-13462.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.V4zNmFms3l ++ mktemp + local LAST_ERR=/tmp/tmp.NiT8jAHhBZ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-696897d69b-sc75l -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-rs0-0.some-name-rs0.demand-backup-if-unhealthy-13462.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.V4zNmFms3l + cat /tmp/tmp.NiT8jAHhBZ + rm /tmp/tmp.V4zNmFms3l /tmp/tmp.NiT8jAHhBZ + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/e2e-tests/demand-backup-if-unhealthy/compare/find.json /tmp/tmp.5ZphUZDmGh/find + for i in $(seq 0 $last_pod) + compare_mongo_cmd find myApp:myPass@some-name-rs0-1.some-name-rs0.demand-backup-if-unhealthy-13462 + local command=find + local uri=myApp:myPass@some-name-rs0-1.some-name-rs0.demand-backup-if-unhealthy-13462 + local postfix= + local suffix= + local database=myApp + local collection=test + local sort= + local tls=false + local replicaset= + local 'full_command=db.test.find()' + [[ -n '' ]] + log 'running db.test.find() in myApp' + set +o xtrace [2025-12-22T07:48:19+0000] running db.test.find() in myApp + [[ false == \t\r\u\e ]] + mongo_command=run_mongo + run_mongo 'use myApp\n db.test.find()' myApp:myPass@some-name-rs0-1.some-name-rs0.demand-backup-if-unhealthy-13462 mongodb '' '' + local 'command=use myApp\n db.test.find()' + local uri=myApp:myPass@some-name-rs0-1.some-name-rs0.demand-backup-if-unhealthy-13462 + local driver=mongodb + local suffix=.svc.cluster.local + local mongo_flag= + egrep -v 'I NETWORK|W NETWORK|F NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:|Started a new thread for the timer service' + local replica_set=rs0 + [[ myApp:myPass@some-name-rs0-1.some-name-rs0.demand-backup-if-unhealthy-13462 == *cfg* ]] + /usr/sbin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp egrep: warning: egrep is obsolescent; using grep -E ++ local LAST_OUT=/tmp/tmp.4GJeCAIxME +++ mktemp ++ local LAST_ERR=/tmp/tmp.zZtKHeXKF8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4GJeCAIxME ++ cat /tmp/tmp.zZtKHeXKF8 ++ rm /tmp/tmp.4GJeCAIxME /tmp/tmp.zZtKHeXKF8 ++ return 0 + local client_container=psmdb-client-696897d69b-sc75l + kubectl_bin exec psmdb-client-696897d69b-sc75l -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-rs0-1.some-name-rs0.demand-backup-if-unhealthy-13462.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.8qUcQxmWP2 ++ mktemp + local LAST_ERR=/tmp/tmp.a2P2apO2Mu + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-696897d69b-sc75l -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-rs0-1.some-name-rs0.demand-backup-if-unhealthy-13462.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.8qUcQxmWP2 + cat /tmp/tmp.a2P2apO2Mu + rm /tmp/tmp.8qUcQxmWP2 /tmp/tmp.a2P2apO2Mu + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/e2e-tests/demand-backup-if-unhealthy/compare/find.json /tmp/tmp.5ZphUZDmGh/find + for i in $(seq 0 $last_pod) + compare_mongo_cmd find myApp:myPass@some-name-rs0-2.some-name-rs0.demand-backup-if-unhealthy-13462 + local command=find + local uri=myApp:myPass@some-name-rs0-2.some-name-rs0.demand-backup-if-unhealthy-13462 + local postfix= + local suffix= + local database=myApp + local collection=test + local sort= + local tls=false + local replicaset= + local 'full_command=db.test.find()' + [[ -n '' ]] + log 'running db.test.find() in myApp' + set +o xtrace [2025-12-22T07:48:21+0000] running db.test.find() in myApp + [[ false == \t\r\u\e ]] + mongo_command=run_mongo + egrep -v 'I NETWORK|W NETWORK|F NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:|Started a new thread for the timer service' + run_mongo 'use myApp\n db.test.find()' myApp:myPass@some-name-rs0-2.some-name-rs0.demand-backup-if-unhealthy-13462 mongodb '' '' + /usr/sbin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' + local 'command=use myApp\n db.test.find()' + local uri=myApp:myPass@some-name-rs0-2.some-name-rs0.demand-backup-if-unhealthy-13462 + local driver=mongodb + local suffix=.svc.cluster.local + local mongo_flag= + local replica_set=rs0 + [[ myApp:myPass@some-name-rs0-2.some-name-rs0.demand-backup-if-unhealthy-13462 == *cfg* ]] ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' egrep: warning: egrep is obsolescent; using grep -E +++ mktemp ++ local LAST_OUT=/tmp/tmp.fQOjvKNq9o +++ mktemp ++ local LAST_ERR=/tmp/tmp.lnSCkkGjxu ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.fQOjvKNq9o ++ cat /tmp/tmp.lnSCkkGjxu ++ rm /tmp/tmp.fQOjvKNq9o /tmp/tmp.lnSCkkGjxu ++ return 0 + local client_container=psmdb-client-696897d69b-sc75l + kubectl_bin exec psmdb-client-696897d69b-sc75l -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-rs0-2.some-name-rs0.demand-backup-if-unhealthy-13462.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.6ob1Ni62sE ++ mktemp + local LAST_ERR=/tmp/tmp.BYFpCbJwgm + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-696897d69b-sc75l -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-rs0-2.some-name-rs0.demand-backup-if-unhealthy-13462.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.6ob1Ni62sE + cat /tmp/tmp.BYFpCbJwgm + rm /tmp/tmp.6ob1Ni62sE /tmp/tmp.BYFpCbJwgm + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/e2e-tests/demand-backup-if-unhealthy/compare/find.json /tmp/tmp.5ZphUZDmGh/find + destroy demand-backup-if-unhealthy-13462 + local namespace=demand-backup-if-unhealthy-13462 + local ignore_logs=true + [[ 0 == 1 ]] + desc 'destroy cluster/operator and all other resources' + set +o xtrace ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + '[' true == false ']' + delete_backups + desc 'Delete psmdb-backup' + set +o xtrace ----------------------------------------------------------------------------------- Delete psmdb-backup ----------------------------------------------------------------------------------- ++ wc -l ++ kubectl_bin get psmdb-backup --no-headers +++ mktemp ++ local LAST_OUT=/tmp/tmp.iKsarhoRA2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.siATBpORDo ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb-backup --no-headers ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.iKsarhoRA2 ++ cat /tmp/tmp.siATBpORDo ++ rm /tmp/tmp.iKsarhoRA2 /tmp/tmp.siATBpORDo ++ return 0 + '[' 1 '!=' 0 ']' + kubectl_bin get psmdb-backup ++ mktemp + local LAST_OUT=/tmp/tmp.4xxDZR8RYV ++ mktemp + local LAST_ERR=/tmp/tmp.Lib430jL4s + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get psmdb-backup + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.4xxDZR8RYV NAME CLUSTER STORAGE DESTINATION TYPE SIZE STATUS COMPLETED AGE backup-minio-logical some-name minio s3://operator-testing/2025-12-22T07:46:27Z logical 53.62KB ready 104s 119s + cat /tmp/tmp.Lib430jL4s + rm /tmp/tmp.4xxDZR8RYV /tmp/tmp.Lib430jL4s + return 0 + kubectl_bin delete psmdb-backup --all ++ mktemp + local LAST_OUT=/tmp/tmp.A4qZZoYW1P ++ mktemp + local LAST_ERR=/tmp/tmp.wU6nC0BJ56 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete psmdb-backup --all + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.A4qZZoYW1P perconaservermongodbbackup.psmdb.percona.com "backup-minio-logical" deleted from demand-backup-if-unhealthy-13462 namespace + cat /tmp/tmp.wU6nC0BJ56 + rm /tmp/tmp.A4qZZoYW1P /tmp/tmp.wU6nC0BJ56 + return 0 + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.mTISMAkvTV ++ mktemp + local LAST_ERR=/tmp/tmp.LO4nMB8Gbo + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.mTISMAkvTV customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.LO4nMB8Gbo + rm /tmp/tmp.mTISMAkvTV /tmp/tmp.LO4nMB8Gbo + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/deploy/crd.yaml ++ grep -v '\-\-\-' grep: warning: stray \ before - grep: warning: stray \ before - + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.yJRzZpYvnE ++ mktemp + local LAST_ERR=/tmp/tmp.5SpSph5hmq + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.yJRzZpYvnE + cat /tmp/tmp.5SpSph5hmq + rm /tmp/tmp.yJRzZpYvnE /tmp/tmp.5SpSph5hmq + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.DeRUbkPY0B ++ mktemp + local LAST_ERR=/tmp/tmp.8TWgthRSbe + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.DeRUbkPY0B + cat /tmp/tmp.8TWgthRSbe + rm /tmp/tmp.DeRUbkPY0B /tmp/tmp.8TWgthRSbe + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' No resources found + kubectl patch perconaservermongodbs.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbs" + : + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.kDkYWwFz50 ++ mktemp + local LAST_ERR=/tmp/tmp.XSOUQDGJN2 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.kDkYWwFz50 + cat /tmp/tmp.XSOUQDGJN2 + rm /tmp/tmp.kDkYWwFz50 /tmp/tmp.XSOUQDGJN2 + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.MxozpKLX53 ++ mktemp + local LAST_ERR=/tmp/tmp.OS4Y0uiIpd + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2159/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.MxozpKLX53 clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.OS4Y0uiIpd + rm /tmp/tmp.MxozpKLX53 /tmp/tmp.OS4Y0uiIpd + return 0 + destroy_cert_manager + kubectl_bin delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.398Ni0nXD2 ++ mktemp + local LAST_ERR=/tmp/tmp.74ySUdp3q3 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.398Ni0nXD2 + cat /tmp/tmp.74ySUdp3q3 Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 0 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.398Ni0nXD2 + cat /tmp/tmp.74ySUdp3q3 Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 4 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.398Ni0nXD2 + cat /tmp/tmp.74ySUdp3q3 Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 8 + cat /tmp/tmp.398Ni0nXD2 + cat /tmp/tmp.74ySUdp3q3 Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + rm /tmp/tmp.398Ni0nXD2 /tmp/tmp.74ySUdp3q3 + return 1 + true + '[' -n '' ']' + '[' -n psmdb-operator ']' + kubectl_bin delete --grace-period=0 --force=true namespace demand-backup-if-unhealthy-13462 + rm -rf /tmp/tmp.5ZphUZDmGh + kubectl_bin delete --grace-period=0 --force=true namespace psmdb-operator ++ mktemp ++ mktemp + desc 'test passed' + set +o xtrace ----------------------------------------------------------------------------------- test passed ----------------------------------------------------------------------------------- + local LAST_OUT=/tmp/tmp.iyIqtr0w19 ++ mktemp + local LAST_OUT=/tmp/tmp.U8Jn6NnCYZ ++ mktemp + local LAST_ERR=/tmp/tmp.VfM3K4RqEa + local exit_status=0 + local timeout=4 + local LAST_ERR=/tmp/tmp.ZS8OQeHI2h + local exit_status=0 + local timeout=4 ++ seq 0 2 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete --grace-period=0 --force=true namespace demand-backup-if-unhealthy-13462 + for i in $(seq 0 2) + set +e + kubectl delete --grace-period=0 --force=true namespace psmdb-operator