Log: /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/logs/demand-backup-if-unhealthy.log Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 + create_infra demand-backup-if-unhealthy-355 + local ns=demand-backup-if-unhealthy-355 + [[ 1 == 1 ]] + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.1O1RD5Qcys ++ mktemp + local LAST_ERR=/tmp/tmp.35EnHW5YEX + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.1O1RD5Qcys customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.35EnHW5YEX + rm /tmp/tmp.1O1RD5Qcys /tmp/tmp.35EnHW5YEX + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/crd.yaml ++ grep -v '\-\-\-' grep: warning: stray \ before - grep: warning: stray \ before - + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.1SnKfDo7yS ++ mktemp + local LAST_ERR=/tmp/tmp.oLNTybdfPr + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.1SnKfDo7yS + cat /tmp/tmp.oLNTybdfPr + rm /tmp/tmp.1SnKfDo7yS /tmp/tmp.oLNTybdfPr + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.61oCHNHo8h ++ mktemp + local LAST_ERR=/tmp/tmp.lkjmlhdbii + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.61oCHNHo8h + cat /tmp/tmp.lkjmlhdbii + rm /tmp/tmp.61oCHNHo8h /tmp/tmp.lkjmlhdbii + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbs" + kubectl patch perconaservermongodbs.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbs" + : + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.BDnwej7S0e ++ mktemp + local LAST_ERR=/tmp/tmp.wzvvLz9wwr + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.BDnwej7S0e + cat /tmp/tmp.wzvvLz9wwr + rm /tmp/tmp.BDnwej7S0e /tmp/tmp.wzvvLz9wwr + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.rTNkdklFyq ++ mktemp + local LAST_ERR=/tmp/tmp.ZrDtO1BaYR + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.rTNkdklFyq clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.ZrDtO1BaYR + rm /tmp/tmp.rTNkdklFyq /tmp/tmp.ZrDtO1BaYR + return 0 + check_crd_for_deletion PR-2266-f9fc55604 + local git_tag=PR-2266-f9fc55604 ++ curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/PR-2266-f9fc55604/deploy/crd.yaml ++ yq eval .metadata.name ++ /usr/sbin/sed s/---//g ++ /usr/sbin/sed ':a;N;$!ba;s/\n/ /g' + for crd_name in $(curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/${git_tag}/deploy/crd.yaml | yq eval '.metadata.name' | $sed 's/---//g' | $sed ':a;N;$!ba;s/\n/ /g') ++ kubectl_bin get crd/null -o 'jsonpath={.status.conditions[-1].type}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xqoB2reb5v +++ mktemp ++ local LAST_ERR=/tmp/tmp.duUVo5addN ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.xqoB2reb5v ++ cat /tmp/tmp.duUVo5addN Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 0 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.xqoB2reb5v ++ cat /tmp/tmp.duUVo5addN Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 4 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.xqoB2reb5v ++ cat /tmp/tmp.duUVo5addN Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 8 ++ cat /tmp/tmp.xqoB2reb5v ++ cat /tmp/tmp.duUVo5addN Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ rm /tmp/tmp.xqoB2reb5v /tmp/tmp.duUVo5addN ++ return 1 + [[ '' == \T\e\r\m\i\n\a\t\i\n\g ]] + '[' -n psmdb-operator ']' + create_namespace psmdb-operator + local namespace=psmdb-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ awk '{print $1}' ++ grep chaos-mesh.org + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + grep -E -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' + awk '{print$1}' + '[' -n '' ']' + desc 'cleaned up old namespaces psmdb-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces psmdb-operator++ mktemp ----------------------------------------------------------------------------------- + kubectl_bin delete namespace psmdb-operator --ignore-not-found + xargs kubectl delete ns ++ mktemp + local LAST_OUT=/tmp/tmp.H2UWbmQ6Ia ++ mktemp + local LAST_OUT=/tmp/tmp.L94qlQ6cG6 ++ mktemp + local LAST_ERR=/tmp/tmp.3bDKtTJfyR + local exit_status=0 + local timeout=4 ++ seq 0 2 + local LAST_ERR=/tmp/tmp.fSf5CQrC0m + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get ns + for i in $(seq 0 2) + set +e + kubectl delete namespace psmdb-operator --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.H2UWbmQ6Ia + cat /tmp/tmp.3bDKtTJfyR + rm /tmp/tmp.H2UWbmQ6Ia /tmp/tmp.3bDKtTJfyR + return 0 namespace "demand-backup-if-unhealthy-3484" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.L94qlQ6cG6 namespace "psmdb-operator" deleted + cat /tmp/tmp.fSf5CQrC0m + rm /tmp/tmp.L94qlQ6cG6 /tmp/tmp.fSf5CQrC0m + return 0 + kubectl_bin wait --for=delete namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.KSo4GUKC9y ++ mktemp + local LAST_ERR=/tmp/tmp.1rkorfex9j + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.KSo4GUKC9y + cat /tmp/tmp.1rkorfex9j + rm /tmp/tmp.KSo4GUKC9y /tmp/tmp.1rkorfex9j + return 0 + desc 'create namespace psmdb-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.fBza85MC4F ++ mktemp + local LAST_ERR=/tmp/tmp.IP6qUr7JVn + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.fBza85MC4F namespace/psmdb-operator created + cat /tmp/tmp.IP6qUr7JVn + rm /tmp/tmp.fBza85MC4F /tmp/tmp.IP6qUr7JVn + return 0 + set_kube_ctx psmdb-operator + local namespace=psmdb-operator ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.nZWGSI3aWG +++ mktemp ++ local LAST_ERR=/tmp/tmp.PGQc8d46c8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.nZWGSI3aWG ++ cat /tmp/tmp.PGQc8d46c8 ++ rm /tmp/tmp.nZWGSI3aWG /tmp/tmp.PGQc8d46c8 ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2266-f9fc55604-9-cluster14 --namespace=psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.G1drExz38X ++ mktemp + local LAST_ERR=/tmp/tmp.eGU31kkB05 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2266-f9fc55604-9-cluster14 --namespace=psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.G1drExz38X Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2266-f9fc55604-9-cluster14" modified. + cat /tmp/tmp.eGU31kkB05 + rm /tmp/tmp.G1drExz38X /tmp/tmp.eGU31kkB05 + return 0 + deploy_operator + desc 'start PSMDB operator: docker.io/perconalab/percona-server-mongodb-operator:PR-2266-f9fc55604' + set +o xtrace ----------------------------------------------------------------------------------- start PSMDB operator: docker.io/perconalab/percona-server-mongodb-operator:PR-2266-f9fc55604 ----------------------------------------------------------------------------------- + local cr_file + '[' -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/demand-backup-if-unhealthy/conf/crd.yaml ']' + cr_file=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/crd.yaml + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.1GM1laBtUP ++ mktemp + local LAST_ERR=/tmp/tmp.1jw6172FFG + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.1GM1laBtUP customresourcedefinition.apiextensions.k8s.io/perconaservermongodbbackups.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbrestores.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com serverside-applied + cat /tmp/tmp.1jw6172FFG + rm /tmp/tmp.1GM1laBtUP /tmp/tmp.1jw6172FFG + return 0 + '[' -n psmdb-operator ']' + apply_rbac cw-rbac + local operator_namespace=psmdb-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: psmdb-operator^' + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.RA3FArk4sr ++ mktemp + local LAST_ERR=/tmp/tmp.z5YZDaX6SH + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.RA3FArk4sr clusterrole.rbac.authorization.k8s.io/percona-server-mongodb-operator created serviceaccount/percona-server-mongodb-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-server-mongodb-operator created + cat /tmp/tmp.z5YZDaX6SH + rm /tmp/tmp.RA3FArk4sr /tmp/tmp.z5YZDaX6SH + return 0 + yq eval ' (.spec.template.spec.containers[].image = "docker.io/perconalab/percona-server-mongodb-operator:PR-2266-f9fc55604") | ((.. | select(.[] == "DISABLE_TELEMETRY")) |= .value="true") | ((.. | select(.[] == "LOG_LEVEL")) |= .value="DEBUG")' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/cw-operator.yaml + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.qSCtsW4cty ++ mktemp + local LAST_ERR=/tmp/tmp.xWaZaJ6ElJ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.qSCtsW4cty deployment.apps/percona-server-mongodb-operator created + cat /tmp/tmp.xWaZaJ6ElJ + rm /tmp/tmp.qSCtsW4cty /tmp/tmp.xWaZaJ6ElJ + return 0 + sleep 20 ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.hmrIuGqDnM +++ mktemp ++ local LAST_ERR=/tmp/tmp.kuE2jjxSll ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.hmrIuGqDnM ++ cat /tmp/tmp.kuE2jjxSll ++ rm /tmp/tmp.hmrIuGqDnM /tmp/tmp.kuE2jjxSll ++ return 0 + wait_operator_pod percona-server-mongodb-operator-5554c4df68-l857v + local pod=percona-server-mongodb-operator-5554c4df68-l857v + set +o xtrace waiting for pod/percona-server-mongodb-operator-5554c4df68-l857v to be ready.OK + echo 'Print operator info from log' Print operator info from log + grep 'Manager starting up' ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.yoROUcSAh8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.4yIU9jpjlj ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.yoROUcSAh8 ++ cat /tmp/tmp.4yIU9jpjlj ++ rm /tmp/tmp.yoROUcSAh8 /tmp/tmp.4yIU9jpjlj ++ return 0 + kubectl_bin logs -n psmdb-operator percona-server-mongodb-operator-5554c4df68-l857v ++ mktemp + local LAST_OUT=/tmp/tmp.uAa3ig8nDU ++ mktemp + local LAST_ERR=/tmp/tmp.IkAnO1rL6V + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl logs -n psmdb-operator percona-server-mongodb-operator-5554c4df68-l857v + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.uAa3ig8nDU + cat /tmp/tmp.IkAnO1rL6V + rm /tmp/tmp.uAa3ig8nDU /tmp/tmp.IkAnO1rL6V + return 0 2026-03-10T06:51:39.595Z INFO setup Manager starting up {"gitCommit": "f9fc5560468cd2df79998ec06ba830d1104a31a3", "gitBranch": "PR-2266-f9fc55604", "buildTime": "", "goVersion": "go1.25.8", "os": "linux", "arch": "amd64"} + create_namespace demand-backup-if-unhealthy-355 + local namespace=demand-backup-if-unhealthy-355 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ helm list --all-namespaces --filter chaos-mesh ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ kubectl get clusterrole ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + grep -E -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' + awk '{print$1}' + '[' -n '' ']' ++ mktemp + desc 'cleaned up old namespaces demand-backup-if-unhealthy-355' + xargs kubectl delete ns + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces demand-backup-if-unhealthy-355 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace demand-backup-if-unhealthy-355 --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.NrcofRjXCY ++ mktemp + local LAST_OUT=/tmp/tmp.x7vCuyMSsL + local LAST_ERR=/tmp/tmp.7xALDNGa8T + local exit_status=0 + local timeout=4 ++ mktemp ++ seq 0 2 + local LAST_ERR=/tmp/tmp.zUgIMkFsOb + local exit_status=0 + local timeout=4 + for i in $(seq 0 2) + set +e + kubectl get ns ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete namespace demand-backup-if-unhealthy-355 --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.NrcofRjXCY + cat /tmp/tmp.7xALDNGa8T + rm /tmp/tmp.NrcofRjXCY /tmp/tmp.7xALDNGa8T + return 0 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.x7vCuyMSsL + cat /tmp/tmp.zUgIMkFsOb + rm /tmp/tmp.x7vCuyMSsL /tmp/tmp.zUgIMkFsOb + return 0 + kubectl_bin wait --for=delete namespace demand-backup-if-unhealthy-355 ++ mktemp + local LAST_OUT=/tmp/tmp.H4KHdsBDeh ++ mktemp + local LAST_ERR=/tmp/tmp.LVCH1MuRn5 + local exit_status=0 + local timeout=4 ++ seq 0 2 error: resource(s) were provided, but no name was specified + for i in $(seq 0 2) + set +e + kubectl wait --for=delete namespace demand-backup-if-unhealthy-355 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.H4KHdsBDeh + cat /tmp/tmp.LVCH1MuRn5 + rm /tmp/tmp.H4KHdsBDeh /tmp/tmp.LVCH1MuRn5 + return 0 + desc 'create namespace demand-backup-if-unhealthy-355' + set +o xtrace ----------------------------------------------------------------------------------- create namespace demand-backup-if-unhealthy-355 ----------------------------------------------------------------------------------- + kubectl_bin create namespace demand-backup-if-unhealthy-355 ++ mktemp + local LAST_OUT=/tmp/tmp.NSglP8KkJu ++ mktemp + local LAST_ERR=/tmp/tmp.0sOAZdudg6 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace demand-backup-if-unhealthy-355 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.NSglP8KkJu namespace/demand-backup-if-unhealthy-355 created + cat /tmp/tmp.0sOAZdudg6 + rm /tmp/tmp.NSglP8KkJu /tmp/tmp.0sOAZdudg6 + return 0 + set_kube_ctx demand-backup-if-unhealthy-355 + local namespace=demand-backup-if-unhealthy-355 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.1NRxdehZRp +++ mktemp ++ local LAST_ERR=/tmp/tmp.kNBLam21Sz ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.1NRxdehZRp ++ cat /tmp/tmp.kNBLam21Sz ++ rm /tmp/tmp.1NRxdehZRp /tmp/tmp.kNBLam21Sz ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2266-f9fc55604-9-cluster14 --namespace=demand-backup-if-unhealthy-355 ++ mktemp + local LAST_OUT=/tmp/tmp.OCOhoBJiy8 ++ mktemp + local LAST_ERR=/tmp/tmp.4URyhZyuvD + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2266-f9fc55604-9-cluster14 --namespace=demand-backup-if-unhealthy-355 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.OCOhoBJiy8 Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2266-f9fc55604-9-cluster14" modified. + cat /tmp/tmp.4URyhZyuvD + rm /tmp/tmp.OCOhoBJiy8 /tmp/tmp.4URyhZyuvD + return 0 + deploy_minio + local cert_secret= + local service_name=minio-service + desc 'install MinIO: minio-service' + set +o xtrace ----------------------------------------------------------------------------------- install MinIO: minio-service ----------------------------------------------------------------------------------- + helm uninstall minio-service + : + helm repo remove minio "minio" has been removed from your repositories + helm repo add minio https://charts.min.io/ "minio" has been added to your repositories + local endpoint=http://minio-service:9000 + minio_args=('--version' '5.4.0' '--set' 'replicas=1' '--set' 'mode=standalone' '--set' 'resources.requests.memory=256Mi' '--set' 'rootUser=rootuser' '--set' 'rootPassword=rootpass123' '--set' 'users[0].accessKey=some-access-key' '--set' 'users[0].secretKey=some-secret-key' '--set' 'users[0].policy=consoleAdmin' '--set' 'service.type=ClusterIP' '--set' 'configPathmc=/tmp/' '--set' 'securityContext.enabled=false' '--set' 'persistence.size=2G' '--set' 'fullnameOverride=minio-service' '--set' 'serviceAccount.create=true' '--set' 'serviceAccount.name=minio-service-sa') + local minio_args + [[ -n '' ]] + retry 10 60 helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G --set fullnameOverride=minio-service --set serviceAccount.create=true --set serviceAccount.name=minio-service-sa minio/minio + local max=10 + local delay=60 + shift 2 + local n=1 + helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/ --set securityContext.enabled=false --set persistence.size=2G --set fullnameOverride=minio-service --set serviceAccount.create=true --set serviceAccount.name=minio-service-sa minio/minio NAME: minio-service LAST DEPLOYED: Tue Mar 10 06:52:15 2026 NAMESPACE: demand-backup-if-unhealthy-355 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: MinIO can be accessed via port 9000 on the following DNS name from within your cluster: minio-service.demand-backup-if-unhealthy-355.cluster.local To access MinIO from localhost, run the below commands: 1. export POD_NAME=$(kubectl get pods --namespace demand-backup-if-unhealthy-355 -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}") 2. kubectl port-forward $POD_NAME 9000 --namespace demand-backup-if-unhealthy-355 Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/ You can now access MinIO server on http://localhost:9000. Follow the below steps to connect to MinIO server with mc client: 1. Download the MinIO mc client - https://min.io/docs/minio/linux/reference/minio-mc.html#quickstart 2. export MC_HOST_minio-service-local=http://$(kubectl get secret --namespace demand-backup-if-unhealthy-355 minio-service -o jsonpath="{.data.rootUser}" | base64 --decode):$(kubectl get secret --namespace demand-backup-if-unhealthy-355 minio-service -o jsonpath="{.data.rootPassword}" | base64 --decode)@localhost:9000 3. mc ls minio-service-local ++ kubectl_bin get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.p0mCJTPgyT +++ mktemp ++ local LAST_ERR=/tmp/tmp.qyNNgcVAjz ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.p0mCJTPgyT ++ cat /tmp/tmp.qyNNgcVAjz ++ rm /tmp/tmp.p0mCJTPgyT /tmp/tmp.qyNNgcVAjz ++ return 0 + local MINIO_POD=minio-service-6d5f646cdc-jpgzf + wait_pod minio-service-6d5f646cdc-jpgzf + local pod=minio-service-6d5f646cdc-jpgzf + set +o xtrace waiting for pod/minio-service-6d5f646cdc-jpgzf to be ready.OK + '[' -n psmdb-operator ']' + kubectl_bin create svc -n psmdb-operator externalname minio-service --external-name=minio-service.demand-backup-if-unhealthy-355.svc.cluster.local --tcp=9000 service/minio-service created + create_minio_bucket operator-testing http://minio-service:9000 + local bucket=operator-testing + local endpoint=http://minio-service:9000 + kubectl_bin run -i --rm aws-cli --image=docker.io/perconalab/awscli --restart=Never -- bash -c 'AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --no-verify-ssl --endpoint-url http://minio-service:9000 s3 mb s3://operator-testing' ++ mktemp + local LAST_OUT=/tmp/tmp.i7wamOu8EC ++ mktemp + local LAST_ERR=/tmp/tmp.tnQeNOvTM9 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl run -i --rm aws-cli --image=docker.io/perconalab/awscli --restart=Never -- bash -c 'AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --no-verify-ssl --endpoint-url http://minio-service:9000 s3 mb s3://operator-testing' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.i7wamOu8EC pod "aws-cli" deleted from demand-backup-if-unhealthy-355 namespace + cat /tmp/tmp.tnQeNOvTM9 All commands and output from this session will be recorded in container logs, including credentials and sensitive information passed through the command prompt. If you don't see a command prompt, try pressing enter. + rm /tmp/tmp.i7wamOu8EC /tmp/tmp.tnQeNOvTM9 + return 0 + create_secrets + echo 'Creating secrets and start client' Creating secrets and start client + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/demand-backup-if-unhealthy/conf/secrets.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.S6DiZiRuKM ++ mktemp + local LAST_ERR=/tmp/tmp.K2xe2RAU24 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/demand-backup-if-unhealthy/conf/secrets.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/conf/client.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.S6DiZiRuKM secret/some-users created deployment.apps/psmdb-client created + cat /tmp/tmp.K2xe2RAU24 + rm /tmp/tmp.S6DiZiRuKM /tmp/tmp.K2xe2RAU24 + return 0 + apply_s3_storage_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.iSSp6fDR6W ++ mktemp + local LAST_ERR=/tmp/tmp.vSZ560I6VT + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.iSSp6fDR6W secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created secret/gcp-cs-sa-key-secret created + cat /tmp/tmp.vSZ560I6VT + rm /tmp/tmp.iSSp6fDR6W /tmp/tmp.vSZ560I6VT + return 0 + cluster=some-name + desc 'create PSMDB cluster some-name' + set +o xtrace ----------------------------------------------------------------------------------- create PSMDB cluster some-name ----------------------------------------------------------------------------------- + apply_cluster /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/demand-backup-if-unhealthy/conf/some-name.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/demand-backup-if-unhealthy/conf/some-name.yml + kubectl_bin apply -f - + yq eval '(.spec | select(.image == null)).image = "docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0"' + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/demand-backup-if-unhealthy/conf/some-name.yml ++ mktemp + yq eval '.spec.upgradeOptions.apply="Never"' + yq eval '(.spec | select(has("backup"))).backup.image = "docker.io/perconalab/percona-server-mongodb-operator:main-backup"' + /usr/sbin/sed -e s/NAME_SPACE/demand-backup-if-unhealthy-355/g + yq eval '(.spec | select(has("initImage"))).initImage = "docker.io/perconalab/percona-server-mongodb-operator:PR-2266-f9fc55604"' + yq eval '(.spec | select(has("pmm"))).pmm.image = "docker.io/percona/pmm-client:2.44.1-1"' + local LAST_OUT=/tmp/tmp.IVbwbV20qz ++ mktemp + local LAST_ERR=/tmp/tmp.Jbl4xHRp6g + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.IVbwbV20qz perconaservermongodb.psmdb.percona.com/some-name created + cat /tmp/tmp.Jbl4xHRp6g + rm /tmp/tmp.IVbwbV20qz /tmp/tmp.Jbl4xHRp6g + return 0 + wait_for_cluster_state some-name ready + local cluster_name=some-name + local target_state=ready + echo -n 'Waiting for psmdb/some-name to reach ready state' Waiting for psmdb/some-name to reach ready state+ local timeout=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YKqHDE4Qsb +++ mktemp ++ local LAST_ERR=/tmp/tmp.hjjvZoMxit ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.YKqHDE4Qsb ++ cat /tmp/tmp.hjjvZoMxit ++ rm /tmp/tmp.YKqHDE4Qsb /tmp/tmp.hjjvZoMxit ++ return 0 + [[ '' =~ ready ]] + sleep 1 + timeout=1 + echo -n . .+ [[ 1 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2snwysh7G8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.zElGqL5tQq ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.2snwysh7G8 ++ cat /tmp/tmp.zElGqL5tQq ++ rm /tmp/tmp.2snwysh7G8 /tmp/tmp.zElGqL5tQq ++ return 0 + [[ '' =~ ready ]] + sleep 1 + timeout=2 + echo -n . .+ [[ 2 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DW5KG8zAzr +++ mktemp ++ local LAST_ERR=/tmp/tmp.F3NFeaUpts ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.DW5KG8zAzr ++ cat /tmp/tmp.F3NFeaUpts ++ rm /tmp/tmp.DW5KG8zAzr /tmp/tmp.F3NFeaUpts ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=3 + echo -n . .+ [[ 3 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.p3smDGGJoJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.HC3VjtzO4N ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.p3smDGGJoJ ++ cat /tmp/tmp.HC3VjtzO4N ++ rm /tmp/tmp.p3smDGGJoJ /tmp/tmp.HC3VjtzO4N ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=4 + echo -n . .+ [[ 4 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.saFT5nw49y +++ mktemp ++ local LAST_ERR=/tmp/tmp.7wmzgeD0hG ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.saFT5nw49y ++ cat /tmp/tmp.7wmzgeD0hG ++ rm /tmp/tmp.saFT5nw49y /tmp/tmp.7wmzgeD0hG ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=5 + echo -n . .+ [[ 5 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HdrVjvz5R7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.rpfRezc7uy ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.HdrVjvz5R7 ++ cat /tmp/tmp.rpfRezc7uy ++ rm /tmp/tmp.HdrVjvz5R7 /tmp/tmp.rpfRezc7uy ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=6 + echo -n . .+ [[ 6 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EIcHDUNgVw +++ mktemp ++ local LAST_ERR=/tmp/tmp.WVHWvomTLf ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.EIcHDUNgVw ++ cat /tmp/tmp.WVHWvomTLf ++ rm /tmp/tmp.EIcHDUNgVw /tmp/tmp.WVHWvomTLf ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=7 + echo -n . .+ [[ 7 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.10qtPUqJ8g +++ mktemp ++ local LAST_ERR=/tmp/tmp.tHwOWG4her ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.10qtPUqJ8g ++ cat /tmp/tmp.tHwOWG4her ++ rm /tmp/tmp.10qtPUqJ8g /tmp/tmp.tHwOWG4her ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=8 + echo -n . .+ [[ 8 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NIAzQisMYU +++ mktemp ++ local LAST_ERR=/tmp/tmp.YGpcOGNqnI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.NIAzQisMYU ++ cat /tmp/tmp.YGpcOGNqnI ++ rm /tmp/tmp.NIAzQisMYU /tmp/tmp.YGpcOGNqnI ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=9 + echo -n . .+ [[ 9 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3QfhdiWKNk +++ mktemp ++ local LAST_ERR=/tmp/tmp.AveL5J1yAm ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.3QfhdiWKNk ++ cat /tmp/tmp.AveL5J1yAm ++ rm /tmp/tmp.3QfhdiWKNk /tmp/tmp.AveL5J1yAm ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=10 + echo -n . .+ [[ 10 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zoYemW9kjg +++ mktemp ++ local LAST_ERR=/tmp/tmp.1AZftZCTY9 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.zoYemW9kjg ++ cat /tmp/tmp.1AZftZCTY9 ++ rm /tmp/tmp.zoYemW9kjg /tmp/tmp.1AZftZCTY9 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=11 + echo -n . .+ [[ 11 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.CUVacKOEqn +++ mktemp ++ local LAST_ERR=/tmp/tmp.vc1mChjo8k ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.CUVacKOEqn ++ cat /tmp/tmp.vc1mChjo8k ++ rm /tmp/tmp.CUVacKOEqn /tmp/tmp.vc1mChjo8k ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=12 + echo -n . .+ [[ 12 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KGewkZV1o8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.5KqgyKjSKN ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.KGewkZV1o8 ++ cat /tmp/tmp.5KqgyKjSKN ++ rm /tmp/tmp.KGewkZV1o8 /tmp/tmp.5KqgyKjSKN ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=13 + echo -n . .+ [[ 13 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZDtJh4pEGb +++ mktemp ++ local LAST_ERR=/tmp/tmp.Yjccqqu4aH ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ZDtJh4pEGb ++ cat /tmp/tmp.Yjccqqu4aH ++ rm /tmp/tmp.ZDtJh4pEGb /tmp/tmp.Yjccqqu4aH ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=14 + echo -n . .+ [[ 14 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.D5aODEz4UA +++ mktemp ++ local LAST_ERR=/tmp/tmp.WFB5Xm4zWo ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.D5aODEz4UA ++ cat /tmp/tmp.WFB5Xm4zWo ++ rm /tmp/tmp.D5aODEz4UA /tmp/tmp.WFB5Xm4zWo ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=15 + echo -n . .+ [[ 15 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JAjXm8jC3O +++ mktemp ++ local LAST_ERR=/tmp/tmp.7OOf9aIoYE ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.JAjXm8jC3O ++ cat /tmp/tmp.7OOf9aIoYE ++ rm /tmp/tmp.JAjXm8jC3O /tmp/tmp.7OOf9aIoYE ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=16 + echo -n . .+ [[ 16 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AvMceLyOSY +++ mktemp ++ local LAST_ERR=/tmp/tmp.LI0XQWy3DC ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.AvMceLyOSY ++ cat /tmp/tmp.LI0XQWy3DC ++ rm /tmp/tmp.AvMceLyOSY /tmp/tmp.LI0XQWy3DC ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=17 + echo -n . .+ [[ 17 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vTlLVuh6Ch +++ mktemp ++ local LAST_ERR=/tmp/tmp.rSBOF5tWZU ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.vTlLVuh6Ch ++ cat /tmp/tmp.rSBOF5tWZU ++ rm /tmp/tmp.vTlLVuh6Ch /tmp/tmp.rSBOF5tWZU ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=18 + echo -n . .+ [[ 18 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vHZcXgFYM5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.ewQFmEZrkI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.vHZcXgFYM5 ++ cat /tmp/tmp.ewQFmEZrkI ++ rm /tmp/tmp.vHZcXgFYM5 /tmp/tmp.ewQFmEZrkI ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=19 + echo -n . .+ [[ 19 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8BZZMPsbTe +++ mktemp ++ local LAST_ERR=/tmp/tmp.VKElLwYl4P ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.8BZZMPsbTe ++ cat /tmp/tmp.VKElLwYl4P ++ rm /tmp/tmp.8BZZMPsbTe /tmp/tmp.VKElLwYl4P ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=20 + echo -n . .+ [[ 20 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wOSJxClzCx +++ mktemp ++ local LAST_ERR=/tmp/tmp.sKwEHJeC12 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.wOSJxClzCx ++ cat /tmp/tmp.sKwEHJeC12 ++ rm /tmp/tmp.wOSJxClzCx /tmp/tmp.sKwEHJeC12 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=21 + echo -n . .+ [[ 21 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5iJpACDWa6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.qAckc4JDr9 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5iJpACDWa6 ++ cat /tmp/tmp.qAckc4JDr9 ++ rm /tmp/tmp.5iJpACDWa6 /tmp/tmp.qAckc4JDr9 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=22 + echo -n . .+ [[ 22 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VpmkOAKbAq +++ mktemp ++ local LAST_ERR=/tmp/tmp.U9Vrzaheqm ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.VpmkOAKbAq ++ cat /tmp/tmp.U9Vrzaheqm ++ rm /tmp/tmp.VpmkOAKbAq /tmp/tmp.U9Vrzaheqm ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=23 + echo -n . .+ [[ 23 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tP8tELaVBI +++ mktemp ++ local LAST_ERR=/tmp/tmp.Rt6kgz6noP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.tP8tELaVBI ++ cat /tmp/tmp.Rt6kgz6noP ++ rm /tmp/tmp.tP8tELaVBI /tmp/tmp.Rt6kgz6noP ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=24 + echo -n . .+ [[ 24 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PYo3EWqxMx +++ mktemp ++ local LAST_ERR=/tmp/tmp.JUucNo7Dcy ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.PYo3EWqxMx ++ cat /tmp/tmp.JUucNo7Dcy ++ rm /tmp/tmp.PYo3EWqxMx /tmp/tmp.JUucNo7Dcy ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=25 + echo -n . .+ [[ 25 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hgNRxqG8ri +++ mktemp ++ local LAST_ERR=/tmp/tmp.DWwwLO07zH ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.hgNRxqG8ri ++ cat /tmp/tmp.DWwwLO07zH ++ rm /tmp/tmp.hgNRxqG8ri /tmp/tmp.DWwwLO07zH ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=26 + echo -n . .+ [[ 26 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pwQGFebcbI +++ mktemp ++ local LAST_ERR=/tmp/tmp.kMqaLQikLI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.pwQGFebcbI ++ cat /tmp/tmp.kMqaLQikLI ++ rm /tmp/tmp.pwQGFebcbI /tmp/tmp.kMqaLQikLI ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=27 + echo -n . .+ [[ 27 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2QDFXSjeE6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.HavsgX0fOi ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.2QDFXSjeE6 ++ cat /tmp/tmp.HavsgX0fOi ++ rm /tmp/tmp.2QDFXSjeE6 /tmp/tmp.HavsgX0fOi ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=28 + echo -n . .+ [[ 28 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vTaLHEvoJO +++ mktemp ++ local LAST_ERR=/tmp/tmp.j4SfTkAEIN ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.vTaLHEvoJO ++ cat /tmp/tmp.j4SfTkAEIN ++ rm /tmp/tmp.vTaLHEvoJO /tmp/tmp.j4SfTkAEIN ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=29 + echo -n . .+ [[ 29 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jwjDLVH8iQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.jAZqjrbdlJ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jwjDLVH8iQ ++ cat /tmp/tmp.jAZqjrbdlJ ++ rm /tmp/tmp.jwjDLVH8iQ /tmp/tmp.jAZqjrbdlJ ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=30 + echo -n . .+ [[ 30 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QAgxL4axdg +++ mktemp ++ local LAST_ERR=/tmp/tmp.RRvQH7NDZo ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.QAgxL4axdg ++ cat /tmp/tmp.RRvQH7NDZo ++ rm /tmp/tmp.QAgxL4axdg /tmp/tmp.RRvQH7NDZo ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=31 + echo -n . .+ [[ 31 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3W3g1G3QO8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Kr831O6Ooy ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.3W3g1G3QO8 ++ cat /tmp/tmp.Kr831O6Ooy ++ rm /tmp/tmp.3W3g1G3QO8 /tmp/tmp.Kr831O6Ooy ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=32 + echo -n . .+ [[ 32 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8HrbxMnPrs +++ mktemp ++ local LAST_ERR=/tmp/tmp.Vz7PwK9I8u ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.8HrbxMnPrs ++ cat /tmp/tmp.Vz7PwK9I8u ++ rm /tmp/tmp.8HrbxMnPrs /tmp/tmp.Vz7PwK9I8u ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=33 + echo -n . .+ [[ 33 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XnKNRofj8w +++ mktemp ++ local LAST_ERR=/tmp/tmp.u3dKFJLWhZ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.XnKNRofj8w ++ cat /tmp/tmp.u3dKFJLWhZ ++ rm /tmp/tmp.XnKNRofj8w /tmp/tmp.u3dKFJLWhZ ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=34 + echo -n . .+ [[ 34 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.d5aRHgnkKM +++ mktemp ++ local LAST_ERR=/tmp/tmp.T0nvgghg2U ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.d5aRHgnkKM ++ cat /tmp/tmp.T0nvgghg2U ++ rm /tmp/tmp.d5aRHgnkKM /tmp/tmp.T0nvgghg2U ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=35 + echo -n . .+ [[ 35 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tPUoT947TI +++ mktemp ++ local LAST_ERR=/tmp/tmp.7sQMjGQf3W ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.tPUoT947TI ++ cat /tmp/tmp.7sQMjGQf3W ++ rm /tmp/tmp.tPUoT947TI /tmp/tmp.7sQMjGQf3W ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=36 + echo -n . .+ [[ 36 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QveVO4cLmG +++ mktemp ++ local LAST_ERR=/tmp/tmp.i9FMQW7BvF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.QveVO4cLmG ++ cat /tmp/tmp.i9FMQW7BvF ++ rm /tmp/tmp.QveVO4cLmG /tmp/tmp.i9FMQW7BvF ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=37 + echo -n . .+ [[ 37 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.x0MqNXNUE3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.08gc9RZegb ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.x0MqNXNUE3 ++ cat /tmp/tmp.08gc9RZegb ++ rm /tmp/tmp.x0MqNXNUE3 /tmp/tmp.08gc9RZegb ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=38 + echo -n . .+ [[ 38 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FwANuiYnEV +++ mktemp ++ local LAST_ERR=/tmp/tmp.fgc6aIqxh3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.FwANuiYnEV ++ cat /tmp/tmp.fgc6aIqxh3 ++ rm /tmp/tmp.FwANuiYnEV /tmp/tmp.fgc6aIqxh3 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=39 + echo -n . .+ [[ 39 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.CUReXYxav0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.rMbiwsnEWF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.CUReXYxav0 ++ cat /tmp/tmp.rMbiwsnEWF ++ rm /tmp/tmp.CUReXYxav0 /tmp/tmp.rMbiwsnEWF ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=40 + echo -n . .+ [[ 40 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nQ4IxzB6rN +++ mktemp ++ local LAST_ERR=/tmp/tmp.WFJ2LbgYf0 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.nQ4IxzB6rN ++ cat /tmp/tmp.WFJ2LbgYf0 ++ rm /tmp/tmp.nQ4IxzB6rN /tmp/tmp.WFJ2LbgYf0 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=41 + echo -n . .+ [[ 41 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ti4LaoqGCv +++ mktemp ++ local LAST_ERR=/tmp/tmp.tBXNUDnjdd ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ti4LaoqGCv ++ cat /tmp/tmp.tBXNUDnjdd ++ rm /tmp/tmp.ti4LaoqGCv /tmp/tmp.tBXNUDnjdd ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=42 + echo -n . .+ [[ 42 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FbbHDYuT1y +++ mktemp ++ local LAST_ERR=/tmp/tmp.qtY83g7PIu ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.FbbHDYuT1y ++ cat /tmp/tmp.qtY83g7PIu ++ rm /tmp/tmp.FbbHDYuT1y /tmp/tmp.qtY83g7PIu ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=43 + echo -n . .+ [[ 43 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fSCRX1l8jC +++ mktemp ++ local LAST_ERR=/tmp/tmp.Rz36JTXGFl ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.fSCRX1l8jC ++ cat /tmp/tmp.Rz36JTXGFl ++ rm /tmp/tmp.fSCRX1l8jC /tmp/tmp.Rz36JTXGFl ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=44 + echo -n . .+ [[ 44 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fY0QqneDe1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.RXdwlCELYN ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.fY0QqneDe1 ++ cat /tmp/tmp.RXdwlCELYN ++ rm /tmp/tmp.fY0QqneDe1 /tmp/tmp.RXdwlCELYN ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=45 + echo -n . .+ [[ 45 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.q6aPVzKPsB +++ mktemp ++ local LAST_ERR=/tmp/tmp.2KEt6yX92z ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.q6aPVzKPsB ++ cat /tmp/tmp.2KEt6yX92z ++ rm /tmp/tmp.q6aPVzKPsB /tmp/tmp.2KEt6yX92z ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=46 + echo -n . .+ [[ 46 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BoHWbc4Lkv +++ mktemp ++ local LAST_ERR=/tmp/tmp.tjWF6nNEFV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.BoHWbc4Lkv ++ cat /tmp/tmp.tjWF6nNEFV ++ rm /tmp/tmp.BoHWbc4Lkv /tmp/tmp.tjWF6nNEFV ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=47 + echo -n . .+ [[ 47 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fL3Ue8O0Fe +++ mktemp ++ local LAST_ERR=/tmp/tmp.fBOO9S30B7 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.fL3Ue8O0Fe ++ cat /tmp/tmp.fBOO9S30B7 ++ rm /tmp/tmp.fL3Ue8O0Fe /tmp/tmp.fBOO9S30B7 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=48 + echo -n . .+ [[ 48 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.D1FnWOs8Xh +++ mktemp ++ local LAST_ERR=/tmp/tmp.dEsml1QNEd ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.D1FnWOs8Xh ++ cat /tmp/tmp.dEsml1QNEd ++ rm /tmp/tmp.D1FnWOs8Xh /tmp/tmp.dEsml1QNEd ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=49 + echo -n . .+ [[ 49 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rfgXxiAMeU +++ mktemp ++ local LAST_ERR=/tmp/tmp.wnF9y6qHDN ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.rfgXxiAMeU ++ cat /tmp/tmp.wnF9y6qHDN ++ rm /tmp/tmp.rfgXxiAMeU /tmp/tmp.wnF9y6qHDN ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=50 + echo -n . .+ [[ 50 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hsUhR5mCfc +++ mktemp ++ local LAST_ERR=/tmp/tmp.7m0ywrNiNF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.hsUhR5mCfc ++ cat /tmp/tmp.7m0ywrNiNF ++ rm /tmp/tmp.hsUhR5mCfc /tmp/tmp.7m0ywrNiNF ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=51 + echo -n . .+ [[ 51 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MnSoqrAaKp +++ mktemp ++ local LAST_ERR=/tmp/tmp.GEkzVHPC1H ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.MnSoqrAaKp ++ cat /tmp/tmp.GEkzVHPC1H ++ rm /tmp/tmp.MnSoqrAaKp /tmp/tmp.GEkzVHPC1H ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=52 + echo -n . .+ [[ 52 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fnG0lndqOf +++ mktemp ++ local LAST_ERR=/tmp/tmp.lsFQYqYQyb ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.fnG0lndqOf ++ cat /tmp/tmp.lsFQYqYQyb ++ rm /tmp/tmp.fnG0lndqOf /tmp/tmp.lsFQYqYQyb ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=53 + echo -n . .+ [[ 53 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.S9ZCMuuWKe +++ mktemp ++ local LAST_ERR=/tmp/tmp.gW3Uhh7cFw ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.S9ZCMuuWKe ++ cat /tmp/tmp.gW3Uhh7cFw ++ rm /tmp/tmp.S9ZCMuuWKe /tmp/tmp.gW3Uhh7cFw ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=54 + echo -n . .+ [[ 54 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TcVyGbx56z +++ mktemp ++ local LAST_ERR=/tmp/tmp.x6pyLsb49A ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.TcVyGbx56z ++ cat /tmp/tmp.x6pyLsb49A ++ rm /tmp/tmp.TcVyGbx56z /tmp/tmp.x6pyLsb49A ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=55 + echo -n . .+ [[ 55 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.iIa1sHTRfx +++ mktemp ++ local LAST_ERR=/tmp/tmp.XSC02Kumne ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.iIa1sHTRfx ++ cat /tmp/tmp.XSC02Kumne ++ rm /tmp/tmp.iIa1sHTRfx /tmp/tmp.XSC02Kumne ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=56 + echo -n . .+ [[ 56 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Bck2LvayfH +++ mktemp ++ local LAST_ERR=/tmp/tmp.cEoRvwcoeX ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Bck2LvayfH ++ cat /tmp/tmp.cEoRvwcoeX ++ rm /tmp/tmp.Bck2LvayfH /tmp/tmp.cEoRvwcoeX ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=57 + echo -n . .+ [[ 57 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7D74Ao60TS +++ mktemp ++ local LAST_ERR=/tmp/tmp.D7Gnql0k6G ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.7D74Ao60TS ++ cat /tmp/tmp.D7Gnql0k6G ++ rm /tmp/tmp.7D74Ao60TS /tmp/tmp.D7Gnql0k6G ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=58 + echo -n . .+ [[ 58 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kWl4B7bn81 +++ mktemp ++ local LAST_ERR=/tmp/tmp.bgAtjf7DI4 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.kWl4B7bn81 ++ cat /tmp/tmp.bgAtjf7DI4 ++ rm /tmp/tmp.kWl4B7bn81 /tmp/tmp.bgAtjf7DI4 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=59 + echo -n . .+ [[ 59 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lVYN18HVrm +++ mktemp ++ local LAST_ERR=/tmp/tmp.OnPe7nArZV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.lVYN18HVrm ++ cat /tmp/tmp.OnPe7nArZV ++ rm /tmp/tmp.lVYN18HVrm /tmp/tmp.OnPe7nArZV ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=60 + echo -n . .+ [[ 60 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.omKryKlt5p +++ mktemp ++ local LAST_ERR=/tmp/tmp.hiUcJ4f5Ii ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.omKryKlt5p ++ cat /tmp/tmp.hiUcJ4f5Ii ++ rm /tmp/tmp.omKryKlt5p /tmp/tmp.hiUcJ4f5Ii ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=61 + echo -n . .+ [[ 61 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5leNKbU6WG +++ mktemp ++ local LAST_ERR=/tmp/tmp.dLQrdibTUs ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5leNKbU6WG ++ cat /tmp/tmp.dLQrdibTUs ++ rm /tmp/tmp.5leNKbU6WG /tmp/tmp.dLQrdibTUs ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=62 + echo -n . .+ [[ 62 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XmgUwDNSRj +++ mktemp ++ local LAST_ERR=/tmp/tmp.6pmnRCk36H ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.XmgUwDNSRj ++ cat /tmp/tmp.6pmnRCk36H ++ rm /tmp/tmp.XmgUwDNSRj /tmp/tmp.6pmnRCk36H ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=63 + echo -n . .+ [[ 63 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2AIthq9nBG +++ mktemp ++ local LAST_ERR=/tmp/tmp.YkiFmJZ3o9 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.2AIthq9nBG ++ cat /tmp/tmp.YkiFmJZ3o9 ++ rm /tmp/tmp.2AIthq9nBG /tmp/tmp.YkiFmJZ3o9 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=64 + echo -n . .+ [[ 64 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.806bVlqVYR +++ mktemp ++ local LAST_ERR=/tmp/tmp.wvUif9i745 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.806bVlqVYR ++ cat /tmp/tmp.wvUif9i745 ++ rm /tmp/tmp.806bVlqVYR /tmp/tmp.wvUif9i745 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=65 + echo -n . .+ [[ 65 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NgMdLNcr4K +++ mktemp ++ local LAST_ERR=/tmp/tmp.yVj3OmpWbJ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.NgMdLNcr4K ++ cat /tmp/tmp.yVj3OmpWbJ ++ rm /tmp/tmp.NgMdLNcr4K /tmp/tmp.yVj3OmpWbJ ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=66 + echo -n . .+ [[ 66 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5Le1YsMpDk +++ mktemp ++ local LAST_ERR=/tmp/tmp.dpb1JnlAWK ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5Le1YsMpDk ++ cat /tmp/tmp.dpb1JnlAWK ++ rm /tmp/tmp.5Le1YsMpDk /tmp/tmp.dpb1JnlAWK ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=67 + echo -n . .+ [[ 67 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Prt8NbbK46 +++ mktemp ++ local LAST_ERR=/tmp/tmp.lXx3LMZEiD ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Prt8NbbK46 ++ cat /tmp/tmp.lXx3LMZEiD ++ rm /tmp/tmp.Prt8NbbK46 /tmp/tmp.lXx3LMZEiD ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=68 + echo -n . .+ [[ 68 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7jQyrrXpWO +++ mktemp ++ local LAST_ERR=/tmp/tmp.nZOU1NRHOU ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.7jQyrrXpWO ++ cat /tmp/tmp.nZOU1NRHOU ++ rm /tmp/tmp.7jQyrrXpWO /tmp/tmp.nZOU1NRHOU ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=69 + echo -n . .+ [[ 69 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ejENX67tyN +++ mktemp ++ local LAST_ERR=/tmp/tmp.Fxca6sYMDq ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ejENX67tyN ++ cat /tmp/tmp.Fxca6sYMDq ++ rm /tmp/tmp.ejENX67tyN /tmp/tmp.Fxca6sYMDq ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=70 + echo -n . .+ [[ 70 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rtpr1F2nbP +++ mktemp ++ local LAST_ERR=/tmp/tmp.ytd1BN3ftb ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.rtpr1F2nbP ++ cat /tmp/tmp.ytd1BN3ftb ++ rm /tmp/tmp.rtpr1F2nbP /tmp/tmp.ytd1BN3ftb ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=71 + echo -n . .+ [[ 71 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.15s9hncYxl +++ mktemp ++ local LAST_ERR=/tmp/tmp.ch4h9IA6Hx ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.15s9hncYxl ++ cat /tmp/tmp.ch4h9IA6Hx ++ rm /tmp/tmp.15s9hncYxl /tmp/tmp.ch4h9IA6Hx ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=72 + echo -n . .+ [[ 72 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Pmlr202MaV +++ mktemp ++ local LAST_ERR=/tmp/tmp.2vh9lCiLUx ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Pmlr202MaV ++ cat /tmp/tmp.2vh9lCiLUx ++ rm /tmp/tmp.Pmlr202MaV /tmp/tmp.2vh9lCiLUx ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=73 + echo -n . .+ [[ 73 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EbOOc5y1i8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.eke64blbSr ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.EbOOc5y1i8 ++ cat /tmp/tmp.eke64blbSr ++ rm /tmp/tmp.EbOOc5y1i8 /tmp/tmp.eke64blbSr ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=74 + echo -n . .+ [[ 74 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.H6ATaiM9ca +++ mktemp ++ local LAST_ERR=/tmp/tmp.nbrcnXUyAK ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.H6ATaiM9ca ++ cat /tmp/tmp.nbrcnXUyAK ++ rm /tmp/tmp.H6ATaiM9ca /tmp/tmp.nbrcnXUyAK ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=75 + echo -n . .+ [[ 75 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XrnSPP0iKj +++ mktemp ++ local LAST_ERR=/tmp/tmp.gV8tOmQLXm ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.XrnSPP0iKj ++ cat /tmp/tmp.gV8tOmQLXm ++ rm /tmp/tmp.XrnSPP0iKj /tmp/tmp.gV8tOmQLXm ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=76 + echo -n . .+ [[ 76 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xg34Lx9TTH +++ mktemp ++ local LAST_ERR=/tmp/tmp.a9JwIDFri7 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xg34Lx9TTH ++ cat /tmp/tmp.a9JwIDFri7 ++ rm /tmp/tmp.xg34Lx9TTH /tmp/tmp.a9JwIDFri7 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=77 + echo -n . .+ [[ 77 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rHNeQmHF7h +++ mktemp ++ local LAST_ERR=/tmp/tmp.nwyzEB8MrM ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.rHNeQmHF7h ++ cat /tmp/tmp.nwyzEB8MrM ++ rm /tmp/tmp.rHNeQmHF7h /tmp/tmp.nwyzEB8MrM ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=78 + echo -n . .+ [[ 78 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.c7VEBx2UZZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.PJ7hKrCfc6 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.c7VEBx2UZZ ++ cat /tmp/tmp.PJ7hKrCfc6 ++ rm /tmp/tmp.c7VEBx2UZZ /tmp/tmp.PJ7hKrCfc6 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=79 + echo -n . .+ [[ 79 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fO7ZGhk4vB +++ mktemp ++ local LAST_ERR=/tmp/tmp.NX4DcJImFm ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.fO7ZGhk4vB ++ cat /tmp/tmp.NX4DcJImFm ++ rm /tmp/tmp.fO7ZGhk4vB /tmp/tmp.NX4DcJImFm ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=80 + echo -n . .+ [[ 80 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DAqBhXA66D +++ mktemp ++ local LAST_ERR=/tmp/tmp.Beaqdzb2p7 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.DAqBhXA66D ++ cat /tmp/tmp.Beaqdzb2p7 ++ rm /tmp/tmp.DAqBhXA66D /tmp/tmp.Beaqdzb2p7 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=81 + echo -n . .+ [[ 81 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xdPAA4NI19 +++ mktemp ++ local LAST_ERR=/tmp/tmp.8klslF3Mlo ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xdPAA4NI19 ++ cat /tmp/tmp.8klslF3Mlo ++ rm /tmp/tmp.xdPAA4NI19 /tmp/tmp.8klslF3Mlo ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=82 + echo -n . .+ [[ 82 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.m2Jo1RMJSh +++ mktemp ++ local LAST_ERR=/tmp/tmp.z3ApQiuEZF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.m2Jo1RMJSh ++ cat /tmp/tmp.z3ApQiuEZF ++ rm /tmp/tmp.m2Jo1RMJSh /tmp/tmp.z3ApQiuEZF ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=83 + echo -n . .+ [[ 83 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.59F2ebwP92 +++ mktemp ++ local LAST_ERR=/tmp/tmp.RuMZPZlGGX ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.59F2ebwP92 ++ cat /tmp/tmp.RuMZPZlGGX ++ rm /tmp/tmp.59F2ebwP92 /tmp/tmp.RuMZPZlGGX ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=84 + echo -n . .+ [[ 84 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6Y8VnIt23z +++ mktemp ++ local LAST_ERR=/tmp/tmp.pJxkewbuWD ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.6Y8VnIt23z ++ cat /tmp/tmp.pJxkewbuWD ++ rm /tmp/tmp.6Y8VnIt23z /tmp/tmp.pJxkewbuWD ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=85 + echo -n . .+ [[ 85 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2vzqLnN6kB +++ mktemp ++ local LAST_ERR=/tmp/tmp.F8i6GTutWT ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.2vzqLnN6kB ++ cat /tmp/tmp.F8i6GTutWT ++ rm /tmp/tmp.2vzqLnN6kB /tmp/tmp.F8i6GTutWT ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=86 + echo -n . .+ [[ 86 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.T7906P6U3d +++ mktemp ++ local LAST_ERR=/tmp/tmp.9SAv5nNQL0 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.T7906P6U3d ++ cat /tmp/tmp.9SAv5nNQL0 ++ rm /tmp/tmp.T7906P6U3d /tmp/tmp.9SAv5nNQL0 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=87 + echo -n . .+ [[ 87 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5q7un4HZYh +++ mktemp ++ local LAST_ERR=/tmp/tmp.P2rflaK7i6 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5q7un4HZYh ++ cat /tmp/tmp.P2rflaK7i6 ++ rm /tmp/tmp.5q7un4HZYh /tmp/tmp.P2rflaK7i6 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=88 + echo -n . .+ [[ 88 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Dsq0f7TVgy +++ mktemp ++ local LAST_ERR=/tmp/tmp.E9MkZZE7uR ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Dsq0f7TVgy ++ cat /tmp/tmp.E9MkZZE7uR ++ rm /tmp/tmp.Dsq0f7TVgy /tmp/tmp.E9MkZZE7uR ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=89 + echo -n . .+ [[ 89 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7g35zMaea9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.rSjOLSQjJ8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.7g35zMaea9 ++ cat /tmp/tmp.rSjOLSQjJ8 ++ rm /tmp/tmp.7g35zMaea9 /tmp/tmp.rSjOLSQjJ8 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=90 + echo -n . .+ [[ 90 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.h8febx9lsF +++ mktemp ++ local LAST_ERR=/tmp/tmp.3A5m042iND ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.h8febx9lsF ++ cat /tmp/tmp.3A5m042iND ++ rm /tmp/tmp.h8febx9lsF /tmp/tmp.3A5m042iND ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=91 + echo -n . .+ [[ 91 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Hyci5xZgwe +++ mktemp ++ local LAST_ERR=/tmp/tmp.u9CFTr2gYV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Hyci5xZgwe ++ cat /tmp/tmp.u9CFTr2gYV ++ rm /tmp/tmp.Hyci5xZgwe /tmp/tmp.u9CFTr2gYV ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=92 + echo -n . .+ [[ 92 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.aZjdSZwO4C +++ mktemp ++ local LAST_ERR=/tmp/tmp.by4KsGScP5 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.aZjdSZwO4C ++ cat /tmp/tmp.by4KsGScP5 ++ rm /tmp/tmp.aZjdSZwO4C /tmp/tmp.by4KsGScP5 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=93 + echo -n . .+ [[ 93 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6jztuaAPLo +++ mktemp ++ local LAST_ERR=/tmp/tmp.G6Xq7R6PDC ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.6jztuaAPLo ++ cat /tmp/tmp.G6Xq7R6PDC ++ rm /tmp/tmp.6jztuaAPLo /tmp/tmp.G6Xq7R6PDC ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=94 + echo -n . .+ [[ 94 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LkMlS8fUd2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.VJioc2PfaL ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.LkMlS8fUd2 ++ cat /tmp/tmp.VJioc2PfaL ++ rm /tmp/tmp.LkMlS8fUd2 /tmp/tmp.VJioc2PfaL ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=95 + echo -n . .+ [[ 95 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.h81o0W5CFH +++ mktemp ++ local LAST_ERR=/tmp/tmp.39PKAll7kt ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.h81o0W5CFH ++ cat /tmp/tmp.39PKAll7kt ++ rm /tmp/tmp.h81o0W5CFH /tmp/tmp.39PKAll7kt ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=96 + echo -n . .+ [[ 96 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.d3UidJGioF +++ mktemp ++ local LAST_ERR=/tmp/tmp.GcGOO681Hr ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.d3UidJGioF ++ cat /tmp/tmp.GcGOO681Hr ++ rm /tmp/tmp.d3UidJGioF /tmp/tmp.GcGOO681Hr ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=97 + echo -n . .+ [[ 97 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.I7wYg0DpCd +++ mktemp ++ local LAST_ERR=/tmp/tmp.pDDUOPoIPP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.I7wYg0DpCd ++ cat /tmp/tmp.pDDUOPoIPP ++ rm /tmp/tmp.I7wYg0DpCd /tmp/tmp.pDDUOPoIPP ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=98 + echo -n . .+ [[ 98 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4eZoZvLPKS +++ mktemp ++ local LAST_ERR=/tmp/tmp.VXNqG8VsUo ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4eZoZvLPKS ++ cat /tmp/tmp.VXNqG8VsUo ++ rm /tmp/tmp.4eZoZvLPKS /tmp/tmp.VXNqG8VsUo ++ return 0 + [[ error =~ ready ]] + sleep 1 + timeout=99 + echo -n . .+ [[ 99 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uwS9UREdGo +++ mktemp ++ local LAST_ERR=/tmp/tmp.c9CqdXERDm ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.uwS9UREdGo ++ cat /tmp/tmp.c9CqdXERDm ++ rm /tmp/tmp.uwS9UREdGo /tmp/tmp.c9CqdXERDm ++ return 0 + [[ error =~ ready ]] + sleep 1 + timeout=100 + echo -n . .+ [[ 100 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WrHzrB9jYu +++ mktemp ++ local LAST_ERR=/tmp/tmp.BXwPnDLSmY ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.WrHzrB9jYu ++ cat /tmp/tmp.BXwPnDLSmY ++ rm /tmp/tmp.WrHzrB9jYu /tmp/tmp.BXwPnDLSmY ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=101 + echo -n . .+ [[ 101 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Yji9KPJ07S +++ mktemp ++ local LAST_ERR=/tmp/tmp.KSoRdFylZx ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Yji9KPJ07S ++ cat /tmp/tmp.KSoRdFylZx ++ rm /tmp/tmp.Yji9KPJ07S /tmp/tmp.KSoRdFylZx ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=102 + echo -n . .+ [[ 102 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rX4xxAFI3f +++ mktemp ++ local LAST_ERR=/tmp/tmp.gT58MbemBe ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.rX4xxAFI3f ++ cat /tmp/tmp.gT58MbemBe ++ rm /tmp/tmp.rX4xxAFI3f /tmp/tmp.gT58MbemBe ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=103 + echo -n . .+ [[ 103 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jFq8PvJ6x0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.PZMnskm8AI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jFq8PvJ6x0 ++ cat /tmp/tmp.PZMnskm8AI ++ rm /tmp/tmp.jFq8PvJ6x0 /tmp/tmp.PZMnskm8AI ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=104 + echo -n . .+ [[ 104 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XRWct0kMpF +++ mktemp ++ local LAST_ERR=/tmp/tmp.QHWAIAu7fl ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.XRWct0kMpF ++ cat /tmp/tmp.QHWAIAu7fl ++ rm /tmp/tmp.XRWct0kMpF /tmp/tmp.QHWAIAu7fl ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=105 + echo -n . .+ [[ 105 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uQfclTSP7p +++ mktemp ++ local LAST_ERR=/tmp/tmp.TKQXdx2ubV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.uQfclTSP7p ++ cat /tmp/tmp.TKQXdx2ubV ++ rm /tmp/tmp.uQfclTSP7p /tmp/tmp.TKQXdx2ubV ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=106 + echo -n . .+ [[ 106 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9kQ6PyKW8V +++ mktemp ++ local LAST_ERR=/tmp/tmp.oF2s5bSGwF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9kQ6PyKW8V ++ cat /tmp/tmp.oF2s5bSGwF ++ rm /tmp/tmp.9kQ6PyKW8V /tmp/tmp.oF2s5bSGwF ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=107 + echo -n . .+ [[ 107 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hwfFGBOMi7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.0xTkKSeSJ0 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.hwfFGBOMi7 ++ cat /tmp/tmp.0xTkKSeSJ0 ++ rm /tmp/tmp.hwfFGBOMi7 /tmp/tmp.0xTkKSeSJ0 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=108 + echo -n . .+ [[ 108 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ppdQE8kCJw +++ mktemp ++ local LAST_ERR=/tmp/tmp.setUIj1XkZ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ppdQE8kCJw ++ cat /tmp/tmp.setUIj1XkZ ++ rm /tmp/tmp.ppdQE8kCJw /tmp/tmp.setUIj1XkZ ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=109 + echo -n . .+ [[ 109 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cytsJHCp8O +++ mktemp ++ local LAST_ERR=/tmp/tmp.M1D9JqBhjv ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.cytsJHCp8O ++ cat /tmp/tmp.M1D9JqBhjv ++ rm /tmp/tmp.cytsJHCp8O /tmp/tmp.M1D9JqBhjv ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=110 + echo -n . .+ [[ 110 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MppjEFcJ4Q +++ mktemp ++ local LAST_ERR=/tmp/tmp.fDLVED2Yeb ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.MppjEFcJ4Q ++ cat /tmp/tmp.fDLVED2Yeb ++ rm /tmp/tmp.MppjEFcJ4Q /tmp/tmp.fDLVED2Yeb ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=111 + echo -n . .+ [[ 111 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.J0PGo4338I +++ mktemp ++ local LAST_ERR=/tmp/tmp.Jn0hc5VyGU ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.J0PGo4338I ++ cat /tmp/tmp.Jn0hc5VyGU ++ rm /tmp/tmp.J0PGo4338I /tmp/tmp.Jn0hc5VyGU ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=112 + echo -n . .+ [[ 112 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.V4yKS0SbJf +++ mktemp ++ local LAST_ERR=/tmp/tmp.14VqKe48i2 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.V4yKS0SbJf ++ cat /tmp/tmp.14VqKe48i2 ++ rm /tmp/tmp.V4yKS0SbJf /tmp/tmp.14VqKe48i2 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=113 + echo -n . .+ [[ 113 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ccIBfoK8aE +++ mktemp ++ local LAST_ERR=/tmp/tmp.JwtpSIHoXe ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ccIBfoK8aE ++ cat /tmp/tmp.JwtpSIHoXe ++ rm /tmp/tmp.ccIBfoK8aE /tmp/tmp.JwtpSIHoXe ++ return 0 + [[ ready =~ ready ]] + echo + log 'psmdb/some-name is ready: OK' + set +o xtrace [2026-03-10T06:56:52+0000] psmdb/some-name is ready: OK + desc 'Adding new user' + set +o xtrace ----------------------------------------------------------------------------------- Adding new user ----------------------------------------------------------------------------------- + run_mongo 'db.createUser({user:"myApp",pwd:"myPass",roles:[{db:"myApp",role:"readWrite"}]})' userAdmin:userAdmin123456@some-name-rs0.demand-backup-if-unhealthy-355 + local 'command=db.createUser({user:"myApp",pwd:"myPass",roles:[{db:"myApp",role:"readWrite"}]})' + local uri=userAdmin:userAdmin123456@some-name-rs0.demand-backup-if-unhealthy-355 + local driver=mongodb+srv + local suffix=.svc.cluster.local + local mongo_flag= + local replica_set=rs0 + [[ userAdmin:userAdmin123456@some-name-rs0.demand-backup-if-unhealthy-355 == *cfg* ]] ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QJXXQtRAxq +++ mktemp ++ local LAST_ERR=/tmp/tmp.9RLAd8JS5S ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.QJXXQtRAxq ++ cat /tmp/tmp.9RLAd8JS5S ++ rm /tmp/tmp.QJXXQtRAxq /tmp/tmp.9RLAd8JS5S ++ return 0 + local client_container=psmdb-client-bb8b97679-wx99q + kubectl_bin exec psmdb-client-bb8b97679-wx99q -- bash -c 'printf '\''db.createUser({user:"myApp",pwd:"myPass",roles:[{db:"myApp",role:"readWrite"}]})\n'\'' | mongo mongodb+srv://userAdmin:userAdmin123456@some-name-rs0.demand-backup-if-unhealthy-355.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.wBvVmMDRh8 ++ mktemp + local LAST_ERR=/tmp/tmp.eYYfajKH1E + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-wx99q -- bash -c 'printf '\''db.createUser({user:"myApp",pwd:"myPass",roles:[{db:"myApp",role:"readWrite"}]})\n'\'' | mongo mongodb+srv://userAdmin:userAdmin123456@some-name-rs0.demand-backup-if-unhealthy-355.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.wBvVmMDRh8 Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-rs0-2.some-name-rs0.demand-backup-if-unhealthy-355.svc.cluster.local:27017,some-name-rs0-0.some-name-rs0.demand-backup-if-unhealthy-355.svc.cluster.local:27017,some-name-rs0-1.some-name-rs0.demand-backup-if-unhealthy-355.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb&replicaSet=rs0&ssl=false Implicit session: session { "id" : UUID("8ffd273a-5d7d-4f2f-a4bb-8afaae1981b1") } Percona Server for MongoDB server version: v8.0.19-7 WARNING: shell and server versions do not match Successfully added user: { "user" : "myApp", "roles" : [ { "db" : "myApp", "role" : "readWrite" } ] } bye + cat /tmp/tmp.eYYfajKH1E + rm /tmp/tmp.wBvVmMDRh8 /tmp/tmp.eYYfajKH1E + return 0 + echo -n 'Verifying user creation...' Verifying user creation...+ until_with_timeout 10 check_user + local timeout=10 + shift + local elapsed=0 + check_user + run_mongo 'db.getUser("myApp")' userAdmin:userAdmin123456@some-name-rs0.demand-backup-if-unhealthy-355 + local 'command=db.getUser("myApp")' + local uri=userAdmin:userAdmin123456@some-name-rs0.demand-backup-if-unhealthy-355 + local driver=mongodb+srv + local suffix=.svc.cluster.local + local mongo_flag= + local replica_set=rs0 + [[ userAdmin:userAdmin123456@some-name-rs0.demand-backup-if-unhealthy-355 == *cfg* ]] + grep -q '"user" : "myApp"' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OKLQghBCI3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.oxP8ick1rU ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.OKLQghBCI3 ++ cat /tmp/tmp.oxP8ick1rU ++ rm /tmp/tmp.OKLQghBCI3 /tmp/tmp.oxP8ick1rU ++ return 0 + local client_container=psmdb-client-bb8b97679-wx99q + kubectl_bin exec psmdb-client-bb8b97679-wx99q -- bash -c 'printf '\''db.getUser("myApp")\n'\'' | mongo mongodb+srv://userAdmin:userAdmin123456@some-name-rs0.demand-backup-if-unhealthy-355.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.Gs0JHujfRb ++ mktemp + local LAST_ERR=/tmp/tmp.fXaVqNeAx5 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-wx99q -- bash -c 'printf '\''db.getUser("myApp")\n'\'' | mongo mongodb+srv://userAdmin:userAdmin123456@some-name-rs0.demand-backup-if-unhealthy-355.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Gs0JHujfRb + cat /tmp/tmp.fXaVqNeAx5 + rm /tmp/tmp.Gs0JHujfRb /tmp/tmp.fXaVqNeAx5 + return 0 + return 0 + desc 'Testing logical backup with an unhealthy container' + set +o xtrace ----------------------------------------------------------------------------------- Testing logical backup with an unhealthy container ----------------------------------------------------------------------------------- + type=logical + write_data 'This is some data for backup if unhealthy' + local 'data=This is some data for backup if unhealthy' + local find_prefix= + run_mongo 'use myApp\n db.test.insert({ x: "This is some data for backup if unhealthy" })' myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-355 + local 'command=use myApp\n db.test.insert({ x: "This is some data for backup if unhealthy" })' + local uri=myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-355 + local driver=mongodb+srv + local suffix=.svc.cluster.local + local mongo_flag= + local replica_set=rs0 + [[ myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-355 == *cfg* ]] ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XC0T9jyPyE +++ mktemp ++ local LAST_ERR=/tmp/tmp.t1Ua3Awpmo ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.XC0T9jyPyE ++ cat /tmp/tmp.t1Ua3Awpmo ++ rm /tmp/tmp.XC0T9jyPyE /tmp/tmp.t1Ua3Awpmo ++ return 0 + local client_container=psmdb-client-bb8b97679-wx99q + kubectl_bin exec psmdb-client-bb8b97679-wx99q -- bash -c 'printf '\''use myApp\n db.test.insert({ x: "This is some data for backup if unhealthy" })\n'\'' | mongo mongodb+srv://myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-355.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.Zeu0Xz6qsy ++ mktemp + local LAST_ERR=/tmp/tmp.9kileHyJ7e + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-wx99q -- bash -c 'printf '\''use myApp\n db.test.insert({ x: "This is some data for backup if unhealthy" })\n'\'' | mongo mongodb+srv://myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-355.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Zeu0Xz6qsy Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-rs0-2.some-name-rs0.demand-backup-if-unhealthy-355.svc.cluster.local:27017,some-name-rs0-0.some-name-rs0.demand-backup-if-unhealthy-355.svc.cluster.local:27017,some-name-rs0-1.some-name-rs0.demand-backup-if-unhealthy-355.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb&replicaSet=rs0&ssl=false Implicit session: session { "id" : UUID("0d15b127-b7a5-4b4e-b1bd-0bd8e15d7bfc") } Percona Server for MongoDB server version: v8.0.19-7 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.9kileHyJ7e + rm /tmp/tmp.Zeu0Xz6qsy /tmp/tmp.9kileHyJ7e + return 0 + compare_mongo_cmd find myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-355 '' .svc.cluster.local myApp test + local command=find + local uri=myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-355 + local postfix= + local suffix=.svc.cluster.local + local database=myApp + local collection=test + local sort= + local tls=false + local replicaset= + local 'full_command=db.test.find()' + [[ -n '' ]] + log 'running db.test.find() in myApp' + set +o xtrace [2026-03-10T06:57:02+0000] running db.test.find() in myApp + [[ false == \t\r\u\e ]] + mongo_command=run_mongo + run_mongo 'use myApp\n db.test.find()' myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-355 mongodb .svc.cluster.local '' + grep -E -v 'I NETWORK|W NETWORK|F NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:|Started a new thread for the timer service' + local 'command=use myApp\n db.test.find()' + local uri=myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-355 + local driver=mongodb + local suffix=.svc.cluster.local + local mongo_flag= + local replica_set=rs0 + [[ myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-355 == *cfg* ]] + /usr/sbin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LLem9mNKLT +++ mktemp ++ local LAST_ERR=/tmp/tmp.ftP3pH5s4U ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.LLem9mNKLT ++ cat /tmp/tmp.ftP3pH5s4U ++ rm /tmp/tmp.LLem9mNKLT /tmp/tmp.ftP3pH5s4U ++ return 0 + local client_container=psmdb-client-bb8b97679-wx99q + kubectl_bin exec psmdb-client-bb8b97679-wx99q -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-355.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.TYSK8PmJPp ++ mktemp + local LAST_ERR=/tmp/tmp.mi3m9j8ePa + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-wx99q -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-355.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.TYSK8PmJPp + cat /tmp/tmp.mi3m9j8ePa + rm /tmp/tmp.TYSK8PmJPp /tmp/tmp.mi3m9j8ePa + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/demand-backup-if-unhealthy/compare/find.json /tmp/tmp.gaQrVt1nmF/find + wait_sts_to_become_not_ready + local sts_name=some-name-rs0 + local timeout=60 + local pod_name=some-name-rs0-1 + local interval=2 + local elapsed=0 + echo 'Updating cluster with invalid image...' Updating cluster with invalid image... + update_with_invalid_db_image + local cluster_name=some-name + yq /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/demand-backup-if-unhealthy/conf/some-name.yml + yq '.spec.image="invalid-image:invalid-tag"' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.0r54PObv1H ++ mktemp + local LAST_ERR=/tmp/tmp.BTKaT4t5rO + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.0r54PObv1H perconaservermongodb.psmdb.percona.com/some-name configured + cat /tmp/tmp.BTKaT4t5rO + rm /tmp/tmp.0r54PObv1H /tmp/tmp.BTKaT4t5rO + return 0 + echo -n 'Wait for statefulset some-name-rs0 to become not ready...' Wait for statefulset some-name-rs0 to become not ready...+ is_sts_ready some-name-rs0 + local sts_name=some-name-rs0 + local replicas ready_replicas + kubectl_bin get sts some-name-rs0 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lKeg3SrakS +++ mktemp ++ local LAST_ERR=/tmp/tmp.xnNTlNnnAO ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.lKeg3SrakS ++ cat /tmp/tmp.xnNTlNnnAO ++ rm /tmp/tmp.lKeg3SrakS /tmp/tmp.xnNTlNnnAO ++ return 0 + replicas=3 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.CZyVbYVbhM +++ mktemp ++ local LAST_ERR=/tmp/tmp.qCQWLmraDI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.CZyVbYVbhM ++ cat /tmp/tmp.qCQWLmraDI ++ rm /tmp/tmp.CZyVbYVbhM /tmp/tmp.qCQWLmraDI ++ return 0 + ready_replicas=3 + [[ 3 -eq 3 ]] + (( elapsed >= timeout )) + sleep 2 + (( elapsed += interval )) + echo -n . .+ is_sts_ready some-name-rs0 + local sts_name=some-name-rs0 + local replicas ready_replicas + kubectl_bin get sts some-name-rs0 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JCVfGBOO5P +++ mktemp ++ local LAST_ERR=/tmp/tmp.xwOhsjrRTV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.JCVfGBOO5P ++ cat /tmp/tmp.xwOhsjrRTV ++ rm /tmp/tmp.JCVfGBOO5P /tmp/tmp.xwOhsjrRTV ++ return 0 + replicas=3 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.856sRDjTEo +++ mktemp ++ local LAST_ERR=/tmp/tmp.IZwflogcTA ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.856sRDjTEo ++ cat /tmp/tmp.IZwflogcTA ++ rm /tmp/tmp.856sRDjTEo /tmp/tmp.IZwflogcTA ++ return 0 + ready_replicas=3 + [[ 3 -eq 3 ]] + (( elapsed >= timeout )) + sleep 2 + (( elapsed += interval )) + echo -n . .+ is_sts_ready some-name-rs0 + local sts_name=some-name-rs0 + local replicas ready_replicas + kubectl_bin get sts some-name-rs0 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ET3c9UfKrk +++ mktemp ++ local LAST_ERR=/tmp/tmp.D94mtwyW8H ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ET3c9UfKrk ++ cat /tmp/tmp.D94mtwyW8H ++ rm /tmp/tmp.ET3c9UfKrk /tmp/tmp.D94mtwyW8H ++ return 0 + replicas=3 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MZliSt7ez3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.t1jblfe0eU ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.MZliSt7ez3 ++ cat /tmp/tmp.t1jblfe0eU ++ rm /tmp/tmp.MZliSt7ez3 /tmp/tmp.t1jblfe0eU ++ return 0 + ready_replicas=3 + [[ 3 -eq 3 ]] + (( elapsed >= timeout )) + sleep 2 + (( elapsed += interval )) + echo -n . .+ is_sts_ready some-name-rs0 + local sts_name=some-name-rs0 + local replicas ready_replicas + kubectl_bin get sts some-name-rs0 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1XFwX4JN1a +++ mktemp ++ local LAST_ERR=/tmp/tmp.ky2gfm3GSC ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.1XFwX4JN1a ++ cat /tmp/tmp.ky2gfm3GSC ++ rm /tmp/tmp.1XFwX4JN1a /tmp/tmp.ky2gfm3GSC ++ return 0 + replicas=3 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3guPKZbmWB +++ mktemp ++ local LAST_ERR=/tmp/tmp.v7UNMq1qwG ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.3guPKZbmWB ++ cat /tmp/tmp.v7UNMq1qwG ++ rm /tmp/tmp.3guPKZbmWB /tmp/tmp.v7UNMq1qwG ++ return 0 + ready_replicas=3 + [[ 3 -eq 3 ]] + (( elapsed >= timeout )) + sleep 2 + (( elapsed += interval )) + echo -n . .+ is_sts_ready some-name-rs0 + local sts_name=some-name-rs0 + local replicas ready_replicas + kubectl_bin get sts some-name-rs0 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.r7zS9b7kIW +++ mktemp ++ local LAST_ERR=/tmp/tmp.wqpET6KZxs ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.r7zS9b7kIW ++ cat /tmp/tmp.wqpET6KZxs ++ rm /tmp/tmp.r7zS9b7kIW /tmp/tmp.wqpET6KZxs ++ return 0 + replicas=3 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.r2WdDyTcRh +++ mktemp ++ local LAST_ERR=/tmp/tmp.YwpeSB8QMj ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.r2WdDyTcRh ++ cat /tmp/tmp.YwpeSB8QMj ++ rm /tmp/tmp.r2WdDyTcRh /tmp/tmp.YwpeSB8QMj ++ return 0 + ready_replicas=2 + [[ 3 -eq 2 ]] + wait_for_cluster_state some-name 'initializing|error' + local cluster_name=some-name + local 'target_state=initializing|error' + echo -n 'Waiting for psmdb/some-name to reach initializing|error state' Waiting for psmdb/some-name to reach initializing|error state+ local timeout=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.h0chn77DKa +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZH4GanUVmW ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.h0chn77DKa ++ cat /tmp/tmp.ZH4GanUVmW ++ rm /tmp/tmp.h0chn77DKa /tmp/tmp.ZH4GanUVmW ++ return 0 + [[ initializing =~ initializing|error ]] + echo + log 'psmdb/some-name is initializing|error: OK' + set +o xtrace [2026-03-10T06:57:35+0000] psmdb/some-name is initializing|error: OK + backup_name_minio=backup-minio-logical + run_backup minio backup-minio-logical logical + local storage=minio + local backup_name=backup-minio-logical + local type=logical + log 'running backup backup-minio-logical' + set +o xtrace [2026-03-10T06:57:35+0000] running backup backup-minio-logical + yq eval '.metadata.name = "backup-minio-logical" | .spec.storageName = "minio" | .spec.type = "logical"' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/demand-backup-if-unhealthy/conf/backup-minio.yml + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.nx4jtApvCI ++ mktemp + local LAST_ERR=/tmp/tmp.ycq4juODAj + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.nx4jtApvCI perconaservermongodbbackup.psmdb.percona.com/backup-minio-logical created + cat /tmp/tmp.ycq4juODAj + rm /tmp/tmp.nx4jtApvCI /tmp/tmp.ycq4juODAj + return 0 + verify_sts_not_ready + local sts_name=some-name-rs0 + is_sts_ready some-name-rs0 + local sts_name=some-name-rs0 + local replicas ready_replicas + kubectl_bin get sts some-name-rs0 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zfT6EgfK2j +++ mktemp ++ local LAST_ERR=/tmp/tmp.YUjmNhJ83Q ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.zfT6EgfK2j ++ cat /tmp/tmp.YUjmNhJ83Q ++ rm /tmp/tmp.zfT6EgfK2j /tmp/tmp.YUjmNhJ83Q ++ return 0 + replicas=3 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5fB92TtS00 +++ mktemp ++ local LAST_ERR=/tmp/tmp.B9v6xZxzYR ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5fB92TtS00 ++ cat /tmp/tmp.B9v6xZxzYR ++ rm /tmp/tmp.5fB92TtS00 /tmp/tmp.B9v6xZxzYR ++ return 0 + ready_replicas=2 + [[ 3 -eq 2 ]] + wait_for_cluster_state some-name 'initializing|error' + local cluster_name=some-name + local 'target_state=initializing|error' + echo -n 'Waiting for psmdb/some-name to reach initializing|error state' Waiting for psmdb/some-name to reach initializing|error state+ local timeout=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yCP11PKXgf +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZHtWTFvMyh ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.yCP11PKXgf ++ cat /tmp/tmp.ZHtWTFvMyh ++ rm /tmp/tmp.yCP11PKXgf /tmp/tmp.ZHtWTFvMyh ++ return 0 + [[ initializing =~ initializing|error ]] + echo + log 'psmdb/some-name is initializing|error: OK' + set +o xtrace [2026-03-10T06:57:41+0000] psmdb/some-name is initializing|error: OK + desc 'Verifying backup-minio-logical and update cluster back to ready state' + set +o xtrace ----------------------------------------------------------------------------------- Verifying backup-minio-logical and update cluster back to ready state ----------------------------------------------------------------------------------- + wait_backup backup-minio-logical ready + local backup_name=backup-minio-logical + local target_state=ready + set +o xtrace waiting for backup-minio-logical to reach ready state.............OK + verify_sts_not_ready + local sts_name=some-name-rs0 + is_sts_ready some-name-rs0 + local sts_name=some-name-rs0 + local replicas ready_replicas + kubectl_bin get sts some-name-rs0 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.u6BA1Xc7di +++ mktemp ++ local LAST_ERR=/tmp/tmp.lKEd4fsFfb ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.spec.replicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.u6BA1Xc7di ++ cat /tmp/tmp.lKEd4fsFfb ++ rm /tmp/tmp.u6BA1Xc7di /tmp/tmp.lKEd4fsFfb ++ return 0 + replicas=3 ++ kubectl_bin get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pA5g5Gh4qn +++ mktemp ++ local LAST_ERR=/tmp/tmp.zjrQRomI5o ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts some-name-rs0 -o 'jsonpath={.status.readyReplicas}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.pA5g5Gh4qn ++ cat /tmp/tmp.zjrQRomI5o ++ rm /tmp/tmp.pA5g5Gh4qn /tmp/tmp.zjrQRomI5o ++ return 0 + ready_replicas=2 + [[ 3 -eq 2 ]] + wait_for_cluster_state some-name 'initializing|error' + local cluster_name=some-name + local 'target_state=initializing|error' + echo -n 'Waiting for psmdb/some-name to reach initializing|error state' Waiting for psmdb/some-name to reach initializing|error state+ local timeout=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VJIjboLwiW +++ mktemp ++ local LAST_ERR=/tmp/tmp.Wlz7qs7Y2c ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.VJIjboLwiW ++ cat /tmp/tmp.Wlz7qs7Y2c ++ rm /tmp/tmp.VJIjboLwiW /tmp/tmp.Wlz7qs7Y2c ++ return 0 + [[ initializing =~ initializing|error ]] + echo + log 'psmdb/some-name is initializing|error: OK' + set +o xtrace [2026-03-10T06:58:09+0000] psmdb/some-name is initializing|error: OK + rollback_invalid_image + local cluster_name=some-name + local timeout= + apply_cluster /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/demand-backup-if-unhealthy/conf/some-name.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/demand-backup-if-unhealthy/conf/some-name.yml + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/demand-backup-if-unhealthy/conf/some-name.yml + yq eval '(.spec | select(.image == null)).image = "docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0"' ++ mktemp + yq eval '(.spec | select(has("pmm"))).pmm.image = "docker.io/percona/pmm-client:2.44.1-1"' + yq eval '(.spec | select(has("initImage"))).initImage = "docker.io/perconalab/percona-server-mongodb-operator:PR-2266-f9fc55604"' + yq eval '(.spec | select(has("backup"))).backup.image = "docker.io/perconalab/percona-server-mongodb-operator:main-backup"' + local LAST_OUT=/tmp/tmp.fIvqohngLF ++ mktemp + /usr/sbin/sed -e s/NAME_SPACE/demand-backup-if-unhealthy-355/g + yq eval '.spec.upgradeOptions.apply="Never"' + local LAST_ERR=/tmp/tmp.XZdCmzc0oS + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.fIvqohngLF perconaservermongodb.psmdb.percona.com/some-name configured + cat /tmp/tmp.XZdCmzc0oS + rm /tmp/tmp.fIvqohngLF /tmp/tmp.XZdCmzc0oS + return 0 + echo -n 'Waiting new statefulset version...' Waiting new statefulset version...+ kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.CdStfhA0zi ++ mktemp + local LAST_ERR=/tmp/tmp.IYwhAeuIqr + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.CdStfhA0zi + cat /tmp/tmp.IYwhAeuIqr + rm /tmp/tmp.CdStfhA0zi /tmp/tmp.IYwhAeuIqr + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.PE0pc67GPd ++ mktemp + local LAST_ERR=/tmp/tmp.2FbVEn68bb + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.PE0pc67GPd + cat /tmp/tmp.2FbVEn68bb + rm /tmp/tmp.PE0pc67GPd /tmp/tmp.2FbVEn68bb + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.mnldvrzqLc ++ mktemp + local LAST_ERR=/tmp/tmp.VrMaL7yHre + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.mnldvrzqLc + cat /tmp/tmp.VrMaL7yHre + rm /tmp/tmp.mnldvrzqLc /tmp/tmp.VrMaL7yHre + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.LZRxbMmP8V ++ mktemp + local LAST_ERR=/tmp/tmp.m9p3wTTOrp + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.LZRxbMmP8V + cat /tmp/tmp.m9p3wTTOrp + rm /tmp/tmp.LZRxbMmP8V /tmp/tmp.m9p3wTTOrp + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.DKxv0r7vGe ++ mktemp + local LAST_ERR=/tmp/tmp.OGFQlrD4or + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.DKxv0r7vGe + cat /tmp/tmp.OGFQlrD4or + rm /tmp/tmp.DKxv0r7vGe /tmp/tmp.OGFQlrD4or + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.im7qkFVYll ++ mktemp + local LAST_ERR=/tmp/tmp.bH89hMO38h + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.im7qkFVYll + cat /tmp/tmp.bH89hMO38h + rm /tmp/tmp.im7qkFVYll /tmp/tmp.bH89hMO38h + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.PE7XjQrlJJ ++ mktemp + local LAST_ERR=/tmp/tmp.wJ0YW5zYSn + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.PE7XjQrlJJ + cat /tmp/tmp.wJ0YW5zYSn + rm /tmp/tmp.PE7XjQrlJJ /tmp/tmp.wJ0YW5zYSn + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.opVSTXNkYr ++ mktemp + local LAST_ERR=/tmp/tmp.YbC8QhCCQX + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.opVSTXNkYr + cat /tmp/tmp.YbC8QhCCQX + rm /tmp/tmp.opVSTXNkYr /tmp/tmp.YbC8QhCCQX + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' ++ mktemp + grep -q invalid-image:invalid-tag + local LAST_OUT=/tmp/tmp.o0HDLZ8BMs ++ mktemp + local LAST_ERR=/tmp/tmp.SJpm0WnNw9 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.o0HDLZ8BMs + cat /tmp/tmp.SJpm0WnNw9 + rm /tmp/tmp.o0HDLZ8BMs /tmp/tmp.SJpm0WnNw9 + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.U1l5cJfATh ++ mktemp + local LAST_ERR=/tmp/tmp.rEEy1sp2zt + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.U1l5cJfATh + cat /tmp/tmp.rEEy1sp2zt + rm /tmp/tmp.U1l5cJfATh /tmp/tmp.rEEy1sp2zt + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.7hFdQzoFMa ++ mktemp + local LAST_ERR=/tmp/tmp.s5w7kSGKyY + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.7hFdQzoFMa + cat /tmp/tmp.s5w7kSGKyY + rm /tmp/tmp.7hFdQzoFMa /tmp/tmp.s5w7kSGKyY + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.C2b9QSHcDD ++ mktemp + local LAST_ERR=/tmp/tmp.hg0da1xJ3E + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.C2b9QSHcDD + cat /tmp/tmp.hg0da1xJ3E + rm /tmp/tmp.C2b9QSHcDD /tmp/tmp.hg0da1xJ3E + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.gz11YWruc0 ++ mktemp + local LAST_ERR=/tmp/tmp.8pqaZlBpXX + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.gz11YWruc0 + cat /tmp/tmp.8pqaZlBpXX + rm /tmp/tmp.gz11YWruc0 /tmp/tmp.8pqaZlBpXX + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.Woqrc07Upa ++ mktemp + local LAST_ERR=/tmp/tmp.pM4iQRTnPZ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Woqrc07Upa + cat /tmp/tmp.pM4iQRTnPZ + rm /tmp/tmp.Woqrc07Upa /tmp/tmp.pM4iQRTnPZ + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.hJCOskgvp4 ++ mktemp + local LAST_ERR=/tmp/tmp.cQC3nt9AK4 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.hJCOskgvp4 + cat /tmp/tmp.cQC3nt9AK4 + rm /tmp/tmp.hJCOskgvp4 /tmp/tmp.cQC3nt9AK4 + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.6M2Au9mmEQ ++ mktemp + local LAST_ERR=/tmp/tmp.ZuTPxbQVog + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.6M2Au9mmEQ + cat /tmp/tmp.ZuTPxbQVog + rm /tmp/tmp.6M2Au9mmEQ /tmp/tmp.ZuTPxbQVog + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + grep -q invalid-image:invalid-tag ++ mktemp + local LAST_OUT=/tmp/tmp.zCH9sRIgyq ++ mktemp + local LAST_ERR=/tmp/tmp.5S0ZvXakS5 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.zCH9sRIgyq + cat /tmp/tmp.5S0ZvXakS5 + rm /tmp/tmp.zCH9sRIgyq /tmp/tmp.5S0ZvXakS5 + return 0 + echo -n . .+ sleep 1 + kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' ++ mktemp + grep -q invalid-image:invalid-tag + local LAST_OUT=/tmp/tmp.icFzY1sPeX ++ mktemp + local LAST_ERR=/tmp/tmp.aZVs72WJ5z + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get statefulset some-name-rs0 -o 'jsonpath={range .spec.template.spec.containers[*]}{.image}{"\n"}{end}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.icFzY1sPeX + cat /tmp/tmp.aZVs72WJ5z + rm /tmp/tmp.icFzY1sPeX /tmp/tmp.aZVs72WJ5z + return 0 + echo + echo 'Deleting crashing pod some-name-rs0-2' Deleting crashing pod some-name-rs0-2 + kubectl_bin delete pod some-name-rs0-2 --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.zLIPMe1wRo ++ mktemp + local LAST_ERR=/tmp/tmp.JczD4ta3mg + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete pod some-name-rs0-2 --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.zLIPMe1wRo pod "some-name-rs0-2" deleted from demand-backup-if-unhealthy-355 namespace + cat /tmp/tmp.JczD4ta3mg + rm /tmp/tmp.zLIPMe1wRo /tmp/tmp.JczD4ta3mg + return 0 + wait_for_cluster_state some-name ready + local cluster_name=some-name + local target_state=ready + echo -n 'Waiting for psmdb/some-name to reach ready state' Waiting for psmdb/some-name to reach ready state+ local timeout=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Wm6PsetJPi +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZyeK7KVGQp ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Wm6PsetJPi ++ cat /tmp/tmp.ZyeK7KVGQp ++ rm /tmp/tmp.Wm6PsetJPi /tmp/tmp.ZyeK7KVGQp ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=1 + echo -n . .+ [[ 1 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mEEgzI9Em2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.GkVcCOeCmI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.mEEgzI9Em2 ++ cat /tmp/tmp.GkVcCOeCmI ++ rm /tmp/tmp.mEEgzI9Em2 /tmp/tmp.GkVcCOeCmI ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=2 + echo -n . .+ [[ 2 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nNkKPDs41k +++ mktemp ++ local LAST_ERR=/tmp/tmp.qbfnTrMEvm ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.nNkKPDs41k ++ cat /tmp/tmp.qbfnTrMEvm ++ rm /tmp/tmp.nNkKPDs41k /tmp/tmp.qbfnTrMEvm ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=3 + echo -n . .+ [[ 3 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GhRuUAsPPQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.X7q4yXWTvA ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.GhRuUAsPPQ ++ cat /tmp/tmp.X7q4yXWTvA ++ rm /tmp/tmp.GhRuUAsPPQ /tmp/tmp.X7q4yXWTvA ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=4 + echo -n . .+ [[ 4 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DyzJ2nI1id +++ mktemp ++ local LAST_ERR=/tmp/tmp.QYyeFoWpdQ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.DyzJ2nI1id ++ cat /tmp/tmp.QYyeFoWpdQ ++ rm /tmp/tmp.DyzJ2nI1id /tmp/tmp.QYyeFoWpdQ ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=5 + echo -n . .+ [[ 5 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sBJfdj86xs +++ mktemp ++ local LAST_ERR=/tmp/tmp.lBPzPFDupO ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.sBJfdj86xs ++ cat /tmp/tmp.lBPzPFDupO ++ rm /tmp/tmp.sBJfdj86xs /tmp/tmp.lBPzPFDupO ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=6 + echo -n . .+ [[ 6 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9PldC1SltB +++ mktemp ++ local LAST_ERR=/tmp/tmp.ciUcKBdTpd ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9PldC1SltB ++ cat /tmp/tmp.ciUcKBdTpd ++ rm /tmp/tmp.9PldC1SltB /tmp/tmp.ciUcKBdTpd ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=7 + echo -n . .+ [[ 7 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xEZd6pJg61 +++ mktemp ++ local LAST_ERR=/tmp/tmp.CeX1NR3Q99 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xEZd6pJg61 ++ cat /tmp/tmp.CeX1NR3Q99 ++ rm /tmp/tmp.xEZd6pJg61 /tmp/tmp.CeX1NR3Q99 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=8 + echo -n . .+ [[ 8 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.E2ftXVIqg9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.PXidrWeBKq ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.E2ftXVIqg9 ++ cat /tmp/tmp.PXidrWeBKq ++ rm /tmp/tmp.E2ftXVIqg9 /tmp/tmp.PXidrWeBKq ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=9 + echo -n . .+ [[ 9 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.80T6OSRQpE +++ mktemp ++ local LAST_ERR=/tmp/tmp.hauH6ddgcO ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.80T6OSRQpE ++ cat /tmp/tmp.hauH6ddgcO ++ rm /tmp/tmp.80T6OSRQpE /tmp/tmp.hauH6ddgcO ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=10 + echo -n . .+ [[ 10 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DJpbWTHdVw +++ mktemp ++ local LAST_ERR=/tmp/tmp.Vk92EmEtao ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.DJpbWTHdVw ++ cat /tmp/tmp.Vk92EmEtao ++ rm /tmp/tmp.DJpbWTHdVw /tmp/tmp.Vk92EmEtao ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=11 + echo -n . .+ [[ 11 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AH73tbHP4z +++ mktemp ++ local LAST_ERR=/tmp/tmp.q03htrtII0 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.AH73tbHP4z ++ cat /tmp/tmp.q03htrtII0 ++ rm /tmp/tmp.AH73tbHP4z /tmp/tmp.q03htrtII0 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=12 + echo -n . .+ [[ 12 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lSlh68IkTy +++ mktemp ++ local LAST_ERR=/tmp/tmp.IU3DM7Iczj ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.lSlh68IkTy ++ cat /tmp/tmp.IU3DM7Iczj ++ rm /tmp/tmp.lSlh68IkTy /tmp/tmp.IU3DM7Iczj ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=13 + echo -n . .+ [[ 13 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.t5Cx4YA6PL +++ mktemp ++ local LAST_ERR=/tmp/tmp.FnOpVKvJwW ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.t5Cx4YA6PL ++ cat /tmp/tmp.FnOpVKvJwW ++ rm /tmp/tmp.t5Cx4YA6PL /tmp/tmp.FnOpVKvJwW ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=14 + echo -n . .+ [[ 14 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NzJ27jCQjA +++ mktemp ++ local LAST_ERR=/tmp/tmp.XMcTmXeM8T ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.NzJ27jCQjA ++ cat /tmp/tmp.XMcTmXeM8T ++ rm /tmp/tmp.NzJ27jCQjA /tmp/tmp.XMcTmXeM8T ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=15 + echo -n . .+ [[ 15 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8PkkO1fz2L +++ mktemp ++ local LAST_ERR=/tmp/tmp.RZnSNLqnbx ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.8PkkO1fz2L ++ cat /tmp/tmp.RZnSNLqnbx ++ rm /tmp/tmp.8PkkO1fz2L /tmp/tmp.RZnSNLqnbx ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=16 + echo -n . .+ [[ 16 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1Du1xydoa1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.OhZcd4Cgpd ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.1Du1xydoa1 ++ cat /tmp/tmp.OhZcd4Cgpd ++ rm /tmp/tmp.1Du1xydoa1 /tmp/tmp.OhZcd4Cgpd ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=17 + echo -n . .+ [[ 17 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HASyTxFnfh +++ mktemp ++ local LAST_ERR=/tmp/tmp.7BNEB2nZmB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.HASyTxFnfh ++ cat /tmp/tmp.7BNEB2nZmB ++ rm /tmp/tmp.HASyTxFnfh /tmp/tmp.7BNEB2nZmB ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=18 + echo -n . .+ [[ 18 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.C8xjkk6BsW +++ mktemp ++ local LAST_ERR=/tmp/tmp.L7rLbkj6Z2 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.C8xjkk6BsW ++ cat /tmp/tmp.L7rLbkj6Z2 ++ rm /tmp/tmp.C8xjkk6BsW /tmp/tmp.L7rLbkj6Z2 ++ return 0 + [[ initializing =~ ready ]] + sleep 1 + timeout=19 + echo -n . .+ [[ 19 -gt 1500 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5ZkorAcOVz +++ mktemp ++ local LAST_ERR=/tmp/tmp.qW4MQa0N4S ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5ZkorAcOVz ++ cat /tmp/tmp.qW4MQa0N4S ++ rm /tmp/tmp.5ZkorAcOVz /tmp/tmp.qW4MQa0N4S ++ return 0 + [[ ready =~ ready ]] + echo + log 'psmdb/some-name is ready: OK' + set +o xtrace [2026-03-10T06:59:30+0000] psmdb/some-name is ready: OK + desc 'Delete data and run restore of backup-minio-logical' + set +o xtrace ----------------------------------------------------------------------------------- Delete data and run restore of backup-minio-logical ----------------------------------------------------------------------------------- + delete_data 'This is some data for backup if unhealthy' + local 'data=This is some data for backup if unhealthy' + run_mongo 'use myApp\n db.test.deleteOne({ x: "This is some data for backup if unhealthy" })' myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-355 + local 'command=use myApp\n db.test.deleteOne({ x: "This is some data for backup if unhealthy" })' + local uri=myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-355 + local driver=mongodb+srv + local suffix=.svc.cluster.local + local mongo_flag= + local replica_set=rs0 + [[ myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-355 == *cfg* ]] ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.T7JBpvQKam +++ mktemp ++ local LAST_ERR=/tmp/tmp.ShyRI2xYeh ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.T7JBpvQKam ++ cat /tmp/tmp.ShyRI2xYeh ++ rm /tmp/tmp.T7JBpvQKam /tmp/tmp.ShyRI2xYeh ++ return 0 + local client_container=psmdb-client-bb8b97679-wx99q + kubectl_bin exec psmdb-client-bb8b97679-wx99q -- bash -c 'printf '\''use myApp\n db.test.deleteOne({ x: "This is some data for backup if unhealthy" })\n'\'' | mongo mongodb+srv://myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-355.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.x2oOi9j8oZ ++ mktemp + local LAST_ERR=/tmp/tmp.yQmVCe24fY + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-wx99q -- bash -c 'printf '\''use myApp\n db.test.deleteOne({ x: "This is some data for backup if unhealthy" })\n'\'' | mongo mongodb+srv://myApp:myPass@some-name-rs0.demand-backup-if-unhealthy-355.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.x2oOi9j8oZ Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-rs0-2.some-name-rs0.demand-backup-if-unhealthy-355.svc.cluster.local:27017,some-name-rs0-0.some-name-rs0.demand-backup-if-unhealthy-355.svc.cluster.local:27017,some-name-rs0-1.some-name-rs0.demand-backup-if-unhealthy-355.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb&replicaSet=rs0&ssl=false Implicit session: session { "id" : UUID("65bf56a4-5f72-46b2-ba13-92484e600125") } Percona Server for MongoDB server version: v8.0.19-7 WARNING: shell and server versions do not match switched to db myApp { "acknowledged" : true, "deletedCount" : 1 } bye + cat /tmp/tmp.yQmVCe24fY + rm /tmp/tmp.x2oOi9j8oZ /tmp/tmp.yQmVCe24fY + return 0 + run_restore backup-minio-logical + local backup_name=backup-minio-logical + local restore_file=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/demand-backup-if-unhealthy/conf/restore.yml + log 'running restore restore-backup-minio-logical' + set +o xtrace [2026-03-10T06:59:32+0000] running restore restore-backup-minio-logical + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/demand-backup-if-unhealthy/conf/restore.yml + /usr/sbin/sed -e 's/name:/name: restore-backup-minio-logical/' + /usr/sbin/sed -e 's/backupName:/backupName: backup-minio-logical/' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.uP9tpuD7X8 ++ mktemp + local LAST_ERR=/tmp/tmp.QftlI4xsf3 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.uP9tpuD7X8 perconaservermongodbrestore.psmdb.percona.com/restore-backup-minio-logical created + cat /tmp/tmp.QftlI4xsf3 + rm /tmp/tmp.uP9tpuD7X8 /tmp/tmp.QftlI4xsf3 + return 0 + wait_restore backup-minio-logical some-name + local backup_name=backup-minio-logical + local cluster_name=some-name + local target_state=ready + local wait_cluster_consistency=1 + local wait_time=1780 + local ok_if_ready=0 + set +o xtrace Waiting for the psmdb-restore/restore-backup-minio-logical object to be created.OK Waiting psmdb-restore/restore-backup-minio-logical to reach state "ready" .OK after 0 minutes + [[ 1 -eq 1 ]] + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qiqLEp8A1n +++ mktemp ++ local LAST_ERR=/tmp/tmp.Jh59WUGrk5 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.qiqLEp8A1n ++ cat /tmp/tmp.Jh59WUGrk5 ++ rm /tmp/tmp.qiqLEp8A1n /tmp/tmp.Jh59WUGrk5 ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo .OK .OK + wait_for_cluster_state some-name ready + local cluster_name=some-name + local target_state=ready + echo -n 'Waiting for psmdb/some-name to reach ready state' Waiting for psmdb/some-name to reach ready state+ local timeout=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Y4GNA0qID0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Db6SQQHW7y ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Y4GNA0qID0 ++ cat /tmp/tmp.Db6SQQHW7y ++ rm /tmp/tmp.Y4GNA0qID0 /tmp/tmp.Db6SQQHW7y ++ return 0 + [[ ready =~ ready ]] + echo + log 'psmdb/some-name is ready: OK' + set +o xtrace [2026-03-10T07:00:06+0000] psmdb/some-name is ready: OK + desc 'Verify data is back to the cluster' + set +o xtrace ----------------------------------------------------------------------------------- Verify data is back to the cluster ----------------------------------------------------------------------------------- + simple_data_check some-name 3 0 -rs0 + local cluster_name=some-name + let last_pod=3-1 + local isSharded=0 + local cluster_pfx=-rs0 + '[' 0 -eq 1 ']' ++ seq 0 2 + for i in $(seq 0 $last_pod) + compare_mongo_cmd find myApp:myPass@some-name-rs0-0.some-name-rs0.demand-backup-if-unhealthy-355 + local command=find + local uri=myApp:myPass@some-name-rs0-0.some-name-rs0.demand-backup-if-unhealthy-355 + local postfix= + local suffix= + local database=myApp + local collection=test + local sort= + local tls=false + local replicaset= + local 'full_command=db.test.find()' + [[ -n '' ]] + log 'running db.test.find() in myApp' + set +o xtrace [2026-03-10T07:00:06+0000] running db.test.find() in myApp + [[ false == \t\r\u\e ]] + mongo_command=run_mongo + run_mongo 'use myApp\n db.test.find()' myApp:myPass@some-name-rs0-0.some-name-rs0.demand-backup-if-unhealthy-355 mongodb '' '' + local 'command=use myApp\n db.test.find()' + local uri=myApp:myPass@some-name-rs0-0.some-name-rs0.demand-backup-if-unhealthy-355 + local driver=mongodb + local suffix=.svc.cluster.local + local mongo_flag= + local replica_set=rs0 + [[ myApp:myPass@some-name-rs0-0.some-name-rs0.demand-backup-if-unhealthy-355 == *cfg* ]] + grep -E -v 'I NETWORK|W NETWORK|F NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:|Started a new thread for the timer service' + /usr/sbin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OdtpvKa8wj +++ mktemp ++ local LAST_ERR=/tmp/tmp.Nk75TIaa49 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.OdtpvKa8wj ++ cat /tmp/tmp.Nk75TIaa49 ++ rm /tmp/tmp.OdtpvKa8wj /tmp/tmp.Nk75TIaa49 ++ return 0 + local client_container=psmdb-client-bb8b97679-wx99q + kubectl_bin exec psmdb-client-bb8b97679-wx99q -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-rs0-0.some-name-rs0.demand-backup-if-unhealthy-355.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.0DnCZTs96M ++ mktemp + local LAST_ERR=/tmp/tmp.W6fTf7jJ7Q + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-wx99q -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-rs0-0.some-name-rs0.demand-backup-if-unhealthy-355.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.0DnCZTs96M + cat /tmp/tmp.W6fTf7jJ7Q + rm /tmp/tmp.0DnCZTs96M /tmp/tmp.W6fTf7jJ7Q + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/demand-backup-if-unhealthy/compare/find.json /tmp/tmp.gaQrVt1nmF/find + for i in $(seq 0 $last_pod) + compare_mongo_cmd find myApp:myPass@some-name-rs0-1.some-name-rs0.demand-backup-if-unhealthy-355 + local command=find + local uri=myApp:myPass@some-name-rs0-1.some-name-rs0.demand-backup-if-unhealthy-355 + local postfix= + local suffix= + local database=myApp + local collection=test + local sort= + local tls=false + local replicaset= + local 'full_command=db.test.find()' + [[ -n '' ]] + log 'running db.test.find() in myApp' + set +o xtrace [2026-03-10T07:00:09+0000] running db.test.find() in myApp + [[ false == \t\r\u\e ]] + mongo_command=run_mongo + run_mongo 'use myApp\n db.test.find()' myApp:myPass@some-name-rs0-1.some-name-rs0.demand-backup-if-unhealthy-355 mongodb '' '' + local 'command=use myApp\n db.test.find()' + local uri=myApp:myPass@some-name-rs0-1.some-name-rs0.demand-backup-if-unhealthy-355 + local driver=mongodb + local suffix=.svc.cluster.local + local mongo_flag= + grep -E -v 'I NETWORK|W NETWORK|F NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:|Started a new thread for the timer service' + local replica_set=rs0 + [[ myApp:myPass@some-name-rs0-1.some-name-rs0.demand-backup-if-unhealthy-355 == *cfg* ]] + /usr/sbin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2oKhc6vLdr +++ mktemp ++ local LAST_ERR=/tmp/tmp.N6Qn5fB6EX ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.2oKhc6vLdr ++ cat /tmp/tmp.N6Qn5fB6EX ++ rm /tmp/tmp.2oKhc6vLdr /tmp/tmp.N6Qn5fB6EX ++ return 0 + local client_container=psmdb-client-bb8b97679-wx99q + kubectl_bin exec psmdb-client-bb8b97679-wx99q -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-rs0-1.some-name-rs0.demand-backup-if-unhealthy-355.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.l2r6CCBwUb ++ mktemp + local LAST_ERR=/tmp/tmp.JYFtwIlAnY + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-wx99q -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-rs0-1.some-name-rs0.demand-backup-if-unhealthy-355.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.l2r6CCBwUb + cat /tmp/tmp.JYFtwIlAnY + rm /tmp/tmp.l2r6CCBwUb /tmp/tmp.JYFtwIlAnY + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/demand-backup-if-unhealthy/compare/find.json /tmp/tmp.gaQrVt1nmF/find + for i in $(seq 0 $last_pod) + compare_mongo_cmd find myApp:myPass@some-name-rs0-2.some-name-rs0.demand-backup-if-unhealthy-355 + local command=find + local uri=myApp:myPass@some-name-rs0-2.some-name-rs0.demand-backup-if-unhealthy-355 + local postfix= + local suffix= + local database=myApp + local collection=test + local sort= + local tls=false + local replicaset= + local 'full_command=db.test.find()' + [[ -n '' ]] + log 'running db.test.find() in myApp' + set +o xtrace [2026-03-10T07:00:11+0000] running db.test.find() in myApp + [[ false == \t\r\u\e ]] + mongo_command=run_mongo + run_mongo 'use myApp\n db.test.find()' myApp:myPass@some-name-rs0-2.some-name-rs0.demand-backup-if-unhealthy-355 mongodb '' '' + local 'command=use myApp\n db.test.find()' + local uri=myApp:myPass@some-name-rs0-2.some-name-rs0.demand-backup-if-unhealthy-355 + local driver=mongodb + local suffix=.svc.cluster.local + local mongo_flag= + local replica_set=rs0 + [[ myApp:myPass@some-name-rs0-2.some-name-rs0.demand-backup-if-unhealthy-355 == *cfg* ]] + grep -E -v 'I NETWORK|W NETWORK|F NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:|Started a new thread for the timer service' + /usr/sbin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.iECQoMGa6W +++ mktemp ++ local LAST_ERR=/tmp/tmp.1wtkk7ImNJ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.iECQoMGa6W ++ cat /tmp/tmp.1wtkk7ImNJ ++ rm /tmp/tmp.iECQoMGa6W /tmp/tmp.1wtkk7ImNJ ++ return 0 + local client_container=psmdb-client-bb8b97679-wx99q + kubectl_bin exec psmdb-client-bb8b97679-wx99q -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-rs0-2.some-name-rs0.demand-backup-if-unhealthy-355.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.PENURhKdVt ++ mktemp + local LAST_ERR=/tmp/tmp.lqq6Y9sxty + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-wx99q -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-rs0-2.some-name-rs0.demand-backup-if-unhealthy-355.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.PENURhKdVt + cat /tmp/tmp.lqq6Y9sxty + rm /tmp/tmp.PENURhKdVt /tmp/tmp.lqq6Y9sxty + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/demand-backup-if-unhealthy/compare/find.json /tmp/tmp.gaQrVt1nmF/find + destroy demand-backup-if-unhealthy-355 + local namespace=demand-backup-if-unhealthy-355 + local ignore_logs=true + [[ 0 == 1 ]] + desc 'destroy cluster/operator and all other resources' + set +o xtrace ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + '[' true == false ']' + delete_backups + desc 'Delete psmdb-backup' + set +o xtrace ----------------------------------------------------------------------------------- Delete psmdb-backup ----------------------------------------------------------------------------------- ++ kubectl_bin get psmdb-backup --no-headers ++ wc -l +++ mktemp ++ local LAST_OUT=/tmp/tmp.m2NRZEjhCG +++ mktemp ++ local LAST_ERR=/tmp/tmp.CXktp0hEln ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb-backup --no-headers ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.m2NRZEjhCG ++ cat /tmp/tmp.CXktp0hEln ++ rm /tmp/tmp.m2NRZEjhCG /tmp/tmp.CXktp0hEln ++ return 0 + '[' 1 '!=' 0 ']' + kubectl_bin get psmdb-backup ++ mktemp + local LAST_OUT=/tmp/tmp.VOwIY8imz6 ++ mktemp + local LAST_ERR=/tmp/tmp.6agT2anmLf + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get psmdb-backup + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.VOwIY8imz6 NAME CLUSTER STORAGE DESTINATION TYPE SIZE STATUS COMPLETED AGE backup-minio-logical some-name minio s3://operator-testing/2026-03-10T06:57:38Z logical 57.77KB ready 2m13s 2m38s + cat /tmp/tmp.6agT2anmLf + rm /tmp/tmp.VOwIY8imz6 /tmp/tmp.6agT2anmLf + return 0 + kubectl_bin delete psmdb-backup --all ++ mktemp + local LAST_OUT=/tmp/tmp.1Te8pqo569 ++ mktemp + local LAST_ERR=/tmp/tmp.oHZTE2kawU + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete psmdb-backup --all + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.1Te8pqo569 perconaservermongodbbackup.psmdb.percona.com "backup-minio-logical" deleted from demand-backup-if-unhealthy-355 namespace + cat /tmp/tmp.oHZTE2kawU + rm /tmp/tmp.1Te8pqo569 /tmp/tmp.oHZTE2kawU + return 0 + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.ddE76PepLr ++ mktemp + local LAST_ERR=/tmp/tmp.gulCMdIG2N + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ddE76PepLr customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.gulCMdIG2N + rm /tmp/tmp.ddE76PepLr /tmp/tmp.gulCMdIG2N + return 0 ++ grep -v '\-\-\-' ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/crd.yaml grep: warning: stray \ before - grep: warning: stray \ before - + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.RAffU4DdOu ++ mktemp + local LAST_ERR=/tmp/tmp.zLFQeul1Xx + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.RAffU4DdOu + cat /tmp/tmp.zLFQeul1Xx + rm /tmp/tmp.RAffU4DdOu /tmp/tmp.zLFQeul1Xx + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.ANArg5vtrP ++ mktemp + local LAST_ERR=/tmp/tmp.VM4plXcOKo + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ANArg5vtrP + cat /tmp/tmp.VM4plXcOKo + rm /tmp/tmp.ANArg5vtrP /tmp/tmp.VM4plXcOKo + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' No resources found + kubectl patch perconaservermongodbs.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbs" + : + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.izK3ULQGBF ++ mktemp + local LAST_ERR=/tmp/tmp.zBDMYNtgXc + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.izK3ULQGBF + cat /tmp/tmp.zBDMYNtgXc + rm /tmp/tmp.izK3ULQGBF /tmp/tmp.zBDMYNtgXc + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.utB7ZLnfLM ++ mktemp + local LAST_ERR=/tmp/tmp.AlinVOP7dV + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.utB7ZLnfLM clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.AlinVOP7dV + rm /tmp/tmp.utB7ZLnfLM /tmp/tmp.AlinVOP7dV + return 0 + destroy_cert_manager + kubectl_bin delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.2wsBfVz59G ++ mktemp + local LAST_ERR=/tmp/tmp.UBpTHFU06j + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.2wsBfVz59G + cat /tmp/tmp.UBpTHFU06j Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 0 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.2wsBfVz59G + cat /tmp/tmp.UBpTHFU06j Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 4 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.2wsBfVz59G + cat /tmp/tmp.UBpTHFU06j Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 8 + cat /tmp/tmp.2wsBfVz59G + cat /tmp/tmp.UBpTHFU06j Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + rm /tmp/tmp.2wsBfVz59G /tmp/tmp.UBpTHFU06j + return 1 + true + '[' -n '' ']' + '[' -n psmdb-operator ']' + kubectl_bin delete --grace-period=0 --force=true namespace demand-backup-if-unhealthy-355 + rm -rf /tmp/tmp.gaQrVt1nmF + kubectl_bin delete --grace-period=0 --force=true namespace psmdb-operator ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.osqGwcfE60 + desc 'test passed' + set +o xtrace ----------------------------------------------------------------------------------- test passed ----------------------------------------------------------------------------------- + local LAST_OUT=/tmp/tmp.DMKFAqq7YM ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.NxNxTx71OZ + local exit_status=0 + local timeout=4 ++ seq 0 2 + local LAST_ERR=/tmp/tmp.kpze1tkjWe + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete --grace-period=0 --force=true namespace psmdb-operator + for i in $(seq 0 2) + set +e + kubectl delete --grace-period=0 --force=true namespace demand-backup-if-unhealthy-355