Log: /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/e2e-tests/logs/pvc-resize.log Warning: version difference between client (1.36) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.36) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.36) and server (1.32) exceeds the supported minor version skew of +/-1 + '[' 0 == 1 ']' + ensure_default_sc_allows_expansion + local default_sc ++ get_default_storageclass ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8c1IQH9WhM +++ mktemp ++ local LAST_ERR=/tmp/tmp.NofBozZ5k4 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.8c1IQH9WhM ++ cat /tmp/tmp.NofBozZ5k4 ++ rm /tmp/tmp.8c1IQH9WhM /tmp/tmp.NofBozZ5k4 ++ return 0 + default_sc=standard-rwo + echo 'Checking if default storageclass standard-rwo allows volume expansion' Checking if default storageclass standard-rwo allows volume expansion + local allowVolumeExpansion ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.name=="standard-rwo")].allowVolumeExpansion}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.n5u4EuU6fd +++ mktemp ++ local LAST_ERR=/tmp/tmp.zhmHaqdzAT ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.name=="standard-rwo")].allowVolumeExpansion}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.n5u4EuU6fd ++ cat /tmp/tmp.zhmHaqdzAT ++ rm /tmp/tmp.n5u4EuU6fd /tmp/tmp.zhmHaqdzAT ++ return 0 + allowVolumeExpansion=true + [[ true != true ]] + create_infra pvc-resize-2243 + local ns=pvc-resize-2243 + [[ 1 == 1 ]] + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.lJAlSD7orA ++ mktemp + local LAST_ERR=/tmp/tmp.kT7ciH8NZ6 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.lJAlSD7orA customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.kT7ciH8NZ6 + rm /tmp/tmp.lJAlSD7orA /tmp/tmp.kT7ciH8NZ6 + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/deploy/crd.yaml ++ grep -v '\-\-\-' grep: warning: stray \ before - grep: warning: stray \ before - + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + grep -v NAMESPACE + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.CbLAROa25T ++ mktemp + local LAST_ERR=/tmp/tmp.gSFEQH8TLr + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.CbLAROa25T + cat /tmp/tmp.gSFEQH8TLr + rm /tmp/tmp.CbLAROa25T /tmp/tmp.gSFEQH8TLr + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.I6Y3Wg4qNC ++ mktemp + local LAST_ERR=/tmp/tmp.hdQsxuQs6m + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.I6Y3Wg4qNC + cat /tmp/tmp.hdQsxuQs6m + rm /tmp/tmp.I6Y3Wg4qNC /tmp/tmp.hdQsxuQs6m + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' No resources found + kubectl patch perconaservermongodbs.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbs" + : + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.MiADqEncgm ++ mktemp + local LAST_ERR=/tmp/tmp.ZoTuUW0WRY + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.MiADqEncgm + cat /tmp/tmp.ZoTuUW0WRY + rm /tmp/tmp.MiADqEncgm /tmp/tmp.ZoTuUW0WRY + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.WzFo3NUqdF ++ mktemp + local LAST_ERR=/tmp/tmp.jKRL3gfGTT + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.WzFo3NUqdF clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.jKRL3gfGTT + rm /tmp/tmp.WzFo3NUqdF /tmp/tmp.jKRL3gfGTT + return 0 + check_crd_for_deletion PR-2274-acb3b334 + local git_tag=PR-2274-acb3b334 ++ curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/PR-2274-acb3b334/deploy/crd.yaml ++ yq eval .metadata.name ++ /usr/sbin/sed s/---//g ++ /usr/sbin/sed ':a;N;$!ba;s/\n/ /g' + for crd_name in $(curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/${git_tag}/deploy/crd.yaml | yq eval '.metadata.name' | $sed 's/---//g' | $sed ':a;N;$!ba;s/\n/ /g') ++ kubectl_bin get crd/null -o 'jsonpath={.status.conditions[-1].type}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NSkKBp8EPa +++ mktemp ++ local LAST_ERR=/tmp/tmp.D95ulpQru8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.NSkKBp8EPa ++ cat /tmp/tmp.D95ulpQru8 Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 0 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.NSkKBp8EPa ++ cat /tmp/tmp.D95ulpQru8 Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 4 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.NSkKBp8EPa ++ cat /tmp/tmp.D95ulpQru8 Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 8 ++ cat /tmp/tmp.NSkKBp8EPa ++ cat /tmp/tmp.D95ulpQru8 Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ rm /tmp/tmp.NSkKBp8EPa /tmp/tmp.D95ulpQru8 ++ return 1 + [[ '' == Terminating ]] + '[' -n psmdb-operator ']' + create_namespace psmdb-operator + local namespace=psmdb-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + '[' -n '' ']' + awk '{print$1}' + desc 'cleaned up old namespaces psmdb-operator' + grep -E -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace psmdb-operator --ignore-not-found + xargs kubectl delete ns ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.Ottqhz6S8r + local LAST_OUT=/tmp/tmp.twKuKWwTkl ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.oj3Mnrz18y + local exit_status=0 + local timeout=4 + local LAST_ERR=/tmp/tmp.WYy9ZLhCRl + local exit_status=0 + local timeout=4 ++ seq 0 2 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get ns + for i in $(seq 0 2) + set +e + kubectl delete namespace psmdb-operator --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Ottqhz6S8r + cat /tmp/tmp.oj3Mnrz18y + rm /tmp/tmp.Ottqhz6S8r /tmp/tmp.oj3Mnrz18y + return 0 namespace "pvc-resize-17725" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.twKuKWwTkl namespace "psmdb-operator" deleted + cat /tmp/tmp.WYy9ZLhCRl + rm /tmp/tmp.twKuKWwTkl /tmp/tmp.WYy9ZLhCRl + return 0 + kubectl_bin wait --for=delete namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.aAcDoN4jEd ++ mktemp + local LAST_ERR=/tmp/tmp.mvo8vy8Ywv + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.aAcDoN4jEd + cat /tmp/tmp.mvo8vy8Ywv + rm /tmp/tmp.aAcDoN4jEd /tmp/tmp.mvo8vy8Ywv + return 0 + desc 'create namespace psmdb-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.fD2iAMi46Y ++ mktemp + local LAST_ERR=/tmp/tmp.DF59H7h0mB + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.fD2iAMi46Y namespace/psmdb-operator created + cat /tmp/tmp.DF59H7h0mB + rm /tmp/tmp.fD2iAMi46Y /tmp/tmp.DF59H7h0mB + return 0 + set_kube_ctx psmdb-operator + local namespace=psmdb-operator ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.OjhEMMDniI +++ mktemp ++ local LAST_ERR=/tmp/tmp.qZ2sPYzmSt ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.OjhEMMDniI ++ cat /tmp/tmp.qZ2sPYzmSt ++ rm /tmp/tmp.OjhEMMDniI /tmp/tmp.qZ2sPYzmSt ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2274-acb3b334-6-cluster6 --namespace=psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.jbJOmQilFd ++ mktemp + local LAST_ERR=/tmp/tmp.4H0pK9X4ZL + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2274-acb3b334-6-cluster6 --namespace=psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.jbJOmQilFd Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2274-acb3b334-6-cluster6" modified. + cat /tmp/tmp.4H0pK9X4ZL + rm /tmp/tmp.jbJOmQilFd /tmp/tmp.4H0pK9X4ZL + return 0 + deploy_operator + desc 'start PSMDB operator: docker.io/perconalab/percona-server-mongodb-operator:PR-2274-acb3b334' + set +o xtrace ----------------------------------------------------------------------------------- start PSMDB operator: docker.io/perconalab/percona-server-mongodb-operator:PR-2274-acb3b334 ----------------------------------------------------------------------------------- + local cr_file + '[' -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/e2e-tests/pvc-resize/conf/crd.yaml ']' + cr_file=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/deploy/crd.yaml + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.PrihvJiJ5t ++ mktemp + local LAST_ERR=/tmp/tmp.HoaH9ho5PU + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.PrihvJiJ5t customresourcedefinition.apiextensions.k8s.io/perconaservermongodbbackups.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbrestores.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com serverside-applied + cat /tmp/tmp.HoaH9ho5PU + rm /tmp/tmp.PrihvJiJ5t /tmp/tmp.HoaH9ho5PU + return 0 + '[' -n psmdb-operator ']' + apply_rbac cw-rbac + local operator_namespace=psmdb-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: psmdb-operator^' + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.kKRq4WZY6k ++ mktemp + local LAST_ERR=/tmp/tmp.KzeFeEBdIJ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.kKRq4WZY6k clusterrole.rbac.authorization.k8s.io/percona-server-mongodb-operator created serviceaccount/percona-server-mongodb-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-server-mongodb-operator created + cat /tmp/tmp.KzeFeEBdIJ + rm /tmp/tmp.kKRq4WZY6k /tmp/tmp.KzeFeEBdIJ + return 0 + yq eval $'\n\t\t\t(.spec.template.spec.containers[].image = "docker.io/perconalab/percona-server-mongodb-operator:PR-2274-acb3b334") |\n\t\t\t((.. | select(.[] == "DISABLE_TELEMETRY")) |= .value="true") |\n\t\t\t((.. | select(.[] == "LOG_LEVEL")) |= .value="DEBUG")' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/deploy/cw-operator.yaml + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.dWbid9HDVe ++ mktemp + local LAST_ERR=/tmp/tmp.sZQDL8pnir + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.dWbid9HDVe deployment.apps/percona-server-mongodb-operator created + cat /tmp/tmp.sZQDL8pnir + rm /tmp/tmp.dWbid9HDVe /tmp/tmp.sZQDL8pnir + return 0 + sleep 20 ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.yp1d04tZv9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.5teMxMKcZT ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.yp1d04tZv9 ++ cat /tmp/tmp.5teMxMKcZT ++ rm /tmp/tmp.yp1d04tZv9 /tmp/tmp.5teMxMKcZT ++ return 0 + wait_operator_pod percona-server-mongodb-operator-578bbff7d7-2688j + local pod=percona-server-mongodb-operator-578bbff7d7-2688j + set +o xtrace waiting for pod/percona-server-mongodb-operator-578bbff7d7-2688j to be ready.OK + echo 'Print operator info from log' Print operator info from log + grep 'Manager starting up' ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.0Q9EYwj7RO +++ mktemp ++ local LAST_ERR=/tmp/tmp.btxZdK84eh ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.0Q9EYwj7RO ++ cat /tmp/tmp.btxZdK84eh ++ rm /tmp/tmp.0Q9EYwj7RO /tmp/tmp.btxZdK84eh ++ return 0 + kubectl_bin logs -n psmdb-operator percona-server-mongodb-operator-578bbff7d7-2688j ++ mktemp + local LAST_OUT=/tmp/tmp.6z0R2xccG0 ++ mktemp + local LAST_ERR=/tmp/tmp.4OhwyRf3bi + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl logs -n psmdb-operator percona-server-mongodb-operator-578bbff7d7-2688j + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.6z0R2xccG0 + cat /tmp/tmp.4OhwyRf3bi + rm /tmp/tmp.6z0R2xccG0 /tmp/tmp.4OhwyRf3bi + return 0 2026-04-23T09:54:39.106Z INFO setup Manager starting up {"gitCommit": "acb3b334b50f3e35633eb97dbf38451e80253352", "gitBranch": "PR-2274-acb3b334", "buildTime": "", "goVersion": "go1.25.9", "os": "linux", "arch": "amd64"} + create_namespace pvc-resize-2243 + local namespace=pvc-resize-2243 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + grep -E -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' + kubectl_bin get ns + '[' -n '' ']' + desc 'cleaned up old namespaces pvc-resize-2243' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pvc-resize-2243 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pvc-resize-2243 --ignore-not-found + awk '{print$1}' + xargs kubectl delete ns ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.dImpyjmhDz ++ mktemp + local LAST_OUT=/tmp/tmp.y2FGIHHjk9 + local LAST_ERR=/tmp/tmp.wMZvBYuwZh + local exit_status=0 + local timeout=4 ++ mktemp ++ seq 0 2 + local LAST_ERR=/tmp/tmp.WnOGYVBfDg + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete namespace pvc-resize-2243 --ignore-not-found + for i in $(seq 0 2) + set +e + kubectl get ns + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.y2FGIHHjk9 + cat /tmp/tmp.WnOGYVBfDg + rm /tmp/tmp.y2FGIHHjk9 /tmp/tmp.WnOGYVBfDg + return 0 error: resource(s) were provided, but no name was specified + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.dImpyjmhDz + cat /tmp/tmp.wMZvBYuwZh + rm /tmp/tmp.dImpyjmhDz /tmp/tmp.wMZvBYuwZh + return 0 + kubectl_bin wait --for=delete namespace pvc-resize-2243 ++ mktemp + local LAST_OUT=/tmp/tmp.ZlF3ZUiuAk ++ mktemp + local LAST_ERR=/tmp/tmp.hnoWfYxjWs + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete namespace pvc-resize-2243 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ZlF3ZUiuAk + cat /tmp/tmp.hnoWfYxjWs + rm /tmp/tmp.ZlF3ZUiuAk /tmp/tmp.hnoWfYxjWs + return 0 + desc 'create namespace pvc-resize-2243' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pvc-resize-2243 ----------------------------------------------------------------------------------- + kubectl_bin create namespace pvc-resize-2243 ++ mktemp + local LAST_OUT=/tmp/tmp.dYAcRcY3Uo ++ mktemp + local LAST_ERR=/tmp/tmp.nd8Dt5CkKK + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace pvc-resize-2243 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.dYAcRcY3Uo namespace/pvc-resize-2243 created + cat /tmp/tmp.nd8Dt5CkKK + rm /tmp/tmp.dYAcRcY3Uo /tmp/tmp.nd8Dt5CkKK + return 0 + set_kube_ctx pvc-resize-2243 + local namespace=pvc-resize-2243 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.Lt2enHwF6q +++ mktemp ++ local LAST_ERR=/tmp/tmp.ktwnl4o223 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Lt2enHwF6q ++ cat /tmp/tmp.ktwnl4o223 ++ rm /tmp/tmp.Lt2enHwF6q /tmp/tmp.ktwnl4o223 ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2274-acb3b334-6-cluster6 --namespace=pvc-resize-2243 ++ mktemp + local LAST_OUT=/tmp/tmp.VP42cVz2qm ++ mktemp + local LAST_ERR=/tmp/tmp.wEZIcZ9ajr + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2274-acb3b334-6-cluster6 --namespace=pvc-resize-2243 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.VP42cVz2qm Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2274-acb3b334-6-cluster6" modified. + cat /tmp/tmp.wEZIcZ9ajr + rm /tmp/tmp.VP42cVz2qm /tmp/tmp.wEZIcZ9ajr + return 0 + desc 'create secrets and psmdb client' + set +o xtrace ----------------------------------------------------------------------------------- create secrets and psmdb client ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/e2e-tests/conf/secrets.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.ZOMQ5sipou ++ mktemp + local LAST_ERR=/tmp/tmp.atcL6pDWAu + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/e2e-tests/conf/secrets.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/e2e-tests/conf/client.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ZOMQ5sipou secret/some-users created deployment.apps/psmdb-client created + cat /tmp/tmp.atcL6pDWAu + rm /tmp/tmp.ZOMQ5sipou /tmp/tmp.atcL6pDWAu + return 0 + desc 'create PSMDB cluster' + set +o xtrace ----------------------------------------------------------------------------------- create PSMDB cluster ----------------------------------------------------------------------------------- + cluster=some-name + '[' 0 == 1 ']' + spinup_psmdb some-name-rs0 /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/e2e-tests/pvc-resize/conf/some-name.yml + local cluster=some-name-rs0 + local config=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/e2e-tests/pvc-resize/conf/some-name.yml + local size=3 + desc 'create first PSMDB cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PSMDB cluster ----------------------------------------------------------------------------------- + apply_cluster /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/e2e-tests/pvc-resize/conf/some-name.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/e2e-tests/pvc-resize/conf/some-name.yml + kubectl_bin apply -f - + yq eval '(.spec | select(has("pmm"))).pmm.image = "docker.io/percona/pmm-client:2.44.1-1"' + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/e2e-tests/pvc-resize/conf/some-name.yml + yq eval '(.spec | select(has("initImage"))).initImage = "docker.io/perconalab/percona-server-mongodb-operator:PR-2274-acb3b334"' + yq eval '(.spec | select(.image == null)).image = "docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0"' ++ mktemp + yq eval '(.spec | select(has("backup"))).backup.image = "docker.io/perconalab/percona-server-mongodb-operator:main-backup"' + yq eval '.spec.upgradeOptions.apply="Never"' + /usr/sbin/sed -e s/NAME_SPACE/pvc-resize-2243/g + local LAST_OUT=/tmp/tmp.LCcJMr1yP7 ++ mktemp + local LAST_ERR=/tmp/tmp.VSS2fzNUUQ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.LCcJMr1yP7 perconaservermongodb.psmdb.percona.com/some-name created + cat /tmp/tmp.VSS2fzNUUQ + rm /tmp/tmp.LCcJMr1yP7 /tmp/tmp.VSS2fzNUUQ + return 0 + desc 'check if Pod is started' + set +o xtrace ----------------------------------------------------------------------------------- check if Pod is started ----------------------------------------------------------------------------------- + wait_for_running some-name-rs0 3 + local name=some-name-rs0 + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=rs0 + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-rs0-0 + local pod=some-name-rs0-0 + set +o xtrace waiting for pod/some-name-rs0-0 to be ready.............OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-rs0-1 + local pod=some-name-rs0-1 + set +o xtrace waiting for pod/some-name-rs0-1 to be ready.............OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.a291TPQuj6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.jhryfzenug ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.a291TPQuj6 ++ cat /tmp/tmp.jhryfzenug ++ rm /tmp/tmp.a291TPQuj6 /tmp/tmp.jhryfzenug ++ return 0 + [[ '' == true ]] + wait_pod some-name-rs0-2 + local pod=some-name-rs0-2 + set +o xtrace waiting for pod/some-name-rs0-2 to be ready...........OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ENHuiZvx1v +++ mktemp ++ local LAST_ERR=/tmp/tmp.UIpHOOBouc ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ENHuiZvx1v ++ cat /tmp/tmp.UIpHOOBouc ++ rm /tmp/tmp.ENHuiZvx1v /tmp/tmp.UIpHOOBouc ++ return 0 + [[ '' == true ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9ULsfaElQH +++ mktemp ++ local LAST_ERR=/tmp/tmp.O0kYUg8AiM ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9ULsfaElQH ++ cat /tmp/tmp.O0kYUg8AiM ++ rm /tmp/tmp.9ULsfaElQH /tmp/tmp.O0kYUg8AiM ++ return 0 + [[ '' == true ]] + sleep 10 + [[ true == true ]] + set +x Waiting for cluster readyness + sleep 20 + compare_kubectl statefulset/some-name-rs0 + local resource=statefulset/some-name-rs0 + local postfix= + local skip_generation_check= + local expected_result=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/e2e-tests/pvc-resize/compare/statefulset_some-name-rs0.yml + local new_result=/tmp/tmp.GLPwH4gC3u/statefulset_some-name-rs0.yml + '[' -n '' -a -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/e2e-tests/pvc-resize/compare/statefulset_some-name-rs0-oc.yml ']' + kubectl_bin get -o yaml statefulset/some-name-rs0 + yq eval $'\n\t\t\tdel(.metadata.ownerReferences[].apiVersion) |\n\t\t\tdel(.metadata.managedFields) |\n\t\t\tdel(.. | select(has("creationTimestamp")).creationTimestamp) |\n\t\t\tdel(.. | select(has("namespace")).namespace) |\n\t\t\tdel(.. | select(has("uid")).uid) |\n\t\t\tdel(.metadata.resourceVersion) |\n\t\t\tdel(.spec.template.spec.containers[].env[] | select(.name == "NAMESPACE")) |\n\t\t\tdel(.metadata.selfLink) |\n\t\t\tdel(.metadata.annotations."cloud.google.com/neg") |\n\t\t\tdel(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") |\n\t\t\tdel(.. | select(has("image")).image) |\n\t\t\tdel(.. | select(has("clusterIP")).clusterIP) |\n\t\t\tdel(.. | select(has("clusterIPs")).clusterIPs) |\n\t\t\tdel(.. | select(has("dataSource")).dataSource) |\n\t\t\tdel(.. | select(has("procMount")).procMount) |\n\t\t\tdel(.. | select(has("storageClassName")).storageClassName) |\n\t\t\tdel(.. | select(has("finalizers")).finalizers) |\n\t\t\tdel(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") |\n\t\t\tdel(.. | select(has("volumeName")).volumeName) |\n\t\t\tdel(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") |\n\t\t\tdel(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") |\n\t\t\tdel(.spec.volumeMode) |\n\t\t\tdel(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") |\n\t\t\tdel(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") |\n\t\t\tdel(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") |\n\t\t\tdel(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") |\n\t\t\tdel(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") |\n\t\t\tdel(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) |\n\t\t\tdel(.. | select(has("healthCheckNodePort")).healthCheckNodePort) |\n\t\t\tdel(.. | select(has("nodePort")).nodePort) |\n\t\t\tdel(.status) |\n\t\t\t(.. | select(tag == "!!str")) |= sub("pvc-resize-2243", "NAME_SPACE") |\n\t\t\tdel(.spec.volumeClaimTemplates[].apiVersion) |\n\t\t\tdel(.spec.volumeClaimTemplates[].kind) |\n\t\t\tdel(.spec.ipFamilies) |\n\t\t\tdel(.spec.ipFamilyPolicy) |\n\t\t\t(.. | select(. == "extensions/v1beta1")) = "apps/v1" |\n\t\t\t(.. | select(. == "batch/v1beta1")) = "batch/v1" ' - ++ mktemp + local LAST_OUT=/tmp/tmp.BGqTTDrINT ++ mktemp + local LAST_ERR=/tmp/tmp.0CoQ0xiLrz + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get -o yaml statefulset/some-name-rs0 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.BGqTTDrINT + cat /tmp/tmp.0CoQ0xiLrz + rm /tmp/tmp.BGqTTDrINT /tmp/tmp.0CoQ0xiLrz + return 0 + yq -i eval 'del(.spec.persistentVolumeClaimRetentionPolicy)' /tmp/tmp.GLPwH4gC3u/statefulset_some-name-rs0.yml + version_gt 1.22 ++ echo '1.32 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + yq -i eval 'del(.spec.internalTrafficPolicy)' /tmp/tmp.GLPwH4gC3u/statefulset_some-name-rs0.yml + yq -i eval 'del(.spec.allocateLoadBalancerNodePorts)' /tmp/tmp.GLPwH4gC3u/statefulset_some-name-rs0.yml + [[ /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/e2e-tests/pvc-resize/compare/statefulset_some-name-rs0.yml == */cronjob* ]] + '[' -n '' ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/e2e-tests/pvc-resize/compare/statefulset_some-name-rs0.yml /tmp/tmp.GLPwH4gC3u/statefulset_some-name-rs0.yml + log 'compare_kubectl: statefulset/some-name-rs0 OK' + set +o xtrace [2026-04-23T09:57:12+0000] compare_kubectl: statefulset/some-name-rs0 OK + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + run_mongo 'db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})' userAdmin:userAdmin123456@some-name-rs0.pvc-resize-2243 + local 'command=db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})' + local uri=userAdmin:userAdmin123456@some-name-rs0.pvc-resize-2243 + local driver=mongodb+srv + local suffix=.svc.cluster.local + local mongo_flag= + local replica_set=rs0 + [[ userAdmin:userAdmin123456@some-name-rs0.pvc-resize-2243 == *cfg* ]] ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.C2S8LfFHmo +++ mktemp ++ local LAST_ERR=/tmp/tmp.jNcNOy1qsg ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.C2S8LfFHmo ++ cat /tmp/tmp.jNcNOy1qsg ++ rm /tmp/tmp.C2S8LfFHmo /tmp/tmp.jNcNOy1qsg ++ return 0 + local client_container=psmdb-client-bb8b97679-nsdcj + kubectl_bin exec psmdb-client-bb8b97679-nsdcj -- bash -c 'printf '\''db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})\n'\'' | mongo mongodb+srv://userAdmin:userAdmin123456@some-name-rs0.pvc-resize-2243.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.2R0RSQ4Cvk ++ mktemp + local LAST_ERR=/tmp/tmp.b5h58XwUbA + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-nsdcj -- bash -c 'printf '\''db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})\n'\'' | mongo mongodb+srv://userAdmin:userAdmin123456@some-name-rs0.pvc-resize-2243.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.2R0RSQ4Cvk Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-rs0-1.some-name-rs0.pvc-resize-2243.svc.cluster.local:27017,some-name-rs0-0.some-name-rs0.pvc-resize-2243.svc.cluster.local:27017,some-name-rs0-2.some-name-rs0.pvc-resize-2243.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb&replicaSet=rs0&ssl=false Implicit session: session { "id" : UUID("545bf5fc-b60c-4c70-abec-7cbdcac22a6b") } Percona Server for MongoDB server version: v8.0.20-8 WARNING: shell and server versions do not match Successfully added user: { "user" : "myApp", "roles" : [ { "db" : "myApp", "role" : "readWrite" } ] } bye + cat /tmp/tmp.b5h58XwUbA + rm /tmp/tmp.2R0RSQ4Cvk /tmp/tmp.b5h58XwUbA + return 0 + run_mongo 'use myApp\n db.test.insert({ x: 100500 })' myApp:myPass@some-name-rs0.pvc-resize-2243 + local 'command=use myApp\n db.test.insert({ x: 100500 })' + local uri=myApp:myPass@some-name-rs0.pvc-resize-2243 + local driver=mongodb+srv + local suffix=.svc.cluster.local + local mongo_flag= + local replica_set=rs0 + [[ myApp:myPass@some-name-rs0.pvc-resize-2243 == *cfg* ]] ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zTrm5jzFh4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.DIPYwCJqYk ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.zTrm5jzFh4 ++ cat /tmp/tmp.DIPYwCJqYk ++ rm /tmp/tmp.zTrm5jzFh4 /tmp/tmp.DIPYwCJqYk ++ return 0 + local client_container=psmdb-client-bb8b97679-nsdcj + kubectl_bin exec psmdb-client-bb8b97679-nsdcj -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb+srv://myApp:myPass@some-name-rs0.pvc-resize-2243.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.7ND8umWy4g ++ mktemp + local LAST_ERR=/tmp/tmp.ClJ4AGjauI + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-nsdcj -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb+srv://myApp:myPass@some-name-rs0.pvc-resize-2243.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.7ND8umWy4g Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-rs0-2.some-name-rs0.pvc-resize-2243.svc.cluster.local:27017,some-name-rs0-1.some-name-rs0.pvc-resize-2243.svc.cluster.local:27017,some-name-rs0-0.some-name-rs0.pvc-resize-2243.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb&replicaSet=rs0&ssl=false Implicit session: session { "id" : UUID("85cc1ba1-eef6-415b-9ebb-f9b9324de459") } Percona Server for MongoDB server version: v8.0.20-8 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.ClJ4AGjauI + rm /tmp/tmp.7ND8umWy4g /tmp/tmp.ClJ4AGjauI + return 0 + patch_pvc_request some-name 2G + local cluster=some-name + local size=2G + echo 'Patching PVC request to 2G in some-name' Patching PVC request to 2G in some-name + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"2G"}]' ++ mktemp + local LAST_OUT=/tmp/tmp.lthy2GwUZa ++ mktemp + local LAST_ERR=/tmp/tmp.KJ4hjckJ2d + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"2G"}]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.lthy2GwUZa perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.KJ4hjckJ2d + rm /tmp/tmp.lthy2GwUZa /tmp/tmp.KJ4hjckJ2d + return 0 + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.aXR6TWZuEU +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZKj7FXkyls ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.aXR6TWZuEU ++ cat /tmp/tmp.ZKj7FXkyls ++ rm /tmp/tmp.aXR6TWZuEU /tmp/tmp.ZKj7FXkyls ++ return 0 + [[ ready == ready ]] + echo .OK .OK + echo + wait_all_pvc_resize 2Gi 120 1 + local expected_size=2Gi + local max_retry=120 + local sleep_time=1 ++ kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name +++ mktemp ++ local LAST_OUT=/tmp/tmp.QGWbQvrqnv +++ mktemp ++ local LAST_ERR=/tmp/tmp.cQQ8nFzrLB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc -l app.kubernetes.io/component=mongod -o name ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.QGWbQvrqnv ++ cat /tmp/tmp.cQQ8nFzrLB ++ rm /tmp/tmp.QGWbQvrqnv /tmp/tmp.cQQ8nFzrLB ++ return 0 + for pvc in $(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name) + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-0 2Gi 120 1 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-0 + local expected_size=2Gi + local max_retry=120 + local sleep_time=1 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kjgqyC1o91 +++ mktemp ++ local LAST_ERR=/tmp/tmp.yEy3QUHnDE ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.kjgqyC1o91 ++ cat /tmp/tmp.yEy3QUHnDE ++ rm /tmp/tmp.kjgqyC1o91 /tmp/tmp.yEy3QUHnDE ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 0 -ge 120 ]] + echo -n . .+ sleep 1 + retry=1 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MqljMveINa +++ mktemp ++ local LAST_ERR=/tmp/tmp.6XAMo6hLVu ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.MqljMveINa ++ cat /tmp/tmp.6XAMo6hLVu ++ rm /tmp/tmp.MqljMveINa /tmp/tmp.6XAMo6hLVu ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 1 -ge 120 ]] + echo -n . .+ sleep 1 + retry=2 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SriboT6TmL +++ mktemp ++ local LAST_ERR=/tmp/tmp.ymBRPMrsiC ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.SriboT6TmL ++ cat /tmp/tmp.ymBRPMrsiC ++ rm /tmp/tmp.SriboT6TmL /tmp/tmp.ymBRPMrsiC ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 2 -ge 120 ]] + echo -n . .+ sleep 1 + retry=3 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jE7tXPznjv +++ mktemp ++ local LAST_ERR=/tmp/tmp.l1M2lc2W87 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jE7tXPznjv ++ cat /tmp/tmp.l1M2lc2W87 ++ rm /tmp/tmp.jE7tXPznjv /tmp/tmp.l1M2lc2W87 ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 3 -ge 120 ]] + echo -n . .+ sleep 1 + retry=4 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vQgeyQon3n +++ mktemp ++ local LAST_ERR=/tmp/tmp.McqO7la9LJ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.vQgeyQon3n ++ cat /tmp/tmp.McqO7la9LJ ++ rm /tmp/tmp.vQgeyQon3n /tmp/tmp.McqO7la9LJ ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 4 -ge 120 ]] + echo -n . .+ sleep 1 + retry=5 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.P42hTi4msJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.KJF2DQdgqJ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.P42hTi4msJ ++ cat /tmp/tmp.KJF2DQdgqJ ++ rm /tmp/tmp.P42hTi4msJ /tmp/tmp.KJF2DQdgqJ ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 5 -ge 120 ]] + echo -n . .+ sleep 1 + retry=6 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2SeOzwyc4Q +++ mktemp ++ local LAST_ERR=/tmp/tmp.jCfRk23TPq ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.2SeOzwyc4Q ++ cat /tmp/tmp.jCfRk23TPq ++ rm /tmp/tmp.2SeOzwyc4Q /tmp/tmp.jCfRk23TPq ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 6 -ge 120 ]] + echo -n . .+ sleep 1 + retry=7 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.A1azcUZBF4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.VKRoR1P9p4 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.A1azcUZBF4 ++ cat /tmp/tmp.VKRoR1P9p4 ++ rm /tmp/tmp.A1azcUZBF4 /tmp/tmp.VKRoR1P9p4 ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 7 -ge 120 ]] + echo -n . .+ sleep 1 + retry=8 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RgIMPUKtqZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.2FsTY6WfUe ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.RgIMPUKtqZ ++ cat /tmp/tmp.2FsTY6WfUe ++ rm /tmp/tmp.RgIMPUKtqZ /tmp/tmp.2FsTY6WfUe ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 8 -ge 120 ]] + echo -n . .+ sleep 1 + retry=9 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QAMM0whJqx +++ mktemp ++ local LAST_ERR=/tmp/tmp.qh2x3ArIkT ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.QAMM0whJqx ++ cat /tmp/tmp.qh2x3ArIkT ++ rm /tmp/tmp.QAMM0whJqx /tmp/tmp.qh2x3ArIkT ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 9 -ge 120 ]] + echo -n . .+ sleep 1 + retry=10 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3i99Fgxt3D +++ mktemp ++ local LAST_ERR=/tmp/tmp.Brr12F8Wau ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.3i99Fgxt3D ++ cat /tmp/tmp.Brr12F8Wau ++ rm /tmp/tmp.3i99Fgxt3D /tmp/tmp.Brr12F8Wau ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 10 -ge 120 ]] + echo -n . .+ sleep 1 + retry=11 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wJMZ4nnqjI +++ mktemp ++ local LAST_ERR=/tmp/tmp.F4klBEQF7V ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.wJMZ4nnqjI ++ cat /tmp/tmp.F4klBEQF7V ++ rm /tmp/tmp.wJMZ4nnqjI /tmp/tmp.F4klBEQF7V ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 11 -ge 120 ]] + echo -n . .+ sleep 1 + retry=12 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dydwzkwVIW +++ mktemp ++ local LAST_ERR=/tmp/tmp.5PVuCWjWtn ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.dydwzkwVIW ++ cat /tmp/tmp.5PVuCWjWtn ++ rm /tmp/tmp.dydwzkwVIW /tmp/tmp.5PVuCWjWtn ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 12 -ge 120 ]] + echo -n . .+ sleep 1 + retry=13 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.aa4bfKbNWL +++ mktemp ++ local LAST_ERR=/tmp/tmp.2iqMJypXax ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.aa4bfKbNWL ++ cat /tmp/tmp.2iqMJypXax ++ rm /tmp/tmp.aa4bfKbNWL /tmp/tmp.2iqMJypXax ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 13 -ge 120 ]] + echo -n . .+ sleep 1 + retry=14 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gLpjBQyZdZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.q7j9lYGZVD ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.gLpjBQyZdZ ++ cat /tmp/tmp.q7j9lYGZVD ++ rm /tmp/tmp.gLpjBQyZdZ /tmp/tmp.q7j9lYGZVD ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 14 -ge 120 ]] + echo -n . .+ sleep 1 + retry=15 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ooS2PRqy2m +++ mktemp ++ local LAST_ERR=/tmp/tmp.vcMZmrypuM ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ooS2PRqy2m ++ cat /tmp/tmp.vcMZmrypuM ++ rm /tmp/tmp.ooS2PRqy2m /tmp/tmp.vcMZmrypuM ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 15 -ge 120 ]] + echo -n . .+ sleep 1 + retry=16 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oA7f5VHrAO +++ mktemp ++ local LAST_ERR=/tmp/tmp.zhLPqZucOF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.oA7f5VHrAO ++ cat /tmp/tmp.zhLPqZucOF ++ rm /tmp/tmp.oA7f5VHrAO /tmp/tmp.zhLPqZucOF ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 16 -ge 120 ]] + echo -n . .+ sleep 1 + retry=17 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eYxoooPYa6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.fyzlYsHicy ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.eYxoooPYa6 ++ cat /tmp/tmp.fyzlYsHicy ++ rm /tmp/tmp.eYxoooPYa6 /tmp/tmp.fyzlYsHicy ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 17 -ge 120 ]] + echo -n . .+ sleep 1 + retry=18 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nl20EV7XJ4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.C2rUbIUl5C ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.nl20EV7XJ4 ++ cat /tmp/tmp.C2rUbIUl5C ++ rm /tmp/tmp.nl20EV7XJ4 /tmp/tmp.C2rUbIUl5C ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 18 -ge 120 ]] + echo -n . .+ sleep 1 + retry=19 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IaiFXCz9nn +++ mktemp ++ local LAST_ERR=/tmp/tmp.p4Y7K1s3XI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.IaiFXCz9nn ++ cat /tmp/tmp.p4Y7K1s3XI ++ rm /tmp/tmp.IaiFXCz9nn /tmp/tmp.p4Y7K1s3XI ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 19 -ge 120 ]] + echo -n . .+ sleep 1 + retry=20 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2rx8skCU7y +++ mktemp ++ local LAST_ERR=/tmp/tmp.1YxPyGYXqi ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.2rx8skCU7y ++ cat /tmp/tmp.1YxPyGYXqi ++ rm /tmp/tmp.2rx8skCU7y /tmp/tmp.1YxPyGYXqi ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 20 -ge 120 ]] + echo -n . .+ sleep 1 + retry=21 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KHtvARySmh +++ mktemp ++ local LAST_ERR=/tmp/tmp.nxBSm12HjM ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.KHtvARySmh ++ cat /tmp/tmp.nxBSm12HjM ++ rm /tmp/tmp.KHtvARySmh /tmp/tmp.nxBSm12HjM ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 21 -ge 120 ]] + echo -n . .+ sleep 1 + retry=22 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XxUsa4adiL +++ mktemp ++ local LAST_ERR=/tmp/tmp.gWXmVcOWaC ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.XxUsa4adiL ++ cat /tmp/tmp.gWXmVcOWaC ++ rm /tmp/tmp.XxUsa4adiL /tmp/tmp.gWXmVcOWaC ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 22 -ge 120 ]] + echo -n . .+ sleep 1 + retry=23 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.580hdmjjWM +++ mktemp ++ local LAST_ERR=/tmp/tmp.afaagKXP4q ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.580hdmjjWM ++ cat /tmp/tmp.afaagKXP4q ++ rm /tmp/tmp.580hdmjjWM /tmp/tmp.afaagKXP4q ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 23 -ge 120 ]] + echo -n . .+ sleep 1 + retry=24 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5ufGEKaYSs +++ mktemp ++ local LAST_ERR=/tmp/tmp.SoPf5mqf2u ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5ufGEKaYSs ++ cat /tmp/tmp.SoPf5mqf2u ++ rm /tmp/tmp.5ufGEKaYSs /tmp/tmp.SoPf5mqf2u ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 24 -ge 120 ]] + echo -n . .+ sleep 1 + retry=25 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.e35wkxUof9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Kwd6KcAiGj ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.e35wkxUof9 ++ cat /tmp/tmp.Kwd6KcAiGj ++ rm /tmp/tmp.e35wkxUof9 /tmp/tmp.Kwd6KcAiGj ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 25 -ge 120 ]] + echo -n . .+ sleep 1 + retry=26 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6q6mdOs1GW +++ mktemp ++ local LAST_ERR=/tmp/tmp.mweIPTkE2O ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.6q6mdOs1GW ++ cat /tmp/tmp.mweIPTkE2O ++ rm /tmp/tmp.6q6mdOs1GW /tmp/tmp.mweIPTkE2O ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 26 -ge 120 ]] + echo -n . .+ sleep 1 + retry=27 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3m1dTBUtJ0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.7c2foIMG5v ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.3m1dTBUtJ0 ++ cat /tmp/tmp.7c2foIMG5v ++ rm /tmp/tmp.3m1dTBUtJ0 /tmp/tmp.7c2foIMG5v ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 27 -ge 120 ]] + echo -n . .+ sleep 1 + retry=28 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zJTXsmxLzt +++ mktemp ++ local LAST_ERR=/tmp/tmp.XpXYPIWkWh ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.zJTXsmxLzt ++ cat /tmp/tmp.XpXYPIWkWh ++ rm /tmp/tmp.zJTXsmxLzt /tmp/tmp.XpXYPIWkWh ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 28 -ge 120 ]] + echo -n . .+ sleep 1 + retry=29 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xFVc03iTbk +++ mktemp ++ local LAST_ERR=/tmp/tmp.JBdenNh6Nq ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xFVc03iTbk ++ cat /tmp/tmp.JBdenNh6Nq ++ rm /tmp/tmp.xFVc03iTbk /tmp/tmp.JBdenNh6Nq ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 29 -ge 120 ]] + echo -n . .+ sleep 1 + retry=30 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ggGn85j8IL +++ mktemp ++ local LAST_ERR=/tmp/tmp.rFEKUYbcqq ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ggGn85j8IL ++ cat /tmp/tmp.rFEKUYbcqq ++ rm /tmp/tmp.ggGn85j8IL /tmp/tmp.rFEKUYbcqq ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 30 -ge 120 ]] + echo -n . .+ sleep 1 + retry=31 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LXHSzI4K0r +++ mktemp ++ local LAST_ERR=/tmp/tmp.8tgFvCsgvt ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.LXHSzI4K0r ++ cat /tmp/tmp.8tgFvCsgvt ++ rm /tmp/tmp.LXHSzI4K0r /tmp/tmp.8tgFvCsgvt ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 31 -ge 120 ]] + echo -n . .+ sleep 1 + retry=32 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DUOT8qrI3K +++ mktemp ++ local LAST_ERR=/tmp/tmp.BdJjacsIwt ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.DUOT8qrI3K ++ cat /tmp/tmp.BdJjacsIwt ++ rm /tmp/tmp.DUOT8qrI3K /tmp/tmp.BdJjacsIwt ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 32 -ge 120 ]] + echo -n . .+ sleep 1 + retry=33 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PBa3DQO614 +++ mktemp ++ local LAST_ERR=/tmp/tmp.VYzaYt8SGu ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.PBa3DQO614 ++ cat /tmp/tmp.VYzaYt8SGu ++ rm /tmp/tmp.PBa3DQO614 /tmp/tmp.VYzaYt8SGu ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 33 -ge 120 ]] + echo -n . .+ sleep 1 + retry=34 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RqIpaCsGLi +++ mktemp ++ local LAST_ERR=/tmp/tmp.MhAov2gnZc ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.RqIpaCsGLi ++ cat /tmp/tmp.MhAov2gnZc ++ rm /tmp/tmp.RqIpaCsGLi /tmp/tmp.MhAov2gnZc ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 34 -ge 120 ]] + echo -n . .+ sleep 1 + retry=35 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XsoNxMFJZa +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ul8XE3DKKT ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.XsoNxMFJZa ++ cat /tmp/tmp.Ul8XE3DKKT ++ rm /tmp/tmp.XsoNxMFJZa /tmp/tmp.Ul8XE3DKKT ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 35 -ge 120 ]] + echo -n . .+ sleep 1 + retry=36 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mOGuIGpZai +++ mktemp ++ local LAST_ERR=/tmp/tmp.JE8wKNppet ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.mOGuIGpZai ++ cat /tmp/tmp.JE8wKNppet ++ rm /tmp/tmp.mOGuIGpZai /tmp/tmp.JE8wKNppet ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 36 -ge 120 ]] + echo -n . .+ sleep 1 + retry=37 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mAGaXb2jRz +++ mktemp ++ local LAST_ERR=/tmp/tmp.5wfMh4CA05 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.mAGaXb2jRz ++ cat /tmp/tmp.5wfMh4CA05 ++ rm /tmp/tmp.mAGaXb2jRz /tmp/tmp.5wfMh4CA05 ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 37 -ge 120 ]] + echo -n . .+ sleep 1 + retry=38 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ngTlxllMxQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.euUC9g0s17 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ngTlxllMxQ ++ cat /tmp/tmp.euUC9g0s17 ++ rm /tmp/tmp.ngTlxllMxQ /tmp/tmp.euUC9g0s17 ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 38 -ge 120 ]] + echo -n . .+ sleep 1 + retry=39 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZLs4W9XaRi +++ mktemp ++ local LAST_ERR=/tmp/tmp.mE4lnBXH5r ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ZLs4W9XaRi ++ cat /tmp/tmp.mE4lnBXH5r ++ rm /tmp/tmp.ZLs4W9XaRi /tmp/tmp.mE4lnBXH5r ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 39 -ge 120 ]] + echo -n . .+ sleep 1 + retry=40 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xPVu6o5aHn +++ mktemp ++ local LAST_ERR=/tmp/tmp.gfdjG4ebcJ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xPVu6o5aHn ++ cat /tmp/tmp.gfdjG4ebcJ ++ rm /tmp/tmp.xPVu6o5aHn /tmp/tmp.gfdjG4ebcJ ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 40 -ge 120 ]] + echo -n . .+ sleep 1 + retry=41 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.g669Jdb6xj +++ mktemp ++ local LAST_ERR=/tmp/tmp.tYQ4gl5vjG ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.g669Jdb6xj ++ cat /tmp/tmp.tYQ4gl5vjG ++ rm /tmp/tmp.g669Jdb6xj /tmp/tmp.tYQ4gl5vjG ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 41 -ge 120 ]] + echo -n . .+ sleep 1 + retry=42 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gWnBx5dl25 +++ mktemp ++ local LAST_ERR=/tmp/tmp.DaI7oQO0ss ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.gWnBx5dl25 ++ cat /tmp/tmp.DaI7oQO0ss ++ rm /tmp/tmp.gWnBx5dl25 /tmp/tmp.DaI7oQO0ss ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 42 -ge 120 ]] + echo -n . .+ sleep 1 + retry=43 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2nuvckQFbB +++ mktemp ++ local LAST_ERR=/tmp/tmp.P93dOtk4Yp ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.2nuvckQFbB ++ cat /tmp/tmp.P93dOtk4Yp ++ rm /tmp/tmp.2nuvckQFbB /tmp/tmp.P93dOtk4Yp ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 43 -ge 120 ]] + echo -n . .+ sleep 1 + retry=44 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2wLdVbZz17 +++ mktemp ++ local LAST_ERR=/tmp/tmp.bNMqGEN1xf ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.2wLdVbZz17 ++ cat /tmp/tmp.bNMqGEN1xf ++ rm /tmp/tmp.2wLdVbZz17 /tmp/tmp.bNMqGEN1xf ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 44 -ge 120 ]] + echo -n . .+ sleep 1 + retry=45 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MHRhMvBI89 +++ mktemp ++ local LAST_ERR=/tmp/tmp.dSZTizeq9M ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.MHRhMvBI89 ++ cat /tmp/tmp.dSZTizeq9M ++ rm /tmp/tmp.MHRhMvBI89 /tmp/tmp.dSZTizeq9M ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 45 -ge 120 ]] + echo -n . .+ sleep 1 + retry=46 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Jxw70NeH8M +++ mktemp ++ local LAST_ERR=/tmp/tmp.voAndVZyRC ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Jxw70NeH8M ++ cat /tmp/tmp.voAndVZyRC ++ rm /tmp/tmp.Jxw70NeH8M /tmp/tmp.voAndVZyRC ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 46 -ge 120 ]] + echo -n . .+ sleep 1 + retry=47 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Uzmwkf6lIl +++ mktemp ++ local LAST_ERR=/tmp/tmp.F4b0rk8VGF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Uzmwkf6lIl ++ cat /tmp/tmp.F4b0rk8VGF ++ rm /tmp/tmp.Uzmwkf6lIl /tmp/tmp.F4b0rk8VGF ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 47 -ge 120 ]] + echo -n . .+ sleep 1 + retry=48 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zVG2jQA3rR +++ mktemp ++ local LAST_ERR=/tmp/tmp.DKRaaOr7kf ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.zVG2jQA3rR ++ cat /tmp/tmp.DKRaaOr7kf ++ rm /tmp/tmp.zVG2jQA3rR /tmp/tmp.DKRaaOr7kf ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 48 -ge 120 ]] + echo -n . .+ sleep 1 + retry=49 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6IjIwLGOrh +++ mktemp ++ local LAST_ERR=/tmp/tmp.AFpQUBGr3n ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.6IjIwLGOrh ++ cat /tmp/tmp.AFpQUBGr3n ++ rm /tmp/tmp.6IjIwLGOrh /tmp/tmp.AFpQUBGr3n ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 49 -ge 120 ]] + echo -n . .+ sleep 1 + retry=50 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.K6QLw2bPS6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Q6hbYmsoLJ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.K6QLw2bPS6 ++ cat /tmp/tmp.Q6hbYmsoLJ ++ rm /tmp/tmp.K6QLw2bPS6 /tmp/tmp.Q6hbYmsoLJ ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 50 -ge 120 ]] + echo -n . .+ sleep 1 + retry=51 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BnoBDUS2AV +++ mktemp ++ local LAST_ERR=/tmp/tmp.nxh6NqT1X6 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.BnoBDUS2AV ++ cat /tmp/tmp.nxh6NqT1X6 ++ rm /tmp/tmp.BnoBDUS2AV /tmp/tmp.nxh6NqT1X6 ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 51 -ge 120 ]] + echo -n . .+ sleep 1 + retry=52 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Wtmpja4sCZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.2ujZ5SkVds ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Wtmpja4sCZ ++ cat /tmp/tmp.2ujZ5SkVds ++ rm /tmp/tmp.Wtmpja4sCZ /tmp/tmp.2ujZ5SkVds ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 52 -ge 120 ]] + echo -n . .+ sleep 1 + retry=53 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bCR2APH3ZO +++ mktemp ++ local LAST_ERR=/tmp/tmp.XJ18s227In ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.bCR2APH3ZO ++ cat /tmp/tmp.XJ18s227In ++ rm /tmp/tmp.bCR2APH3ZO /tmp/tmp.XJ18s227In ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 53 -ge 120 ]] + echo -n . .+ sleep 1 + retry=54 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xpMs8s4A5s +++ mktemp ++ local LAST_ERR=/tmp/tmp.E3zjU9Q0kx ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xpMs8s4A5s ++ cat /tmp/tmp.E3zjU9Q0kx ++ rm /tmp/tmp.xpMs8s4A5s /tmp/tmp.E3zjU9Q0kx ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 54 -ge 120 ]] + echo -n . .+ sleep 1 + retry=55 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cLnFVgOH8R +++ mktemp ++ local LAST_ERR=/tmp/tmp.37VzGhen8F ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.cLnFVgOH8R ++ cat /tmp/tmp.37VzGhen8F ++ rm /tmp/tmp.cLnFVgOH8R /tmp/tmp.37VzGhen8F ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 55 -ge 120 ]] + echo -n . .+ sleep 1 + retry=56 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.E4BrAOeTrB +++ mktemp ++ local LAST_ERR=/tmp/tmp.sP03vXQrFD ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.E4BrAOeTrB ++ cat /tmp/tmp.sP03vXQrFD ++ rm /tmp/tmp.E4BrAOeTrB /tmp/tmp.sP03vXQrFD ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 56 -ge 120 ]] + echo -n . .+ sleep 1 + retry=57 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gZEix3mmPy +++ mktemp ++ local LAST_ERR=/tmp/tmp.0qjvyqaupw ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.gZEix3mmPy ++ cat /tmp/tmp.0qjvyqaupw ++ rm /tmp/tmp.gZEix3mmPy /tmp/tmp.0qjvyqaupw ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 57 -ge 120 ]] + echo -n . .+ sleep 1 + retry=58 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VdMm7c1tfw +++ mktemp ++ local LAST_ERR=/tmp/tmp.cMwZjJqV88 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.VdMm7c1tfw ++ cat /tmp/tmp.cMwZjJqV88 ++ rm /tmp/tmp.VdMm7c1tfw /tmp/tmp.cMwZjJqV88 ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 58 -ge 120 ]] + echo -n . .+ sleep 1 + retry=59 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lAifN5KYGa +++ mktemp ++ local LAST_ERR=/tmp/tmp.pdKUxf1F9H ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.lAifN5KYGa ++ cat /tmp/tmp.pdKUxf1F9H ++ rm /tmp/tmp.lAifN5KYGa /tmp/tmp.pdKUxf1F9H ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 59 -ge 120 ]] + echo -n . .+ sleep 1 + retry=60 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.l2KpKwEbMF +++ mktemp ++ local LAST_ERR=/tmp/tmp.hT3lgV1yoa ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.l2KpKwEbMF ++ cat /tmp/tmp.hT3lgV1yoa ++ rm /tmp/tmp.l2KpKwEbMF /tmp/tmp.hT3lgV1yoa ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 60 -ge 120 ]] + echo -n . .+ sleep 1 + retry=61 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wah4HGTK7d +++ mktemp ++ local LAST_ERR=/tmp/tmp.Gqdrd3kzeB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.wah4HGTK7d ++ cat /tmp/tmp.Gqdrd3kzeB ++ rm /tmp/tmp.wah4HGTK7d /tmp/tmp.Gqdrd3kzeB ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 61 -ge 120 ]] + echo -n . .+ sleep 1 + retry=62 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UXXVDKdI8e +++ mktemp ++ local LAST_ERR=/tmp/tmp.pksJAPT563 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.UXXVDKdI8e ++ cat /tmp/tmp.pksJAPT563 ++ rm /tmp/tmp.UXXVDKdI8e /tmp/tmp.pksJAPT563 ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 62 -ge 120 ]] + echo -n . .+ sleep 1 + retry=63 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sxM2mQrJA3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.hEVglCIfc4 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.sxM2mQrJA3 ++ cat /tmp/tmp.hEVglCIfc4 ++ rm /tmp/tmp.sxM2mQrJA3 /tmp/tmp.hEVglCIfc4 ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 63 -ge 120 ]] + echo -n . .+ sleep 1 + retry=64 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JVjQOq9Yqb +++ mktemp ++ local LAST_ERR=/tmp/tmp.xKoemAeW8S ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.JVjQOq9Yqb ++ cat /tmp/tmp.xKoemAeW8S ++ rm /tmp/tmp.JVjQOq9Yqb /tmp/tmp.xKoemAeW8S ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 64 -ge 120 ]] + echo -n . .+ sleep 1 + retry=65 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.e1ICeGIO6Y +++ mktemp ++ local LAST_ERR=/tmp/tmp.vDgMIyosen ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.e1ICeGIO6Y ++ cat /tmp/tmp.vDgMIyosen ++ rm /tmp/tmp.e1ICeGIO6Y /tmp/tmp.vDgMIyosen ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 65 -ge 120 ]] + echo -n . .+ sleep 1 + retry=66 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EkHxuM5wOD +++ mktemp ++ local LAST_ERR=/tmp/tmp.V8uDlEyuz3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.EkHxuM5wOD ++ cat /tmp/tmp.V8uDlEyuz3 ++ rm /tmp/tmp.EkHxuM5wOD /tmp/tmp.V8uDlEyuz3 ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 66 -ge 120 ]] + echo -n . .+ sleep 1 + retry=67 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4ICcZfhn8X +++ mktemp ++ local LAST_ERR=/tmp/tmp.j8lVAd2mPU ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4ICcZfhn8X ++ cat /tmp/tmp.j8lVAd2mPU ++ rm /tmp/tmp.4ICcZfhn8X /tmp/tmp.j8lVAd2mPU ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 67 -ge 120 ]] + echo -n . .+ sleep 1 + retry=68 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dK65UnIUmf +++ mktemp ++ local LAST_ERR=/tmp/tmp.y0JcdgD4fW ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.dK65UnIUmf ++ cat /tmp/tmp.y0JcdgD4fW ++ rm /tmp/tmp.dK65UnIUmf /tmp/tmp.y0JcdgD4fW ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 68 -ge 120 ]] + echo -n . .+ sleep 1 + retry=69 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LSHVuVgaXS +++ mktemp ++ local LAST_ERR=/tmp/tmp.pmR1uiKwob ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.LSHVuVgaXS ++ cat /tmp/tmp.pmR1uiKwob ++ rm /tmp/tmp.LSHVuVgaXS /tmp/tmp.pmR1uiKwob ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 69 -ge 120 ]] + echo -n . .+ sleep 1 + retry=70 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eXTeO4H0kS +++ mktemp ++ local LAST_ERR=/tmp/tmp.JCkqSw8KWF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.eXTeO4H0kS ++ cat /tmp/tmp.JCkqSw8KWF ++ rm /tmp/tmp.eXTeO4H0kS /tmp/tmp.JCkqSw8KWF ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 70 -ge 120 ]] + echo -n . .+ sleep 1 + retry=71 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rXpT8Cf7BY +++ mktemp ++ local LAST_ERR=/tmp/tmp.RMXF6LQZjV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.rXpT8Cf7BY ++ cat /tmp/tmp.RMXF6LQZjV ++ rm /tmp/tmp.rXpT8Cf7BY /tmp/tmp.RMXF6LQZjV ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 71 -ge 120 ]] + echo -n . .+ sleep 1 + retry=72 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UZUTyPc5An +++ mktemp ++ local LAST_ERR=/tmp/tmp.NV8wDSQyl3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.UZUTyPc5An ++ cat /tmp/tmp.NV8wDSQyl3 ++ rm /tmp/tmp.UZUTyPc5An /tmp/tmp.NV8wDSQyl3 ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 72 -ge 120 ]] + echo -n . .+ sleep 1 + retry=73 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BeQTpdGEVB +++ mktemp ++ local LAST_ERR=/tmp/tmp.XprO0njusB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.BeQTpdGEVB ++ cat /tmp/tmp.XprO0njusB ++ rm /tmp/tmp.BeQTpdGEVB /tmp/tmp.XprO0njusB ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 73 -ge 120 ]] + echo -n . .+ sleep 1 + retry=74 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Vc6bncBZKn +++ mktemp ++ local LAST_ERR=/tmp/tmp.vrsnDnkkT2 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Vc6bncBZKn ++ cat /tmp/tmp.vrsnDnkkT2 ++ rm /tmp/tmp.Vc6bncBZKn /tmp/tmp.vrsnDnkkT2 ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 74 -ge 120 ]] + echo -n . .+ sleep 1 + retry=75 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AgqUJVcJap +++ mktemp ++ local LAST_ERR=/tmp/tmp.idZgDEjaQ2 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.AgqUJVcJap ++ cat /tmp/tmp.idZgDEjaQ2 ++ rm /tmp/tmp.AgqUJVcJap /tmp/tmp.idZgDEjaQ2 ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 75 -ge 120 ]] + echo -n . .+ sleep 1 + retry=76 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.37FrguC91F +++ mktemp ++ local LAST_ERR=/tmp/tmp.O2Ra46wUAk ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.37FrguC91F ++ cat /tmp/tmp.O2Ra46wUAk ++ rm /tmp/tmp.37FrguC91F /tmp/tmp.O2Ra46wUAk ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 76 -ge 120 ]] + echo -n . .+ sleep 1 + retry=77 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cUkzoznmj6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.cSsmsjQtB4 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.cUkzoznmj6 ++ cat /tmp/tmp.cSsmsjQtB4 ++ rm /tmp/tmp.cUkzoznmj6 /tmp/tmp.cSsmsjQtB4 ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 77 -ge 120 ]] + echo -n . .+ sleep 1 + retry=78 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bRX4lOnOLX +++ mktemp ++ local LAST_ERR=/tmp/tmp.W5hU64WMGP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.bRX4lOnOLX ++ cat /tmp/tmp.W5hU64WMGP ++ rm /tmp/tmp.bRX4lOnOLX /tmp/tmp.W5hU64WMGP ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 78 -ge 120 ]] + echo -n . .+ sleep 1 + retry=79 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.akOvUKS7Wm +++ mktemp ++ local LAST_ERR=/tmp/tmp.zAfnn7xvIF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.akOvUKS7Wm ++ cat /tmp/tmp.zAfnn7xvIF ++ rm /tmp/tmp.akOvUKS7Wm /tmp/tmp.zAfnn7xvIF ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 79 -ge 120 ]] + echo -n . .+ sleep 1 + retry=80 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mBgmERCRpX +++ mktemp ++ local LAST_ERR=/tmp/tmp.JBhMHeYDU8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.mBgmERCRpX ++ cat /tmp/tmp.JBhMHeYDU8 ++ rm /tmp/tmp.mBgmERCRpX /tmp/tmp.JBhMHeYDU8 ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 80 -ge 120 ]] + echo -n . .+ sleep 1 + retry=81 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HoqDOkDB0R +++ mktemp ++ local LAST_ERR=/tmp/tmp.9Pv785RRMV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.HoqDOkDB0R ++ cat /tmp/tmp.9Pv785RRMV ++ rm /tmp/tmp.HoqDOkDB0R /tmp/tmp.9Pv785RRMV ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 81 -ge 120 ]] + echo -n . .+ sleep 1 + retry=82 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.A2DpCWvpEq +++ mktemp ++ local LAST_ERR=/tmp/tmp.AzXvVaiC7f ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.A2DpCWvpEq ++ cat /tmp/tmp.AzXvVaiC7f ++ rm /tmp/tmp.A2DpCWvpEq /tmp/tmp.AzXvVaiC7f ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 82 -ge 120 ]] + echo -n . .+ sleep 1 + retry=83 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dZBH1y0kIc +++ mktemp ++ local LAST_ERR=/tmp/tmp.59f8lkMwY3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.dZBH1y0kIc ++ cat /tmp/tmp.59f8lkMwY3 ++ rm /tmp/tmp.dZBH1y0kIc /tmp/tmp.59f8lkMwY3 ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 83 -ge 120 ]] + echo -n . .+ sleep 1 + retry=84 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xTgLq94mSo +++ mktemp ++ local LAST_ERR=/tmp/tmp.xZsP7xaCHu ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xTgLq94mSo ++ cat /tmp/tmp.xZsP7xaCHu ++ rm /tmp/tmp.xTgLq94mSo /tmp/tmp.xZsP7xaCHu ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 84 -ge 120 ]] + echo -n . .+ sleep 1 + retry=85 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lFNByyIIgx +++ mktemp ++ local LAST_ERR=/tmp/tmp.lcFGaljH9q ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.lFNByyIIgx ++ cat /tmp/tmp.lcFGaljH9q ++ rm /tmp/tmp.lFNByyIIgx /tmp/tmp.lcFGaljH9q ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 85 -ge 120 ]] + echo -n . .+ sleep 1 + retry=86 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JtIMO7HoGT +++ mktemp ++ local LAST_ERR=/tmp/tmp.08rF2Jpvvd ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.JtIMO7HoGT ++ cat /tmp/tmp.08rF2Jpvvd ++ rm /tmp/tmp.JtIMO7HoGT /tmp/tmp.08rF2Jpvvd ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 86 -ge 120 ]] + echo -n . .+ sleep 1 + retry=87 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zTGavVJlcy +++ mktemp ++ local LAST_ERR=/tmp/tmp.mXo3ZNDRbL ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.zTGavVJlcy ++ cat /tmp/tmp.mXo3ZNDRbL ++ rm /tmp/tmp.zTGavVJlcy /tmp/tmp.mXo3ZNDRbL ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 87 -ge 120 ]] + echo -n . .+ sleep 1 + retry=88 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lcIVR530hm +++ mktemp ++ local LAST_ERR=/tmp/tmp.qt0i2sIU9O ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.lcIVR530hm ++ cat /tmp/tmp.qt0i2sIU9O ++ rm /tmp/tmp.lcIVR530hm /tmp/tmp.qt0i2sIU9O ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 88 -ge 120 ]] + echo -n . .+ sleep 1 + retry=89 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hUlXK0B6xD +++ mktemp ++ local LAST_ERR=/tmp/tmp.5yvLpVKha4 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.hUlXK0B6xD ++ cat /tmp/tmp.5yvLpVKha4 ++ rm /tmp/tmp.hUlXK0B6xD /tmp/tmp.5yvLpVKha4 ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 89 -ge 120 ]] + echo -n . .+ sleep 1 + retry=90 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.O7owUXCqx1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.P2NI3TAD74 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.O7owUXCqx1 ++ cat /tmp/tmp.P2NI3TAD74 ++ rm /tmp/tmp.O7owUXCqx1 /tmp/tmp.P2NI3TAD74 ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 90 -ge 120 ]] + echo -n . .+ sleep 1 + retry=91 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mn2LQtRDUG +++ mktemp ++ local LAST_ERR=/tmp/tmp.MTJVoyKLxt ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.mn2LQtRDUG ++ cat /tmp/tmp.MTJVoyKLxt ++ rm /tmp/tmp.mn2LQtRDUG /tmp/tmp.MTJVoyKLxt ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 91 -ge 120 ]] + echo -n . .+ sleep 1 + retry=92 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MjBAVk9CeG +++ mktemp ++ local LAST_ERR=/tmp/tmp.YPL4wAO5TH ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.MjBAVk9CeG ++ cat /tmp/tmp.YPL4wAO5TH ++ rm /tmp/tmp.MjBAVk9CeG /tmp/tmp.YPL4wAO5TH ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 92 -ge 120 ]] + echo -n . .+ sleep 1 + retry=93 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0q5zKb64kN +++ mktemp ++ local LAST_ERR=/tmp/tmp.YUwhq0Mpnz ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.0q5zKb64kN ++ cat /tmp/tmp.YUwhq0Mpnz ++ rm /tmp/tmp.0q5zKb64kN /tmp/tmp.YUwhq0Mpnz ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 93 -ge 120 ]] + echo -n . .+ sleep 1 + retry=94 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9JMrVA8ApB +++ mktemp ++ local LAST_ERR=/tmp/tmp.gTfP6JkxwJ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9JMrVA8ApB ++ cat /tmp/tmp.gTfP6JkxwJ ++ rm /tmp/tmp.9JMrVA8ApB /tmp/tmp.gTfP6JkxwJ ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 94 -ge 120 ]] + echo -n . .+ sleep 1 + retry=95 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qS9SNnu4rQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.77J7Y2ILgA ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.qS9SNnu4rQ ++ cat /tmp/tmp.77J7Y2ILgA ++ rm /tmp/tmp.qS9SNnu4rQ /tmp/tmp.77J7Y2ILgA ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 95 -ge 120 ]] + echo -n . .+ sleep 1 + retry=96 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7qp5zSTwXC +++ mktemp ++ local LAST_ERR=/tmp/tmp.5VCjHYoZ0Y ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.7qp5zSTwXC ++ cat /tmp/tmp.5VCjHYoZ0Y ++ rm /tmp/tmp.7qp5zSTwXC /tmp/tmp.5VCjHYoZ0Y ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 96 -ge 120 ]] + echo -n . .+ sleep 1 + retry=97 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xBiEFaf9sn +++ mktemp ++ local LAST_ERR=/tmp/tmp.VJQ8lZmDfh ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xBiEFaf9sn ++ cat /tmp/tmp.VJQ8lZmDfh ++ rm /tmp/tmp.xBiEFaf9sn /tmp/tmp.VJQ8lZmDfh ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 97 -ge 120 ]] + echo -n . .+ sleep 1 + retry=98 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Cvs2hO6hNA +++ mktemp ++ local LAST_ERR=/tmp/tmp.yUWRdCzwgl ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Cvs2hO6hNA ++ cat /tmp/tmp.yUWRdCzwgl ++ rm /tmp/tmp.Cvs2hO6hNA /tmp/tmp.yUWRdCzwgl ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 98 -ge 120 ]] + echo -n . .+ sleep 1 + retry=99 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Dy0QRxOx8l +++ mktemp ++ local LAST_ERR=/tmp/tmp.LtwRY4vzBI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Dy0QRxOx8l ++ cat /tmp/tmp.LtwRY4vzBI ++ rm /tmp/tmp.Dy0QRxOx8l /tmp/tmp.LtwRY4vzBI ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 99 -ge 120 ]] + echo -n . .+ sleep 1 + retry=100 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.iwhuag2d99 +++ mktemp ++ local LAST_ERR=/tmp/tmp.ErqqtSLeoR ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.iwhuag2d99 ++ cat /tmp/tmp.ErqqtSLeoR ++ rm /tmp/tmp.iwhuag2d99 /tmp/tmp.ErqqtSLeoR ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 100 -ge 120 ]] + echo -n . .+ sleep 1 + retry=101 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pZ4jhLoSfB +++ mktemp ++ local LAST_ERR=/tmp/tmp.EnWAQ2JfPC ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.pZ4jhLoSfB ++ cat /tmp/tmp.EnWAQ2JfPC ++ rm /tmp/tmp.pZ4jhLoSfB /tmp/tmp.EnWAQ2JfPC ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 101 -ge 120 ]] + echo -n . .+ sleep 1 + retry=102 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bUn8UGu1HM +++ mktemp ++ local LAST_ERR=/tmp/tmp.LewgfbyYaP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.bUn8UGu1HM ++ cat /tmp/tmp.LewgfbyYaP ++ rm /tmp/tmp.bUn8UGu1HM /tmp/tmp.LewgfbyYaP ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 102 -ge 120 ]] + echo -n . .+ sleep 1 + retry=103 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9Drmdtoly9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.wicmjg0DTo ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9Drmdtoly9 ++ cat /tmp/tmp.wicmjg0DTo ++ rm /tmp/tmp.9Drmdtoly9 /tmp/tmp.wicmjg0DTo ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 103 -ge 120 ]] + echo -n . .+ sleep 1 + retry=104 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eMye30jP1Q +++ mktemp ++ local LAST_ERR=/tmp/tmp.e2vLHev4Mq ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.eMye30jP1Q ++ cat /tmp/tmp.e2vLHev4Mq ++ rm /tmp/tmp.eMye30jP1Q /tmp/tmp.e2vLHev4Mq ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 104 -ge 120 ]] + echo -n . .+ sleep 1 + retry=105 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7x0fNthGM6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.VzEjwJarqB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.7x0fNthGM6 ++ cat /tmp/tmp.VzEjwJarqB ++ rm /tmp/tmp.7x0fNthGM6 /tmp/tmp.VzEjwJarqB ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 105 -ge 120 ]] + echo -n . .+ sleep 1 + retry=106 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ltLpJWoCCS +++ mktemp ++ local LAST_ERR=/tmp/tmp.TWL8i6NHRd ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ltLpJWoCCS ++ cat /tmp/tmp.TWL8i6NHRd ++ rm /tmp/tmp.ltLpJWoCCS /tmp/tmp.TWL8i6NHRd ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 106 -ge 120 ]] + echo -n . .+ sleep 1 + retry=107 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VVs9islPNC +++ mktemp ++ local LAST_ERR=/tmp/tmp.iGAE6rvmoE ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.VVs9islPNC ++ cat /tmp/tmp.iGAE6rvmoE ++ rm /tmp/tmp.VVs9islPNC /tmp/tmp.iGAE6rvmoE ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 107 -ge 120 ]] + echo -n . .+ sleep 1 + retry=108 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Peu6y5uwrk +++ mktemp ++ local LAST_ERR=/tmp/tmp.vF7X1KO2bC ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Peu6y5uwrk ++ cat /tmp/tmp.vF7X1KO2bC ++ rm /tmp/tmp.Peu6y5uwrk /tmp/tmp.vF7X1KO2bC ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 108 -ge 120 ]] + echo -n . .+ sleep 1 + retry=109 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YfuuE4Rf2d +++ mktemp ++ local LAST_ERR=/tmp/tmp.cffQWP1iug ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.YfuuE4Rf2d ++ cat /tmp/tmp.cffQWP1iug ++ rm /tmp/tmp.YfuuE4Rf2d /tmp/tmp.cffQWP1iug ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 109 -ge 120 ]] + echo -n . .+ sleep 1 + retry=110 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.J2WtT4Ja67 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Vr0BVra9xv ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.J2WtT4Ja67 ++ cat /tmp/tmp.Vr0BVra9xv ++ rm /tmp/tmp.J2WtT4Ja67 /tmp/tmp.Vr0BVra9xv ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 110 -ge 120 ]] + echo -n . .+ sleep 1 + retry=111 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.h1WU0T7ezo +++ mktemp ++ local LAST_ERR=/tmp/tmp.mEFcPqsMPY ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.h1WU0T7ezo ++ cat /tmp/tmp.mEFcPqsMPY ++ rm /tmp/tmp.h1WU0T7ezo /tmp/tmp.mEFcPqsMPY ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 111 -ge 120 ]] + echo -n . .+ sleep 1 + retry=112 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KBVq2xiFDW +++ mktemp ++ local LAST_ERR=/tmp/tmp.kbn6Z3IlnB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.KBVq2xiFDW ++ cat /tmp/tmp.kbn6Z3IlnB ++ rm /tmp/tmp.KBVq2xiFDW /tmp/tmp.kbn6Z3IlnB ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 112 -ge 120 ]] + echo -n . .+ sleep 1 + retry=113 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6LKOXJ5fRG +++ mktemp ++ local LAST_ERR=/tmp/tmp.wUaNNk4VbS ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.6LKOXJ5fRG ++ cat /tmp/tmp.wUaNNk4VbS ++ rm /tmp/tmp.6LKOXJ5fRG /tmp/tmp.wUaNNk4VbS ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 113 -ge 120 ]] + echo -n . .+ sleep 1 + retry=114 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hdXzt4AmVJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.tuwYVDNrmk ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.hdXzt4AmVJ ++ cat /tmp/tmp.tuwYVDNrmk ++ rm /tmp/tmp.hdXzt4AmVJ /tmp/tmp.tuwYVDNrmk ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 114 -ge 120 ]] + echo -n . .+ sleep 1 + retry=115 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vJeJHkHKXk +++ mktemp ++ local LAST_ERR=/tmp/tmp.5siKoa7B4i ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.vJeJHkHKXk ++ cat /tmp/tmp.5siKoa7B4i ++ rm /tmp/tmp.vJeJHkHKXk /tmp/tmp.5siKoa7B4i ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 115 -ge 120 ]] + echo -n . .+ sleep 1 + retry=116 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hPDaawQhh1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.pZE6n6YO4Y ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.hPDaawQhh1 ++ cat /tmp/tmp.pZE6n6YO4Y ++ rm /tmp/tmp.hPDaawQhh1 /tmp/tmp.pZE6n6YO4Y ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 116 -ge 120 ]] + echo -n . .+ sleep 1 + retry=117 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UUHuKQpbZr +++ mktemp ++ local LAST_ERR=/tmp/tmp.nlYyYjWwkY ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.UUHuKQpbZr ++ cat /tmp/tmp.nlYyYjWwkY ++ rm /tmp/tmp.UUHuKQpbZr /tmp/tmp.nlYyYjWwkY ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 117 -ge 120 ]] + echo -n . .+ sleep 1 + retry=118 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZOYiG7jomn +++ mktemp ++ local LAST_ERR=/tmp/tmp.cu3WUG6uyt ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ZOYiG7jomn ++ cat /tmp/tmp.cu3WUG6uyt ++ rm /tmp/tmp.ZOYiG7jomn /tmp/tmp.cu3WUG6uyt ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 118 -ge 120 ]] + echo -n . .+ sleep 1 + retry=119 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BHm8uf3p07 +++ mktemp ++ local LAST_ERR=/tmp/tmp.esLFC1yg71 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.BHm8uf3p07 ++ cat /tmp/tmp.esLFC1yg71 ++ rm /tmp/tmp.BHm8uf3p07 /tmp/tmp.esLFC1yg71 ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 119 -ge 120 ]] + echo -n . .+ sleep 1 + retry=120 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pP5pGZeXmq +++ mktemp ++ local LAST_ERR=/tmp/tmp.Af3LhbrLnA ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.pP5pGZeXmq ++ cat /tmp/tmp.Af3LhbrLnA ++ rm /tmp/tmp.pP5pGZeXmq /tmp/tmp.Af3LhbrLnA ++ return 0 + [[ 1Gi == 2Gi ]] + [[ 120 -ge 120 ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-0 was not resized, max retries exceeded' persistentvolumeclaim/mongod-data-some-name-rs0-0 was not resized, max retries exceeded + return 1 + return 1 + echo 'Enabling PVC resize' Enabling PVC resize + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "add", "path": "/spec/storageScaling", "value":{"enableVolumeScaling":true} }]' ++ mktemp + local LAST_OUT=/tmp/tmp.T9d5P7UHgp ++ mktemp + local LAST_ERR=/tmp/tmp.MRmZojX6g2 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "add", "path": "/spec/storageScaling", "value":{"enableVolumeScaling":true} }]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.T9d5P7UHgp perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.MRmZojX6g2 + rm /tmp/tmp.T9d5P7UHgp /tmp/tmp.MRmZojX6g2 + return 0 + sleep 10 + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OnvVilauRF +++ mktemp ++ local LAST_ERR=/tmp/tmp.8kFRykCPyJ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.OnvVilauRF ++ cat /tmp/tmp.8kFRykCPyJ ++ rm /tmp/tmp.OnvVilauRF /tmp/tmp.8kFRykCPyJ ++ return 0 + [[ initializing == ready ]] + let retry+=1 + '[' 1 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hf7IgAkADZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.0y5vA56AP2 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.hf7IgAkADZ ++ cat /tmp/tmp.0y5vA56AP2 ++ rm /tmp/tmp.hf7IgAkADZ /tmp/tmp.0y5vA56AP2 ++ return 0 + [[ initializing == ready ]] + let retry+=1 + '[' 2 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ddJOpEcn0k +++ mktemp ++ local LAST_ERR=/tmp/tmp.YvFffHmOPg ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ddJOpEcn0k ++ cat /tmp/tmp.YvFffHmOPg ++ rm /tmp/tmp.ddJOpEcn0k /tmp/tmp.YvFffHmOPg ++ return 0 + [[ initializing == ready ]] + let retry+=1 + '[' 3 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fSizhqs17R +++ mktemp ++ local LAST_ERR=/tmp/tmp.7X1r4udDD1 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.fSizhqs17R ++ cat /tmp/tmp.7X1r4udDD1 ++ rm /tmp/tmp.fSizhqs17R /tmp/tmp.7X1r4udDD1 ++ return 0 + [[ initializing == ready ]] + let retry+=1 + '[' 4 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vk2ElfzKkY +++ mktemp ++ local LAST_ERR=/tmp/tmp.y6BqhH1kRs ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.vk2ElfzKkY ++ cat /tmp/tmp.y6BqhH1kRs ++ rm /tmp/tmp.vk2ElfzKkY /tmp/tmp.y6BqhH1kRs ++ return 0 + [[ initializing == ready ]] + let retry+=1 + '[' 5 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8tpwH8gGEE +++ mktemp ++ local LAST_ERR=/tmp/tmp.yjkCWDU4Bk ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.8tpwH8gGEE ++ cat /tmp/tmp.yjkCWDU4Bk ++ rm /tmp/tmp.8tpwH8gGEE /tmp/tmp.yjkCWDU4Bk ++ return 0 + [[ initializing == ready ]] + let retry+=1 + '[' 6 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9kOLGKqEs2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.blWXRSfNaE ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9kOLGKqEs2 ++ cat /tmp/tmp.blWXRSfNaE ++ rm /tmp/tmp.9kOLGKqEs2 /tmp/tmp.blWXRSfNaE ++ return 0 + [[ initializing == ready ]] + let retry+=1 + '[' 7 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5YXTWqp7Nn +++ mktemp ++ local LAST_ERR=/tmp/tmp.3Aq1ZcmG11 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5YXTWqp7Nn ++ cat /tmp/tmp.3Aq1ZcmG11 ++ rm /tmp/tmp.5YXTWqp7Nn /tmp/tmp.3Aq1ZcmG11 ++ return 0 + [[ initializing == ready ]] + let retry+=1 + '[' 8 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6Zow4rLyNY +++ mktemp ++ local LAST_ERR=/tmp/tmp.chUQIgRvZ4 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.6Zow4rLyNY ++ cat /tmp/tmp.chUQIgRvZ4 ++ rm /tmp/tmp.6Zow4rLyNY /tmp/tmp.chUQIgRvZ4 ++ return 0 + [[ ready == ready ]] + echo .OK .OK + wait_all_pvc_resize 2Gi + local expected_size=2Gi + local max_retry=120 + local sleep_time=5 ++ kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name +++ mktemp ++ local LAST_OUT=/tmp/tmp.X0EDiheSCC +++ mktemp ++ local LAST_ERR=/tmp/tmp.wYMCuubafG ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc -l app.kubernetes.io/component=mongod -o name ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.X0EDiheSCC ++ cat /tmp/tmp.wYMCuubafG ++ rm /tmp/tmp.X0EDiheSCC /tmp/tmp.wYMCuubafG ++ return 0 + for pvc in $(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name) + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-0 2Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-0 + local expected_size=2Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bBQq5CZE9D +++ mktemp ++ local LAST_ERR=/tmp/tmp.rG49pKNkow ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.bBQq5CZE9D ++ cat /tmp/tmp.rG49pKNkow ++ rm /tmp/tmp.bBQq5CZE9D /tmp/tmp.rG49pKNkow ++ return 0 + [[ 2Gi == 2Gi ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized + return 0 + for pvc in $(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name) + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-1 2Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-1 + local expected_size=2Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-1 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-1 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9UwCQJaSrv +++ mktemp ++ local LAST_ERR=/tmp/tmp.8YjOC7qjzB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9UwCQJaSrv ++ cat /tmp/tmp.8YjOC7qjzB ++ rm /tmp/tmp.9UwCQJaSrv /tmp/tmp.8YjOC7qjzB ++ return 0 + [[ 2Gi == 2Gi ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-1 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-1 was resized + return 0 + for pvc in $(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name) + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-2 2Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-2 + local expected_size=2Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-2 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-2 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GHqINcoJZV +++ mktemp ++ local LAST_ERR=/tmp/tmp.oHBAxTaMoG ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.GHqINcoJZV ++ cat /tmp/tmp.oHBAxTaMoG ++ rm /tmp/tmp.GHqINcoJZV /tmp/tmp.oHBAxTaMoG ++ return 0 + [[ 2Gi == 2Gi ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-2 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-2 was resized + return 0 + return 0 + echo + [[ 0 == 1 ]] + [[ -n '' ]] + desc 'create resourcequota' + set +o xtrace ----------------------------------------------------------------------------------- create resourcequota ----------------------------------------------------------------------------------- + apply_resourcequota 7Gi + local quota=7Gi + local default_sc ++ get_default_storageclass ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KG1MFLbofZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.vMERWhtUCY ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.KG1MFLbofZ ++ cat /tmp/tmp.vMERWhtUCY ++ rm /tmp/tmp.KG1MFLbofZ /tmp/tmp.vMERWhtUCY ++ return 0 + default_sc=standard-rwo + echo 'Applying resourcequota for default storageclass standard-rwo with quota 7Gi' Applying resourcequota for default storageclass standard-rwo with quota 7Gi + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/e2e-tests/pvc-resize/conf/resourcequota.yml + sed s/STORAGECLASS/standard-rwo/ + sed s/QUOTA/7Gi/ + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.cMk0DS0Mdg ++ mktemp + local LAST_ERR=/tmp/tmp.faS050Nj44 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.cMk0DS0Mdg resourcequota/default-storage-quota created + cat /tmp/tmp.faS050Nj44 + rm /tmp/tmp.cMk0DS0Mdg /tmp/tmp.faS050Nj44 + return 0 + patch_pvc_request some-name 3G + local cluster=some-name + local size=3G + echo 'Patching PVC request to 3G in some-name' Patching PVC request to 3G in some-name + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"3G"}]' ++ mktemp + local LAST_OUT=/tmp/tmp.gcspnSm2ND ++ mktemp + local LAST_ERR=/tmp/tmp.WrMKjkSF23 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"3G"}]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.gcspnSm2ND perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.WrMKjkSF23 + rm /tmp/tmp.gcspnSm2ND /tmp/tmp.WrMKjkSF23 + return 0 + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YhCqhPZaBO +++ mktemp ++ local LAST_ERR=/tmp/tmp.1U2pLlzTnK ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.YhCqhPZaBO ++ cat /tmp/tmp.1U2pLlzTnK ++ rm /tmp/tmp.YhCqhPZaBO /tmp/tmp.1U2pLlzTnK ++ return 0 + [[ ready == ready ]] + echo .OK .OK + echo + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-0 3Gi + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-0 + local expected_size=3Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.my6WhwQku6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.WmOY9AOouc ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.my6WhwQku6 ++ cat /tmp/tmp.WmOY9AOouc ++ rm /tmp/tmp.my6WhwQku6 /tmp/tmp.WmOY9AOouc ++ return 0 + [[ 2Gi == 3Gi ]] + [[ 0 -ge 120 ]] + echo -n . .+ sleep 5 + retry=1 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uz0Ebskumn +++ mktemp ++ local LAST_ERR=/tmp/tmp.xXqKUku6RV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.uz0Ebskumn ++ cat /tmp/tmp.xXqKUku6RV ++ rm /tmp/tmp.uz0Ebskumn /tmp/tmp.xXqKUku6RV ++ return 0 + [[ 2Gi == 3Gi ]] + [[ 1 -ge 120 ]] + echo -n . .+ sleep 5 + retry=2 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RN4f7hbPSU +++ mktemp ++ local LAST_ERR=/tmp/tmp.bLphoDfhef ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.RN4f7hbPSU ++ cat /tmp/tmp.bLphoDfhef ++ rm /tmp/tmp.RN4f7hbPSU /tmp/tmp.bLphoDfhef ++ return 0 + [[ 2Gi == 3Gi ]] + [[ 2 -ge 120 ]] + echo -n . .+ sleep 5 + retry=3 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ellVrRNwxl +++ mktemp ++ local LAST_ERR=/tmp/tmp.ArAU7qB8ci ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ellVrRNwxl ++ cat /tmp/tmp.ArAU7qB8ci ++ rm /tmp/tmp.ellVrRNwxl /tmp/tmp.ArAU7qB8ci ++ return 0 + [[ 2Gi == 3Gi ]] + [[ 3 -ge 120 ]] + echo -n . .+ sleep 5 + retry=4 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2Tmm4rrJ6n +++ mktemp ++ local LAST_ERR=/tmp/tmp.C5u98cxBC5 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.2Tmm4rrJ6n ++ cat /tmp/tmp.C5u98cxBC5 ++ rm /tmp/tmp.2Tmm4rrJ6n /tmp/tmp.C5u98cxBC5 ++ return 0 + [[ 3Gi == 3Gi ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized + return 0 + apply_resourcequota 9Gi + local quota=9Gi + local default_sc ++ get_default_storageclass ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.C0aaIcvkjW +++ mktemp ++ local LAST_ERR=/tmp/tmp.mQKIoEiRwR ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.C0aaIcvkjW ++ cat /tmp/tmp.mQKIoEiRwR ++ rm /tmp/tmp.C0aaIcvkjW /tmp/tmp.mQKIoEiRwR ++ return 0 + default_sc=standard-rwo + echo 'Applying resourcequota for default storageclass standard-rwo with quota 9Gi' Applying resourcequota for default storageclass standard-rwo with quota 9Gi + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/e2e-tests/pvc-resize/conf/resourcequota.yml + sed s/STORAGECLASS/standard-rwo/ + sed s/QUOTA/9Gi/ + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.D5UlSiBwpg ++ mktemp + local LAST_ERR=/tmp/tmp.Rwgpv7X5p1 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.D5UlSiBwpg resourcequota/default-storage-quota configured + cat /tmp/tmp.Rwgpv7X5p1 + rm /tmp/tmp.D5UlSiBwpg /tmp/tmp.Rwgpv7X5p1 + return 0 + patch_pvc_request some-name 3G + local cluster=some-name + local size=3G + echo 'Patching PVC request to 3G in some-name' Patching PVC request to 3G in some-name + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"3G"}]' ++ mktemp + local LAST_OUT=/tmp/tmp.ZRGlEPEX4F ++ mktemp + local LAST_ERR=/tmp/tmp.AU5JSTKOwi + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"3G"}]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ZRGlEPEX4F perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.AU5JSTKOwi + rm /tmp/tmp.ZRGlEPEX4F /tmp/tmp.AU5JSTKOwi + return 0 + wait_cluster_consistency some-name 42 + local cluster_name=some-name + local wait_time=42 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mYYlseMjPG +++ mktemp ++ local LAST_ERR=/tmp/tmp.3nSkaOrYMA ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.mYYlseMjPG ++ cat /tmp/tmp.3nSkaOrYMA ++ rm /tmp/tmp.mYYlseMjPG /tmp/tmp.3nSkaOrYMA ++ return 0 + [[ initializing == ready ]] + let retry+=1 + '[' 1 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mE3Bu8mRTV +++ mktemp ++ local LAST_ERR=/tmp/tmp.uWxBkWK9BI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.mE3Bu8mRTV ++ cat /tmp/tmp.uWxBkWK9BI ++ rm /tmp/tmp.mE3Bu8mRTV /tmp/tmp.uWxBkWK9BI ++ return 0 + [[ initializing == ready ]] + let retry+=1 + '[' 2 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0WHT3rn1su +++ mktemp ++ local LAST_ERR=/tmp/tmp.jUGy9qmlkr ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.0WHT3rn1su ++ cat /tmp/tmp.jUGy9qmlkr ++ rm /tmp/tmp.0WHT3rn1su /tmp/tmp.jUGy9qmlkr ++ return 0 + [[ initializing == ready ]] + let retry+=1 + '[' 3 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pJhLqsvdQF +++ mktemp ++ local LAST_ERR=/tmp/tmp.p8GXS9a1sm ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.pJhLqsvdQF ++ cat /tmp/tmp.p8GXS9a1sm ++ rm /tmp/tmp.pJhLqsvdQF /tmp/tmp.p8GXS9a1sm ++ return 0 + [[ initializing == ready ]] + let retry+=1 + '[' 4 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.V8e9ettWbe +++ mktemp ++ local LAST_ERR=/tmp/tmp.R9EOylIdXC ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.V8e9ettWbe ++ cat /tmp/tmp.R9EOylIdXC ++ rm /tmp/tmp.V8e9ettWbe /tmp/tmp.R9EOylIdXC ++ return 0 + [[ initializing == ready ]] + let retry+=1 + '[' 5 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xyZK2cgW4r +++ mktemp ++ local LAST_ERR=/tmp/tmp.lEie40pkQV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xyZK2cgW4r ++ cat /tmp/tmp.lEie40pkQV ++ rm /tmp/tmp.xyZK2cgW4r /tmp/tmp.lEie40pkQV ++ return 0 + [[ initializing == ready ]] + let retry+=1 + '[' 6 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RzuixdinlC +++ mktemp ++ local LAST_ERR=/tmp/tmp.uiHfwZDnz7 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.RzuixdinlC ++ cat /tmp/tmp.uiHfwZDnz7 ++ rm /tmp/tmp.RzuixdinlC /tmp/tmp.uiHfwZDnz7 ++ return 0 + [[ initializing == ready ]] + let retry+=1 + '[' 7 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fuk1l6NLKX +++ mktemp ++ local LAST_ERR=/tmp/tmp.osPkIuTklo ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.fuk1l6NLKX ++ cat /tmp/tmp.osPkIuTklo ++ rm /tmp/tmp.fuk1l6NLKX /tmp/tmp.osPkIuTklo ++ return 0 + [[ initializing == ready ]] + let retry+=1 + '[' 8 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NkYDbDiPsZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZNlPWUMQmS ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.NkYDbDiPsZ ++ cat /tmp/tmp.ZNlPWUMQmS ++ rm /tmp/tmp.NkYDbDiPsZ /tmp/tmp.ZNlPWUMQmS ++ return 0 + [[ ready == ready ]] + echo .OK .OK + echo + wait_all_pvc_resize 3Gi + local expected_size=3Gi + local max_retry=120 + local sleep_time=5 ++ kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name +++ mktemp ++ local LAST_OUT=/tmp/tmp.TMthyHXd8H +++ mktemp ++ local LAST_ERR=/tmp/tmp.9DreForMsA ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc -l app.kubernetes.io/component=mongod -o name ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.TMthyHXd8H ++ cat /tmp/tmp.9DreForMsA ++ rm /tmp/tmp.TMthyHXd8H /tmp/tmp.9DreForMsA ++ return 0 + for pvc in $(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name) + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-0 3Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-0 + local expected_size=3Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.igFXKVrUFJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.LjiBZ9DqHq ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.igFXKVrUFJ ++ cat /tmp/tmp.LjiBZ9DqHq ++ rm /tmp/tmp.igFXKVrUFJ /tmp/tmp.LjiBZ9DqHq ++ return 0 + [[ 3Gi == 3Gi ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized + return 0 + for pvc in $(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name) + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-1 3Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-1 + local expected_size=3Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-1 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-1 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fM0bUsrgU1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.LHVWN2GyKF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.fM0bUsrgU1 ++ cat /tmp/tmp.LHVWN2GyKF ++ rm /tmp/tmp.fM0bUsrgU1 /tmp/tmp.LHVWN2GyKF ++ return 0 + [[ 3Gi == 3Gi ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-1 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-1 was resized + return 0 + for pvc in $(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name) + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-2 3Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-2 + local expected_size=3Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-2 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-2 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UsnF3QggWm +++ mktemp ++ local LAST_ERR=/tmp/tmp.sLyX5CGBiG ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.UsnF3QggWm ++ cat /tmp/tmp.sLyX5CGBiG ++ rm /tmp/tmp.UsnF3QggWm /tmp/tmp.sLyX5CGBiG ++ return 0 + [[ 3Gi == 3Gi ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-2 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-2 was resized + return 0 + return 0 + desc 'test downscale' + set +o xtrace ----------------------------------------------------------------------------------- test downscale ----------------------------------------------------------------------------------- + patch_pvc_request some-name 1G + local cluster=some-name + local size=1G + echo 'Patching PVC request to 1G in some-name' Patching PVC request to 1G in some-name + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"1G"}]' ++ mktemp + local LAST_OUT=/tmp/tmp.nRHNacfXoY ++ mktemp + local LAST_ERR=/tmp/tmp.6UxfD1SEQH + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"1G"}]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.nRHNacfXoY perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.6UxfD1SEQH + rm /tmp/tmp.nRHNacfXoY /tmp/tmp.6UxfD1SEQH + return 0 + wait_cluster_status some-name error + local cluster=some-name + local expected=error + local retry=0 + echo -n 'Waiting for psmdb/some-name status to be error' Waiting for psmdb/some-name status to be error++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZVefvkDiVH +++ mktemp ++ local LAST_ERR=/tmp/tmp.UWQVXhHqEj ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ZVefvkDiVH ++ cat /tmp/tmp.UWQVXhHqEj ++ rm /tmp/tmp.ZVefvkDiVH /tmp/tmp.UWQVXhHqEj ++ return 0 + [[ error == error ]] + echo + echo 'psmdb/some-name status is error' psmdb/some-name status is error + patch_pvc_request some-name 3G + local cluster=some-name + local size=3G + echo 'Patching PVC request to 3G in some-name' Patching PVC request to 3G in some-name + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"3G"}]' ++ mktemp + local LAST_OUT=/tmp/tmp.Ea7pDRzNZn ++ mktemp + local LAST_ERR=/tmp/tmp.2yauUi1rag + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"3G"}]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Ea7pDRzNZn perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.2yauUi1rag + rm /tmp/tmp.Ea7pDRzNZn /tmp/tmp.2yauUi1rag + return 0 + wait_cluster_status some-name ready + local cluster=some-name + local expected=ready + local retry=0 + echo -n 'Waiting for psmdb/some-name status to be ready' Waiting for psmdb/some-name status to be ready++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ky45u4pPMR +++ mktemp ++ local LAST_ERR=/tmp/tmp.HLIyhJPpCB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ky45u4pPMR ++ cat /tmp/tmp.HLIyhJPpCB ++ rm /tmp/tmp.ky45u4pPMR /tmp/tmp.HLIyhJPpCB ++ return 0 + [[ error == ready ]] + [[ 0 -ge 60 ]] + echo -n . .+ sleep 5 + retry=1 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3rVqoTSuZv +++ mktemp ++ local LAST_ERR=/tmp/tmp.QfZIvm94Y4 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.3rVqoTSuZv ++ cat /tmp/tmp.QfZIvm94Y4 ++ rm /tmp/tmp.3rVqoTSuZv /tmp/tmp.QfZIvm94Y4 ++ return 0 + [[ ready == ready ]] + echo + echo 'psmdb/some-name status is ready' psmdb/some-name status is ready + [[ 0 == 1 ]] + [[ -n '' ]] + desc 'test restartedAt annotation preservation during PVC resize' + set +o xtrace ----------------------------------------------------------------------------------- test restartedAt annotation preservation during PVC resize ----------------------------------------------------------------------------------- + echo 'delete resourcequota' delete resourcequota + kubectl_bin delete resourcequota default-storage-quota ++ mktemp + local LAST_OUT=/tmp/tmp.a0135Ivazw ++ mktemp + local LAST_ERR=/tmp/tmp.RN6HC5qw7p + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete resourcequota default-storage-quota + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.a0135Ivazw resourcequota "default-storage-quota" deleted from pvc-resize-2243 namespace + cat /tmp/tmp.RN6HC5qw7p + rm /tmp/tmp.a0135Ivazw /tmp/tmp.RN6HC5qw7p + return 0 + sts_name=some-name-rs0 + kubectl_bin rollout restart statefulset/some-name-rs0 ++ mktemp + local LAST_OUT=/tmp/tmp.b3rg1GsVTE ++ mktemp + local LAST_ERR=/tmp/tmp.tDamE9F1HX + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl rollout restart statefulset/some-name-rs0 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.b3rg1GsVTE statefulset.apps/some-name-rs0 restarted + cat /tmp/tmp.tDamE9F1HX + rm /tmp/tmp.b3rg1GsVTE /tmp/tmp.tDamE9F1HX + return 0 + wait_for_running some-name-rs0 3 + local name=some-name-rs0 + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=rs0 + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-rs0-0 + local pod=some-name-rs0-0 + set +o xtrace waiting for pod/some-name-rs0-0 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-rs0-1 + local pod=some-name-rs0-1 + set +o xtrace waiting for pod/some-name-rs0-1 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.K8YAGKXPvl +++ mktemp ++ local LAST_ERR=/tmp/tmp.SVpDJcMMIF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.K8YAGKXPvl ++ cat /tmp/tmp.SVpDJcMMIF ++ rm /tmp/tmp.K8YAGKXPvl /tmp/tmp.SVpDJcMMIF ++ return 0 + [[ false == true ]] + wait_pod some-name-rs0-2 + local pod=some-name-rs0-2 + set +o xtrace waiting for pod/some-name-rs0-2 to be ready.OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fae6QIM5Bv +++ mktemp ++ local LAST_ERR=/tmp/tmp.tgiBl3s6oa ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.fae6QIM5Bv ++ cat /tmp/tmp.tgiBl3s6oa ++ rm /tmp/tmp.fae6QIM5Bv /tmp/tmp.tgiBl3s6oa ++ return 0 + [[ false == true ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.47EtW2R1n4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.79W942mwha ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.47EtW2R1n4 ++ cat /tmp/tmp.79W942mwha ++ rm /tmp/tmp.47EtW2R1n4 /tmp/tmp.79W942mwha ++ return 0 + [[ false == true ]] + sleep 10 + [[ true == true ]] + set +x Waiting for cluster readyness............................................. ++ kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={.spec.template.metadata.annotations.kubectl\.kubernetes\.io/restartedAt}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nI2A0jhgNy +++ mktemp ++ local LAST_ERR=/tmp/tmp.LX3tXFsOnf ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get statefulset some-name-rs0 -o 'jsonpath={.spec.template.metadata.annotations.kubectl\.kubernetes\.io/restartedAt}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.nI2A0jhgNy ++ cat /tmp/tmp.LX3tXFsOnf ++ rm /tmp/tmp.nI2A0jhgNy /tmp/tmp.LX3tXFsOnf ++ return 0 + restarted_at_before=2026-04-23T10:06:04Z + '[' -z 2026-04-23T10:06:04Z ']' + patch_pvc_request some-name 4G + local cluster=some-name + local size=4G + echo 'Patching PVC request to 4G in some-name' Patching PVC request to 4G in some-name + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"4G"}]' ++ mktemp + local LAST_OUT=/tmp/tmp.sa0Nszb6hE ++ mktemp + local LAST_ERR=/tmp/tmp.tQCZdtPTxs + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"4G"}]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.sa0Nszb6hE perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.tQCZdtPTxs + rm /tmp/tmp.sa0Nszb6hE /tmp/tmp.tQCZdtPTxs + return 0 + wait_all_pvc_resize 4Gi + local expected_size=4Gi + local max_retry=120 + local sleep_time=5 ++ kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name +++ mktemp ++ local LAST_OUT=/tmp/tmp.3k6ZiuhfI7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.G14S6hkGBN ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc -l app.kubernetes.io/component=mongod -o name ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.3k6ZiuhfI7 ++ cat /tmp/tmp.G14S6hkGBN ++ rm /tmp/tmp.3k6ZiuhfI7 /tmp/tmp.G14S6hkGBN ++ return 0 + for pvc in $(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name) + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-0 4Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-0 + local expected_size=4Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sLM8qiJza2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.AfQNooPVlZ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.sLM8qiJza2 ++ cat /tmp/tmp.AfQNooPVlZ ++ rm /tmp/tmp.sLM8qiJza2 /tmp/tmp.AfQNooPVlZ ++ return 0 + [[ 3Gi == 4Gi ]] + [[ 0 -ge 120 ]] + echo -n . .+ sleep 5 + retry=1 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DvI6OzEM33 +++ mktemp ++ local LAST_ERR=/tmp/tmp.rvLqCcrQ93 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.DvI6OzEM33 ++ cat /tmp/tmp.rvLqCcrQ93 ++ rm /tmp/tmp.DvI6OzEM33 /tmp/tmp.rvLqCcrQ93 ++ return 0 + [[ 3Gi == 4Gi ]] + [[ 1 -ge 120 ]] + echo -n . .+ sleep 5 + retry=2 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.11FRjfmLlH +++ mktemp ++ local LAST_ERR=/tmp/tmp.8aFbHyW90z ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.11FRjfmLlH ++ cat /tmp/tmp.8aFbHyW90z ++ rm /tmp/tmp.11FRjfmLlH /tmp/tmp.8aFbHyW90z ++ return 0 + [[ 3Gi == 4Gi ]] + [[ 2 -ge 120 ]] + echo -n . .+ sleep 5 + retry=3 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0Xzb0I7heb +++ mktemp ++ local LAST_ERR=/tmp/tmp.iAK1EJmpYi ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.0Xzb0I7heb ++ cat /tmp/tmp.iAK1EJmpYi ++ rm /tmp/tmp.0Xzb0I7heb /tmp/tmp.iAK1EJmpYi ++ return 0 + [[ 3Gi == 4Gi ]] + [[ 3 -ge 120 ]] + echo -n . .+ sleep 5 + retry=4 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nbNSKhNE0T +++ mktemp ++ local LAST_ERR=/tmp/tmp.AHRaOhPEIx ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.nbNSKhNE0T ++ cat /tmp/tmp.AHRaOhPEIx ++ rm /tmp/tmp.nbNSKhNE0T /tmp/tmp.AHRaOhPEIx ++ return 0 + [[ 3Gi == 4Gi ]] + [[ 4 -ge 120 ]] + echo -n . .+ sleep 5 + retry=5 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wy088PwiGd +++ mktemp ++ local LAST_ERR=/tmp/tmp.AeE9N5Fk6h ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.wy088PwiGd ++ cat /tmp/tmp.AeE9N5Fk6h ++ rm /tmp/tmp.wy088PwiGd /tmp/tmp.AeE9N5Fk6h ++ return 0 + [[ 3Gi == 4Gi ]] + [[ 5 -ge 120 ]] + echo -n . .+ sleep 5 + retry=6 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IubdSmgDK7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.KQ0OndtvSL ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.IubdSmgDK7 ++ cat /tmp/tmp.KQ0OndtvSL ++ rm /tmp/tmp.IubdSmgDK7 /tmp/tmp.KQ0OndtvSL ++ return 0 + [[ 3Gi == 4Gi ]] + [[ 6 -ge 120 ]] + echo -n . .+ sleep 5 + retry=7 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MjVzSP6ALa +++ mktemp ++ local LAST_ERR=/tmp/tmp.0I9WFTRaKo ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.MjVzSP6ALa ++ cat /tmp/tmp.0I9WFTRaKo ++ rm /tmp/tmp.MjVzSP6ALa /tmp/tmp.0I9WFTRaKo ++ return 0 + [[ 3Gi == 4Gi ]] + [[ 7 -ge 120 ]] + echo -n . .+ sleep 5 + retry=8 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.59UY4FCwxY +++ mktemp ++ local LAST_ERR=/tmp/tmp.YfBDoku1rK ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.59UY4FCwxY ++ cat /tmp/tmp.YfBDoku1rK ++ rm /tmp/tmp.59UY4FCwxY /tmp/tmp.YfBDoku1rK ++ return 0 + [[ 3Gi == 4Gi ]] + [[ 8 -ge 120 ]] + echo -n . .+ sleep 5 + retry=9 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Cxesl9AbS7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.OSKfTelWe7 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Cxesl9AbS7 ++ cat /tmp/tmp.OSKfTelWe7 ++ rm /tmp/tmp.Cxesl9AbS7 /tmp/tmp.OSKfTelWe7 ++ return 0 + [[ 3Gi == 4Gi ]] + [[ 9 -ge 120 ]] + echo -n . .+ sleep 5 + retry=10 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tNx8B1lvnQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.7R6LrPWQL5 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.tNx8B1lvnQ ++ cat /tmp/tmp.7R6LrPWQL5 ++ rm /tmp/tmp.tNx8B1lvnQ /tmp/tmp.7R6LrPWQL5 ++ return 0 + [[ 4Gi == 4Gi ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized + return 0 + for pvc in $(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name) + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-1 4Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-1 + local expected_size=4Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-1 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-1 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uuy7icHODW +++ mktemp ++ local LAST_ERR=/tmp/tmp.euyPZaGuDQ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.uuy7icHODW ++ cat /tmp/tmp.euyPZaGuDQ ++ rm /tmp/tmp.uuy7icHODW /tmp/tmp.euyPZaGuDQ ++ return 0 + [[ 4Gi == 4Gi ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-1 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-1 was resized + return 0 + for pvc in $(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name) + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-2 4Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-2 + local expected_size=4Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-2 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-2 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7J8LizdQct +++ mktemp ++ local LAST_ERR=/tmp/tmp.EsmkXKHRKP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.7J8LizdQct ++ cat /tmp/tmp.EsmkXKHRKP ++ rm /tmp/tmp.7J8LizdQct /tmp/tmp.EsmkXKHRKP ++ return 0 + [[ 4Gi == 4Gi ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-2 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-2 was resized + return 0 + return 0 + sleep_with_log 30 'wait for StatefulSet recreation after orphan delete during PVC resize' + local d=30 + local 'msg=wait for StatefulSet recreation after orphan delete during PVC resize' + log 'wait for StatefulSet recreation after orphan delete during PVC resize. sleeping for 30 seconds' + set +o xtrace [2026-04-23T10:09:01+0000] wait for StatefulSet recreation after orphan delete during PVC resize. sleeping for 30 seconds + sleep 30 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.L2KAmytqaJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.oGLG9eafxv ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.L2KAmytqaJ ++ cat /tmp/tmp.oGLG9eafxv ++ rm /tmp/tmp.L2KAmytqaJ /tmp/tmp.oGLG9eafxv ++ return 0 + cluster_status=ready ++ kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={.spec.template.metadata.annotations.kubectl\.kubernetes\.io/restartedAt}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WmI7aBzjMX +++ mktemp ++ local LAST_ERR=/tmp/tmp.r3sxuWco66 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get statefulset some-name-rs0 -o 'jsonpath={.spec.template.metadata.annotations.kubectl\.kubernetes\.io/restartedAt}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.WmI7aBzjMX ++ cat /tmp/tmp.r3sxuWco66 ++ rm /tmp/tmp.WmI7aBzjMX /tmp/tmp.r3sxuWco66 ++ return 0 + restarted_at_after=2026-04-23T10:06:04Z + [[ 2026-04-23T10:06:04Z != 2026\-04\-23T10\:06\:04Z ]] + [[ ready == initializing ]] + wait_cluster_status some-name ready + local cluster=some-name + local expected=ready + local retry=0 + echo -n 'Waiting for psmdb/some-name status to be ready' Waiting for psmdb/some-name status to be ready++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QDWoqENlWr +++ mktemp ++ local LAST_ERR=/tmp/tmp.7oT1FbVrla ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.QDWoqENlWr ++ cat /tmp/tmp.7oT1FbVrla ++ rm /tmp/tmp.QDWoqENlWr /tmp/tmp.7oT1FbVrla ++ return 0 + [[ ready == ready ]] + echo + echo 'psmdb/some-name status is ready' psmdb/some-name status is ready + wait_for_running some-name-rs0 3 + local name=some-name-rs0 + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=rs0 + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-rs0-0 + local pod=some-name-rs0-0 + set +o xtrace waiting for pod/some-name-rs0-0 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-rs0-1 + local pod=some-name-rs0-1 + set +o xtrace waiting for pod/some-name-rs0-1 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GF49DL4zl8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.qna4oQ0rXZ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.GF49DL4zl8 ++ cat /tmp/tmp.qna4oQ0rXZ ++ rm /tmp/tmp.GF49DL4zl8 /tmp/tmp.qna4oQ0rXZ ++ return 0 + [[ false == true ]] + wait_pod some-name-rs0-2 + local pod=some-name-rs0-2 + set +o xtrace waiting for pod/some-name-rs0-2 to be ready.OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.aHmQed7UX8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.KTyk5KwYHt ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.aHmQed7UX8 ++ cat /tmp/tmp.KTyk5KwYHt ++ rm /tmp/tmp.aHmQed7UX8 /tmp/tmp.KTyk5KwYHt ++ return 0 + [[ false == true ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1pA1gfodXv +++ mktemp ++ local LAST_ERR=/tmp/tmp.WuFqnM9PLn ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.1pA1gfodXv ++ cat /tmp/tmp.WuFqnM9PLn ++ rm /tmp/tmp.1pA1gfodXv /tmp/tmp.WuFqnM9PLn ++ return 0 + [[ false == true ]] + sleep 10 + [[ true == true ]] + set +x Waiting for cluster readyness + destroy pvc-resize-2243 + local namespace=pvc-resize-2243 + local ignore_logs=true + [[ 0 == 1 ]] + desc 'destroy cluster/operator and all other resources' + set +o xtrace ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + '[' true == false ']' + delete_backups + desc 'Delete psmdb-backup' + set +o xtrace ----------------------------------------------------------------------------------- Delete psmdb-backup ----------------------------------------------------------------------------------- ++ kubectl_bin get psmdb-backup --no-headers ++ wc -l +++ mktemp ++ local LAST_OUT=/tmp/tmp.VgLDjJ3Fnz +++ mktemp ++ local LAST_ERR=/tmp/tmp.x9SO9wNoXK ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb-backup --no-headers ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.VgLDjJ3Fnz ++ cat /tmp/tmp.x9SO9wNoXK No resources found in pvc-resize-2243 namespace. ++ rm /tmp/tmp.VgLDjJ3Fnz /tmp/tmp.x9SO9wNoXK ++ return 0 + '[' 0 '!=' 0 ']' + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.GgnTSasiOp ++ mktemp + local LAST_ERR=/tmp/tmp.MzpYMxzbIy + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.GgnTSasiOp customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.MzpYMxzbIy + rm /tmp/tmp.GgnTSasiOp /tmp/tmp.MzpYMxzbIy + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/deploy/crd.yaml ++ grep -v '\-\-\-' grep: warning: stray \ before - grep: warning: stray \ before - + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.wRDwAxXanp ++ mktemp + local LAST_ERR=/tmp/tmp.qMyVd2lOIS + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.wRDwAxXanp + cat /tmp/tmp.qMyVd2lOIS + rm /tmp/tmp.wRDwAxXanp /tmp/tmp.qMyVd2lOIS + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.pbX9UFX2bm ++ mktemp + local LAST_ERR=/tmp/tmp.JukgiV0F9R + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.pbX9UFX2bm + cat /tmp/tmp.JukgiV0F9R + rm /tmp/tmp.pbX9UFX2bm /tmp/tmp.JukgiV0F9R + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch perconaservermongodbs.psmdb.percona.com -n pvc-resize-2243 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaservermongodb.psmdb.percona.com/some-name patched + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.mLzEa0OsZV ++ mktemp + local LAST_ERR=/tmp/tmp.1D9TeJj1Yb + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.mLzEa0OsZV customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com condition met + cat /tmp/tmp.1D9TeJj1Yb + rm /tmp/tmp.mLzEa0OsZV /tmp/tmp.1D9TeJj1Yb + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.sItcuQ4nz0 ++ mktemp + local LAST_ERR=/tmp/tmp.wN3aTKxr3M + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2274/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.sItcuQ4nz0 clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.wN3aTKxr3M + rm /tmp/tmp.sItcuQ4nz0 /tmp/tmp.wN3aTKxr3M + return 0 + destroy_cert_manager + kubectl_bin delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.JcXnTbbjjo ++ mktemp + local LAST_ERR=/tmp/tmp.mPbTyJ91bP + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.JcXnTbbjjo + cat /tmp/tmp.mPbTyJ91bP Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 0 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.JcXnTbbjjo + cat /tmp/tmp.mPbTyJ91bP Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 4 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.JcXnTbbjjo + cat /tmp/tmp.mPbTyJ91bP Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 8 + cat /tmp/tmp.JcXnTbbjjo + cat /tmp/tmp.mPbTyJ91bP Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + rm /tmp/tmp.JcXnTbbjjo /tmp/tmp.mPbTyJ91bP + return 1 + true + '[' -n '' ']' + '[' -n psmdb-operator ']' + kubectl_bin delete --grace-period=0 --force=true namespace pvc-resize-2243 + rm -rf /tmp/tmp.GLPwH4gC3u + kubectl_bin delete --grace-period=0 --force=true namespace psmdb-operator ++ mktemp ++ mktemp + desc 'test passed' + set +o xtrace ----------------------------------------------------------------------------------- test passed ----------------------------------------------------------------------------------- + local LAST_OUT=/tmp/tmp.tkonrJbOgY + local LAST_OUT=/tmp/tmp.pRgqYEzDAV ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.8zOkWScDGI + local exit_status=0 + local timeout=4 ++ seq 0 2 + local LAST_ERR=/tmp/tmp.llkpu987Qw + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + for i in $(seq 0 2) + kubectl delete --grace-period=0 --force=true namespace psmdb-operator + set +e + kubectl delete --grace-period=0 --force=true namespace pvc-resize-2243