Log: /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/e2e-tests/logs/pvc-resize.log WARNING: version difference between client (1.33) and server (1.30) exceeds the supported minor version skew of +/-1 WARNING: version difference between client (1.33) and server (1.30) exceeds the supported minor version skew of +/-1 WARNING: version difference between client (1.33) and server (1.30) exceeds the supported minor version skew of +/-1 + '[' 0 == 1 ']' + ensure_default_sc_allows_expansion + local default_sc ++ get_default_storageclass ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nF6cHFtNpT +++ mktemp ++ local LAST_ERR=/tmp/tmp.k2rNSBwmXa ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.nF6cHFtNpT ++ cat /tmp/tmp.k2rNSBwmXa ++ rm /tmp/tmp.nF6cHFtNpT /tmp/tmp.k2rNSBwmXa ++ return 0 + default_sc=standard-rwo + echo 'Checking if default storageclass standard-rwo allows volume expansion' Checking if default storageclass standard-rwo allows volume expansion + local allowVolumeExpansion ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.name=="standard-rwo")].allowVolumeExpansion}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.x7jJGGKO67 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Dt6PrAVii4 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.name=="standard-rwo")].allowVolumeExpansion}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.x7jJGGKO67 ++ cat /tmp/tmp.Dt6PrAVii4 ++ rm /tmp/tmp.x7jJGGKO67 /tmp/tmp.Dt6PrAVii4 ++ return 0 + allowVolumeExpansion=true + [[ true != \t\r\u\e ]] + create_infra pvc-resize-5656 + local ns=pvc-resize-5656 + [[ 1 == 1 ]] + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.0M5nRK44yN ++ mktemp + local LAST_ERR=/tmp/tmp.YOhRrWYINw + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.0M5nRK44yN customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.YOhRrWYINw + rm /tmp/tmp.0M5nRK44yN /tmp/tmp.YOhRrWYINw + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/deploy/crd.yaml ++ grep -v '\-\-\-' + for crd_name in '$(yq eval '\''.metadata.name'\'' "${src_dir}/deploy/crd.yaml" | grep -v '\''\-\-\-'\'')' + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + grep -v NAMESPACE error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.D1Vfv4Yl9v ++ mktemp + local LAST_ERR=/tmp/tmp.1eySs4jiiD + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.D1Vfv4Yl9v + cat /tmp/tmp.1eySs4jiiD + rm /tmp/tmp.D1Vfv4Yl9v /tmp/tmp.1eySs4jiiD + return 0 + for crd_name in '$(yq eval '\''.metadata.name'\'' "${src_dir}/deploy/crd.yaml" | grep -v '\''\-\-\-'\'')' + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + grep -v NAMESPACE + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.EhUk8tFask ++ mktemp + local LAST_ERR=/tmp/tmp.UEwqvnDCCP + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.EhUk8tFask + cat /tmp/tmp.UEwqvnDCCP + rm /tmp/tmp.EhUk8tFask /tmp/tmp.UEwqvnDCCP + return 0 + for crd_name in '$(yq eval '\''.metadata.name'\'' "${src_dir}/deploy/crd.yaml" | grep -v '\''\-\-\-'\'')' + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + grep -v NAMESPACE No resources found + kubectl patch perconaservermongodbs.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbs" + : + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.iuKLJ9CIvz ++ mktemp + local LAST_ERR=/tmp/tmp.fLyZXT92YH + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.iuKLJ9CIvz + cat /tmp/tmp.fLyZXT92YH + rm /tmp/tmp.iuKLJ9CIvz /tmp/tmp.fLyZXT92YH + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.q9UmrfKTno ++ mktemp + local LAST_ERR=/tmp/tmp.8zA1AptfcE + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.q9UmrfKTno clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.8zA1AptfcE + rm /tmp/tmp.q9UmrfKTno /tmp/tmp.8zA1AptfcE + return 0 + check_crd_for_deletion PR-1926-d4d89df0 + local git_tag=PR-1926-d4d89df0 ++ yq eval .metadata.name ++ curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/PR-1926-d4d89df0/deploy/crd.yaml ++ /usr/bin/sed s/---//g ++ /usr/bin/sed ':a;N;$!ba;s/\n/ /g' + for crd_name in '$(curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/${git_tag}/deploy/crd.yaml | yq eval '\''.metadata.name'\'' | $sed '\''s/---//g'\'' | $sed '\'':a;N;$!ba;s/\n/ /g'\'')' ++ kubectl_bin get crd/null -o 'jsonpath={.status.conditions[-1].type}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MrGEqDPfDD +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZDZc7mNRcz ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.MrGEqDPfDD ++ cat /tmp/tmp.ZDZc7mNRcz Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 0 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.MrGEqDPfDD ++ cat /tmp/tmp.ZDZc7mNRcz Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 4 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.MrGEqDPfDD ++ cat /tmp/tmp.ZDZc7mNRcz Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 8 ++ cat /tmp/tmp.MrGEqDPfDD ++ cat /tmp/tmp.ZDZc7mNRcz Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ rm /tmp/tmp.MrGEqDPfDD /tmp/tmp.ZDZc7mNRcz ++ return 1 + [[ '' == \T\e\r\m\i\n\a\t\i\n\g ]] + '[' -n psmdb-operator ']' + create_namespace psmdb-operator + local namespace=psmdb-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ tail -n1 ++ helm list --all-namespaces --filter chaos-mesh ++ sed s/NAMESPACE// ++ awk '-F ' '{print $2}' + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get crd ++ awk '{print $1}' ++ grep chaos-mesh.org + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl get clusterrole ++ grep chaos-mesh + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + '[' -n '' ']' + desc 'cleaned up old namespaces psmdb-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace psmdb-operator --ignore-not-found + awk '{print$1}' ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.jLXtT4gNIf + local LAST_OUT=/tmp/tmp.m4tY5BVZfF + egrep -v '^kube-|^default|Terminating|psmdb-operator|openshift|gke-mcs|^NAME' ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.7lpyknJguW + local exit_status=0 + local timeout=4 + local LAST_ERR=/tmp/tmp.Ml2RQI64wD + local exit_status=0 + local timeout=4 ++ seq 0 2 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace psmdb-operator --ignore-not-found + for i in '$(seq 0 2)' + set +e + kubectl get ns + xargs kubectl delete ns + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.m4tY5BVZfF + cat /tmp/tmp.7lpyknJguW + rm /tmp/tmp.m4tY5BVZfF /tmp/tmp.7lpyknJguW + return 0 namespace "gke-managed-cim" deleted namespace "gke-managed-system" deleted namespace "gmp-public" deleted namespace "gmp-system" deleted namespace "pvc-resize-26492" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.jLXtT4gNIf namespace "psmdb-operator" deleted + cat /tmp/tmp.Ml2RQI64wD + rm /tmp/tmp.jLXtT4gNIf /tmp/tmp.Ml2RQI64wD + return 0 + kubectl_bin wait --for=delete namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.2AjJ7mCEtO ++ mktemp + local LAST_ERR=/tmp/tmp.huR6zI3YoU + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.2AjJ7mCEtO + cat /tmp/tmp.huR6zI3YoU + rm /tmp/tmp.2AjJ7mCEtO /tmp/tmp.huR6zI3YoU + return 0 + desc 'create namespace psmdb-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.wgKaLcVIGH ++ mktemp + local LAST_ERR=/tmp/tmp.D71yXg4Kxa + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.wgKaLcVIGH namespace/psmdb-operator created + cat /tmp/tmp.D71yXg4Kxa + rm /tmp/tmp.wgKaLcVIGH /tmp/tmp.D71yXg4Kxa + return 0 + set_kube_ctx psmdb-operator + local namespace=psmdb-operator ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.bKnKwhAT3k +++ mktemp ++ local LAST_ERR=/tmp/tmp.rlCdX4jUma ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.bKnKwhAT3k ++ cat /tmp/tmp.rlCdX4jUma ++ rm /tmp/tmp.bKnKwhAT3k /tmp/tmp.rlCdX4jUma ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-1926-d4d89df0-31-cluster6 --namespace=psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.SdIbdpG6Bu ++ mktemp + local LAST_ERR=/tmp/tmp.N5MBsE2fGU + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-1926-d4d89df0-31-cluster6 --namespace=psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.SdIbdpG6Bu Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-1926-d4d89df0-31-cluster6" modified. + cat /tmp/tmp.N5MBsE2fGU + rm /tmp/tmp.SdIbdpG6Bu /tmp/tmp.N5MBsE2fGU + return 0 + deploy_operator + desc 'start PSMDB operator' + set +o xtrace ----------------------------------------------------------------------------------- start PSMDB operator ----------------------------------------------------------------------------------- + local cr_file + '[' -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/e2e-tests/pvc-resize/conf/crd.yaml ']' + cr_file=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/deploy/crd.yaml + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.KIbjWrNr5y ++ mktemp + local LAST_ERR=/tmp/tmp.wxitciMUZr + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.KIbjWrNr5y customresourcedefinition.apiextensions.k8s.io/perconaservermongodbbackups.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbrestores.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com serverside-applied + cat /tmp/tmp.wxitciMUZr + rm /tmp/tmp.KIbjWrNr5y /tmp/tmp.wxitciMUZr + return 0 + '[' -n psmdb-operator ']' + apply_rbac cw-rbac + local operator_namespace=psmdb-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: psmdb-operator^' + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.8Qib7D4dyf ++ mktemp + local LAST_ERR=/tmp/tmp.qmkLloB5Dw + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.8Qib7D4dyf clusterrole.rbac.authorization.k8s.io/percona-server-mongodb-operator created serviceaccount/percona-server-mongodb-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-server-mongodb-operator created + cat /tmp/tmp.qmkLloB5Dw + rm /tmp/tmp.8Qib7D4dyf /tmp/tmp.qmkLloB5Dw + return 0 + kubectl_bin apply -f - + yq eval ' (.spec.template.spec.containers[].image = "perconalab/percona-server-mongodb-operator:PR-1926-d4d89df0") | ((.. | select(.[] == "DISABLE_TELEMETRY")) |= .value="true") | ((.. | select(.[] == "LOG_LEVEL")) |= .value="DEBUG")' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/deploy/cw-operator.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.9bNfqBJqb6 ++ mktemp + local LAST_ERR=/tmp/tmp.QLTLtAOjmw + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.9bNfqBJqb6 deployment.apps/percona-server-mongodb-operator created + cat /tmp/tmp.QLTLtAOjmw + rm /tmp/tmp.9bNfqBJqb6 /tmp/tmp.QLTLtAOjmw + return 0 + sleep 2 ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.gTPqAEQk5m +++ mktemp ++ local LAST_ERR=/tmp/tmp.5xo2kzluyI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.gTPqAEQk5m ++ cat /tmp/tmp.5xo2kzluyI ++ rm /tmp/tmp.gTPqAEQk5m /tmp/tmp.5xo2kzluyI ++ return 0 + wait_pod percona-server-mongodb-operator-c5895859f-gzz5b + local pod=percona-server-mongodb-operator-c5895859f-gzz5b + set +o xtrace waiting for pod/percona-server-mongodb-operator-c5895859f-gzz5b to be ready.OK + echo 'Print operator info from log' Print operator info from log + grep 'Manager starting up' ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.TsQFlwBOlA +++ mktemp ++ local LAST_ERR=/tmp/tmp.PqlWxLCFSM ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.TsQFlwBOlA ++ cat /tmp/tmp.PqlWxLCFSM ++ rm /tmp/tmp.TsQFlwBOlA /tmp/tmp.PqlWxLCFSM ++ return 0 + kubectl_bin logs percona-server-mongodb-operator-c5895859f-gzz5b ++ mktemp + local LAST_OUT=/tmp/tmp.qZPWGXg9Wz ++ mktemp + local LAST_ERR=/tmp/tmp.oS5cwd21kC + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl logs percona-server-mongodb-operator-c5895859f-gzz5b + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.qZPWGXg9Wz + cat /tmp/tmp.oS5cwd21kC + rm /tmp/tmp.qZPWGXg9Wz /tmp/tmp.oS5cwd21kC + return 0 2025-06-11T13:06:56.002Z INFO setup Manager starting up {"gitCommit": "d4d89df06387c83c75ec97fa349bfd60d81a9afc", "gitBranch": "PR-1926-d4d89df0", "buildTime": "", "goVersion": "go1.24.4", "os": "linux", "arch": "amd64"} + create_namespace pvc-resize-5656 + local namespace=pvc-resize-5656 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// ++ tail -n1 + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get ValidatingWebhookConfiguration + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep validate-auth ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ grep chaos-mesh.org ++ kubectl get crd ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + '[' -n '' ']' + desc 'cleaned up old namespaces pvc-resize-5656' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pvc-resize-5656 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pvc-resize-5656 --ignore-not-found + awk '{print$1}' + egrep -v '^kube-|^default|Terminating|psmdb-operator|openshift|gke-mcs|^NAME' ++ mktemp + kubectl_bin get ns + local LAST_OUT=/tmp/tmp.wdkfRlHvgL ++ mktemp + local LAST_ERR=/tmp/tmp.0VGt6uUVsA + local exit_status=0 + local timeout=4 ++ mktemp + xargs kubectl delete ns + local LAST_OUT=/tmp/tmp.hodN9AFAqP ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pvc-resize-5656 --ignore-not-found ++ mktemp + local LAST_ERR=/tmp/tmp.AJML1YclrQ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.wdkfRlHvgL + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.hodN9AFAqP + cat /tmp/tmp.0VGt6uUVsA + cat /tmp/tmp.AJML1YclrQ + rm /tmp/tmp.wdkfRlHvgL /tmp/tmp.0VGt6uUVsA + rm /tmp/tmp.hodN9AFAqP /tmp/tmp.AJML1YclrQ + return 0 + kubectl_bin wait --for=delete namespace pvc-resize-5656 + return 0 ++ mktemp + local LAST_OUT=/tmp/tmp.d0pnvrtVKh ++ mktemp + local LAST_ERR=/tmp/tmp.woknIYeDOy + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete namespace pvc-resize-5656 namespace "gke-managed-cim" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.d0pnvrtVKh + cat /tmp/tmp.woknIYeDOy + rm /tmp/tmp.d0pnvrtVKh /tmp/tmp.woknIYeDOy namespace "gke-managed-system" deleted + return 0 + desc 'create namespace pvc-resize-5656' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pvc-resize-5656 ----------------------------------------------------------------------------------- + kubectl_bin create namespace pvc-resize-5656 ++ mktemp + local LAST_OUT=/tmp/tmp.rWb1yLAvUW ++ mktemp + local LAST_ERR=/tmp/tmp.fOqBOejOUO + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pvc-resize-5656 namespace "gmp-public" deleted namespace "gmp-system" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.rWb1yLAvUW namespace/pvc-resize-5656 created + cat /tmp/tmp.fOqBOejOUO + rm /tmp/tmp.rWb1yLAvUW /tmp/tmp.fOqBOejOUO + return 0 + set_kube_ctx pvc-resize-5656 + local namespace=pvc-resize-5656 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.SYIT6G3F5m +++ mktemp ++ local LAST_ERR=/tmp/tmp.SOgzs4S1fq ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.SYIT6G3F5m ++ cat /tmp/tmp.SOgzs4S1fq ++ rm /tmp/tmp.SYIT6G3F5m /tmp/tmp.SOgzs4S1fq ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-1926-d4d89df0-31-cluster6 --namespace=pvc-resize-5656 ++ mktemp + local LAST_OUT=/tmp/tmp.5xFsz7z9fk ++ mktemp + local LAST_ERR=/tmp/tmp.f3UHsmd3Rn + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-1926-d4d89df0-31-cluster6 --namespace=pvc-resize-5656 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.5xFsz7z9fk Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-1926-d4d89df0-31-cluster6" modified. + cat /tmp/tmp.f3UHsmd3Rn + rm /tmp/tmp.5xFsz7z9fk /tmp/tmp.f3UHsmd3Rn + return 0 + desc 'create secrets and psmdb client' + set +o xtrace ----------------------------------------------------------------------------------- create secrets and psmdb client ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/e2e-tests/conf/secrets.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.NbORX47zSH ++ mktemp + local LAST_ERR=/tmp/tmp.6VQrsjhWBo + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/e2e-tests/conf/secrets.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/e2e-tests/conf/client.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.NbORX47zSH secret/some-users created deployment.apps/psmdb-client created + cat /tmp/tmp.6VQrsjhWBo + rm /tmp/tmp.NbORX47zSH /tmp/tmp.6VQrsjhWBo + return 0 + desc 'create PSMDB cluster' + set +o xtrace ----------------------------------------------------------------------------------- create PSMDB cluster ----------------------------------------------------------------------------------- + cluster=some-name + '[' 0 == 1 ']' + spinup_psmdb some-name-rs0 /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/e2e-tests/pvc-resize/conf/some-name.yml + local cluster=some-name-rs0 + local config=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/e2e-tests/pvc-resize/conf/some-name.yml + local size=3 + desc 'create first PSMDB cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PSMDB cluster ----------------------------------------------------------------------------------- + apply_cluster /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/e2e-tests/pvc-resize/conf/some-name.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/e2e-tests/pvc-resize/conf/some-name.yml + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/e2e-tests/pvc-resize/conf/some-name.yml ++ mktemp + yq eval '(.spec | select(.image == null)).image = "perconalab/percona-server-mongodb-operator:main-mongod7.0"' + yq eval '(.spec | select(has("pmm"))).pmm.image = "perconalab/pmm-client:dev-latest"' + yq eval '(.spec | select(has("initImage"))).initImage = "perconalab/percona-server-mongodb-operator:PR-1926-d4d89df0"' + yq eval '(.spec | select(has("backup"))).backup.image = "perconalab/percona-server-mongodb-operator:main-backup"' + local LAST_OUT=/tmp/tmp.g4VnZtyctC + yq eval '.spec.upgradeOptions.apply="Never"' ++ mktemp + local LAST_ERR=/tmp/tmp.x6WapTBGpk + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.g4VnZtyctC perconaservermongodb.psmdb.percona.com/some-name created + cat /tmp/tmp.x6WapTBGpk + rm /tmp/tmp.g4VnZtyctC /tmp/tmp.x6WapTBGpk + return 0 + desc 'check if Pod is started' + set +o xtrace ----------------------------------------------------------------------------------- check if Pod is started ----------------------------------------------------------------------------------- + wait_for_running some-name-rs0 3 + local name=some-name-rs0 + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=rs0 + local cluster_name=some-name ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + [[ 0 -eq 2 ]] + wait_pod some-name-rs0-0 + local pod=some-name-rs0-0 + set +o xtrace waiting for pod/some-name-rs0-0 to be ready.............OK + for i in '$(seq 0 $last_pod)' + [[ 1 -eq 2 ]] + wait_pod some-name-rs0-1 + local pod=some-name-rs0-1 + set +o xtrace waiting for pod/some-name-rs0-1 to be ready........................OK + for i in '$(seq 0 $last_pod)' + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WndveiKqqr +++ mktemp ++ local LAST_ERR=/tmp/tmp.6hhef5MwPu ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.WndveiKqqr ++ cat /tmp/tmp.6hhef5MwPu ++ rm /tmp/tmp.WndveiKqqr /tmp/tmp.6hhef5MwPu ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-rs0-2 + local pod=some-name-rs0-2 + set +o xtrace waiting for pod/some-name-rs0-2 to be ready...........OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].non_voting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HLxGhVoo5F +++ mktemp ++ local LAST_ERR=/tmp/tmp.WjtLb7AorR ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].non_voting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.HLxGhVoo5F ++ cat /tmp/tmp.WjtLb7AorR ++ rm /tmp/tmp.HLxGhVoo5F /tmp/tmp.WjtLb7AorR ++ return 0 + [[ '' == \t\r\u\e ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.amVXS2iAtn +++ mktemp ++ local LAST_ERR=/tmp/tmp.MvFBF822eA ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.amVXS2iAtn ++ cat /tmp/tmp.MvFBF822eA ++ rm /tmp/tmp.amVXS2iAtn /tmp/tmp.MvFBF822eA ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness + sleep 20 + compare_kubectl statefulset/some-name-rs0 + local resource=statefulset/some-name-rs0 + local postfix= + local skip_generation_check= + local expected_result=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/e2e-tests/pvc-resize/compare/statefulset_some-name-rs0.yml + local new_result=/tmp/tmp.kcQwVgU6Ei/statefulset_some-name-rs0.yml + '[' -n '' -a -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/e2e-tests/pvc-resize/compare/statefulset_some-name-rs0-oc.yml ']' + kubectl_bin get -o yaml statefulset/some-name-rs0 ++ mktemp + yq eval ' del(.metadata.ownerReferences[].apiVersion) | del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "NAMESPACE")) | del(.metadata.selfLink) | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.status) | (.. | select(tag == "!!str")) |= sub("pvc-resize-5656", "NAME_SPACE") | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "extensions/v1beta1")) = "apps/v1" | (.. | select(. == "batch/v1beta1")) = "batch/v1" ' - + local LAST_OUT=/tmp/tmp.VmAQTcfIZp ++ mktemp + local LAST_ERR=/tmp/tmp.kMHZM3gIwQ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-rs0 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.VmAQTcfIZp + cat /tmp/tmp.kMHZM3gIwQ + rm /tmp/tmp.VmAQTcfIZp /tmp/tmp.kMHZM3gIwQ + return 0 + yq -i eval 'del(.spec.persistentVolumeClaimRetentionPolicy)' /tmp/tmp.kcQwVgU6Ei/statefulset_some-name-rs0.yml + version_gt 1.22 ++ bc -l ++ echo '1.30 >= 1.22' + '[' 1 -eq 1 ']' + return 0 + yq -i eval 'del(.spec.internalTrafficPolicy)' /tmp/tmp.kcQwVgU6Ei/statefulset_some-name-rs0.yml + yq -i eval 'del(.spec.allocateLoadBalancerNodePorts)' /tmp/tmp.kcQwVgU6Ei/statefulset_some-name-rs0.yml + [[ /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/e2e-tests/pvc-resize/compare/statefulset_some-name-rs0.yml == */cronjob* ]] + '[' -n '' ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/e2e-tests/pvc-resize/compare/statefulset_some-name-rs0.yml /tmp/tmp.kcQwVgU6Ei/statefulset_some-name-rs0.yml + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + run_mongo 'db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})' userAdmin:userAdmin123456@some-name-rs0.pvc-resize-5656 + local 'command=db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})' + local uri=userAdmin:userAdmin123456@some-name-rs0.pvc-resize-5656 + local driver=mongodb+srv + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dZnzP1Nyvs +++ mktemp ++ local LAST_ERR=/tmp/tmp.uJGzZdEOqo ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.dZnzP1Nyvs ++ cat /tmp/tmp.uJGzZdEOqo ++ rm /tmp/tmp.dZnzP1Nyvs /tmp/tmp.uJGzZdEOqo ++ return 0 + local client_container=psmdb-client-66f577db5f-rrtt6 + local mongo_flag= + [[ userAdmin:userAdmin123456@some-name-rs0.pvc-resize-5656 == *cfg* ]] + replica_set=rs0 + kubectl_bin exec psmdb-client-66f577db5f-rrtt6 -- bash -c 'printf '\''db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})\n'\'' | mongo mongodb+srv://userAdmin:userAdmin123456@some-name-rs0.pvc-resize-5656.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.pwTGlyRdRD ++ mktemp + local LAST_ERR=/tmp/tmp.1xj4wqCS9f + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-66f577db5f-rrtt6 -- bash -c 'printf '\''db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})\n'\'' | mongo mongodb+srv://userAdmin:userAdmin123456@some-name-rs0.pvc-resize-5656.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.pwTGlyRdRD Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-rs0-0.some-name-rs0.pvc-resize-5656.svc.cluster.local:27017,some-name-rs0-1.some-name-rs0.pvc-resize-5656.svc.cluster.local:27017,some-name-rs0-2.some-name-rs0.pvc-resize-5656.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb&replicaSet=rs0&ssl=false Implicit session: session { "id" : UUID("36af7208-c0a7-4201-be8e-13565b6302b0") } Percona Server for MongoDB server version: v7.0.18-11 WARNING: shell and server versions do not match Successfully added user: { "user" : "myApp", "roles" : [ { "db" : "myApp", "role" : "readWrite" } ] } bye + cat /tmp/tmp.1xj4wqCS9f + rm /tmp/tmp.pwTGlyRdRD /tmp/tmp.1xj4wqCS9f + return 0 + run_mongo 'use myApp\n db.test.insert({ x: 100500 })' myApp:myPass@some-name-rs0.pvc-resize-5656 + local 'command=use myApp\n db.test.insert({ x: 100500 })' + local uri=myApp:myPass@some-name-rs0.pvc-resize-5656 + local driver=mongodb+srv + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.aYkO4lfLSH +++ mktemp ++ local LAST_ERR=/tmp/tmp.5ZAwSUTXHy ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.aYkO4lfLSH ++ cat /tmp/tmp.5ZAwSUTXHy ++ rm /tmp/tmp.aYkO4lfLSH /tmp/tmp.5ZAwSUTXHy ++ return 0 + local client_container=psmdb-client-66f577db5f-rrtt6 + local mongo_flag= + [[ myApp:myPass@some-name-rs0.pvc-resize-5656 == *cfg* ]] + replica_set=rs0 + kubectl_bin exec psmdb-client-66f577db5f-rrtt6 -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb+srv://myApp:myPass@some-name-rs0.pvc-resize-5656.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.M06pTIwFmP ++ mktemp + local LAST_ERR=/tmp/tmp.BBl6XtRSi0 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-66f577db5f-rrtt6 -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb+srv://myApp:myPass@some-name-rs0.pvc-resize-5656.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.M06pTIwFmP Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-rs0-0.some-name-rs0.pvc-resize-5656.svc.cluster.local:27017,some-name-rs0-1.some-name-rs0.pvc-resize-5656.svc.cluster.local:27017,some-name-rs0-2.some-name-rs0.pvc-resize-5656.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb&replicaSet=rs0&ssl=false Implicit session: session { "id" : UUID("7a701e46-2a69-42bc-8c34-3ceef1b40801") } Percona Server for MongoDB server version: v7.0.18-11 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.BBl6XtRSi0 + rm /tmp/tmp.M06pTIwFmP /tmp/tmp.BBl6XtRSi0 + return 0 + patch_pvc_request some-name 2G + local cluster=some-name + local size=2G + echo 'Patching PVC request to 2G in some-name' Patching PVC request to 2G in some-name + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"2G"}]' ++ mktemp + local LAST_OUT=/tmp/tmp.SN8JKvHix7 ++ mktemp + local LAST_ERR=/tmp/tmp.tn3iQU3lej + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"2G"}]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.SN8JKvHix7 perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.tn3iQU3lej + rm /tmp/tmp.SN8JKvHix7 /tmp/tmp.tn3iQU3lej + return 0 + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Tw7FHdU4mq +++ mktemp ++ local LAST_ERR=/tmp/tmp.RDBVOKbstg ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Tw7FHdU4mq ++ cat /tmp/tmp.RDBVOKbstg ++ rm /tmp/tmp.Tw7FHdU4mq /tmp/tmp.RDBVOKbstg ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo + echo + wait_all_pvc_resize 2Gi 120 1 + local expected_size=2Gi + local max_retry=120 + local sleep_time=1 ++ kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name +++ mktemp ++ local LAST_OUT=/tmp/tmp.yMTvAf5pI7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ad7FrZ37J1 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pvc -l app.kubernetes.io/component=mongod -o name ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.yMTvAf5pI7 ++ cat /tmp/tmp.Ad7FrZ37J1 ++ rm /tmp/tmp.yMTvAf5pI7 /tmp/tmp.Ad7FrZ37J1 ++ return 0 + for pvc in '$(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name)' + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-0 2Gi 120 1 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-0 + local expected_size=2Gi + local max_retry=120 + local sleep_time=1 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.elpwnv0IYi +++ mktemp ++ local LAST_ERR=/tmp/tmp.v4VTuxX6Pl ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.elpwnv0IYi ++ cat /tmp/tmp.v4VTuxX6Pl ++ rm /tmp/tmp.elpwnv0IYi /tmp/tmp.v4VTuxX6Pl ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 0 -ge 120 ]] + echo -n . .+ sleep 1 + retry=1 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.95ja0xnMHa +++ mktemp ++ local LAST_ERR=/tmp/tmp.AnJiRBFkPX ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.95ja0xnMHa ++ cat /tmp/tmp.AnJiRBFkPX ++ rm /tmp/tmp.95ja0xnMHa /tmp/tmp.AnJiRBFkPX ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 1 -ge 120 ]] + echo -n . .+ sleep 1 + retry=2 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yEeyIl55y0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.SAH3fpeMQ0 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.yEeyIl55y0 ++ cat /tmp/tmp.SAH3fpeMQ0 ++ rm /tmp/tmp.yEeyIl55y0 /tmp/tmp.SAH3fpeMQ0 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 2 -ge 120 ]] + echo -n . .+ sleep 1 + retry=3 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mjeinr1bZQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.SLLsFJNibP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.mjeinr1bZQ ++ cat /tmp/tmp.SLLsFJNibP ++ rm /tmp/tmp.mjeinr1bZQ /tmp/tmp.SLLsFJNibP ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 3 -ge 120 ]] + echo -n . .+ sleep 1 + retry=4 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AR2aLSMJLz +++ mktemp ++ local LAST_ERR=/tmp/tmp.3PecXHYKRn ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.AR2aLSMJLz ++ cat /tmp/tmp.3PecXHYKRn ++ rm /tmp/tmp.AR2aLSMJLz /tmp/tmp.3PecXHYKRn ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 4 -ge 120 ]] + echo -n . .+ sleep 1 + retry=5 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BNk4sVT3rV +++ mktemp ++ local LAST_ERR=/tmp/tmp.AvAg95j9RP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.BNk4sVT3rV ++ cat /tmp/tmp.AvAg95j9RP ++ rm /tmp/tmp.BNk4sVT3rV /tmp/tmp.AvAg95j9RP ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 5 -ge 120 ]] + echo -n . .+ sleep 1 + retry=6 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.k8sQkYxN77 +++ mktemp ++ local LAST_ERR=/tmp/tmp.cocOOm18TA ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.k8sQkYxN77 ++ cat /tmp/tmp.cocOOm18TA ++ rm /tmp/tmp.k8sQkYxN77 /tmp/tmp.cocOOm18TA ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 6 -ge 120 ]] + echo -n . .+ sleep 1 + retry=7 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sfhcNahg3l +++ mktemp ++ local LAST_ERR=/tmp/tmp.rpBCtg5vqr ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.sfhcNahg3l ++ cat /tmp/tmp.rpBCtg5vqr ++ rm /tmp/tmp.sfhcNahg3l /tmp/tmp.rpBCtg5vqr ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 7 -ge 120 ]] + echo -n . .+ sleep 1 + retry=8 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dpprFR6pCd +++ mktemp ++ local LAST_ERR=/tmp/tmp.3opwIJ1i7t ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.dpprFR6pCd ++ cat /tmp/tmp.3opwIJ1i7t ++ rm /tmp/tmp.dpprFR6pCd /tmp/tmp.3opwIJ1i7t ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 8 -ge 120 ]] + echo -n . .+ sleep 1 + retry=9 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BLFip581rc +++ mktemp ++ local LAST_ERR=/tmp/tmp.7Zi5UcupBp ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.BLFip581rc ++ cat /tmp/tmp.7Zi5UcupBp ++ rm /tmp/tmp.BLFip581rc /tmp/tmp.7Zi5UcupBp ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 9 -ge 120 ]] + echo -n . .+ sleep 1 + retry=10 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oLgQHhbAgo +++ mktemp ++ local LAST_ERR=/tmp/tmp.KSXeF12OCt ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.oLgQHhbAgo ++ cat /tmp/tmp.KSXeF12OCt ++ rm /tmp/tmp.oLgQHhbAgo /tmp/tmp.KSXeF12OCt ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 10 -ge 120 ]] + echo -n . .+ sleep 1 + retry=11 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Iayqhjbusm +++ mktemp ++ local LAST_ERR=/tmp/tmp.gtGT8ac76Q ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Iayqhjbusm ++ cat /tmp/tmp.gtGT8ac76Q ++ rm /tmp/tmp.Iayqhjbusm /tmp/tmp.gtGT8ac76Q ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 11 -ge 120 ]] + echo -n . .+ sleep 1 + retry=12 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dpeCqzMi6Q +++ mktemp ++ local LAST_ERR=/tmp/tmp.KFGCT0kYgR ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.dpeCqzMi6Q ++ cat /tmp/tmp.KFGCT0kYgR ++ rm /tmp/tmp.dpeCqzMi6Q /tmp/tmp.KFGCT0kYgR ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 12 -ge 120 ]] + echo -n . .+ sleep 1 + retry=13 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OLKjNRyQJE +++ mktemp ++ local LAST_ERR=/tmp/tmp.6by2MxDm3X ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.OLKjNRyQJE ++ cat /tmp/tmp.6by2MxDm3X ++ rm /tmp/tmp.OLKjNRyQJE /tmp/tmp.6by2MxDm3X ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 13 -ge 120 ]] + echo -n . .+ sleep 1 + retry=14 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nQUZEVM8aE +++ mktemp ++ local LAST_ERR=/tmp/tmp.sVnc3q4uWN ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.nQUZEVM8aE ++ cat /tmp/tmp.sVnc3q4uWN ++ rm /tmp/tmp.nQUZEVM8aE /tmp/tmp.sVnc3q4uWN ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 14 -ge 120 ]] + echo -n . .+ sleep 1 + retry=15 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zCRyEir7q6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.oFFzXL4qH7 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.zCRyEir7q6 ++ cat /tmp/tmp.oFFzXL4qH7 ++ rm /tmp/tmp.zCRyEir7q6 /tmp/tmp.oFFzXL4qH7 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 15 -ge 120 ]] + echo -n . .+ sleep 1 + retry=16 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dEiIPGwzIb +++ mktemp ++ local LAST_ERR=/tmp/tmp.5PoHoAxItl ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.dEiIPGwzIb ++ cat /tmp/tmp.5PoHoAxItl ++ rm /tmp/tmp.dEiIPGwzIb /tmp/tmp.5PoHoAxItl ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 16 -ge 120 ]] + echo -n . .+ sleep 1 + retry=17 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.N3Re5dSYAv +++ mktemp ++ local LAST_ERR=/tmp/tmp.D2sBRF0RFy ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.N3Re5dSYAv ++ cat /tmp/tmp.D2sBRF0RFy ++ rm /tmp/tmp.N3Re5dSYAv /tmp/tmp.D2sBRF0RFy ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 17 -ge 120 ]] + echo -n . .+ sleep 1 + retry=18 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fG4n2PSArW +++ mktemp ++ local LAST_ERR=/tmp/tmp.hKhrrXp6cX ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.fG4n2PSArW ++ cat /tmp/tmp.hKhrrXp6cX ++ rm /tmp/tmp.fG4n2PSArW /tmp/tmp.hKhrrXp6cX ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 18 -ge 120 ]] + echo -n . .+ sleep 1 + retry=19 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gXmwngzwSU +++ mktemp ++ local LAST_ERR=/tmp/tmp.fEg9CxFSnM ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.gXmwngzwSU ++ cat /tmp/tmp.fEg9CxFSnM ++ rm /tmp/tmp.gXmwngzwSU /tmp/tmp.fEg9CxFSnM ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 19 -ge 120 ]] + echo -n . .+ sleep 1 + retry=20 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WxjsPnykqT +++ mktemp ++ local LAST_ERR=/tmp/tmp.urN6GzlCkt ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.WxjsPnykqT ++ cat /tmp/tmp.urN6GzlCkt ++ rm /tmp/tmp.WxjsPnykqT /tmp/tmp.urN6GzlCkt ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 20 -ge 120 ]] + echo -n . .+ sleep 1 + retry=21 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sKx0WuUJRm +++ mktemp ++ local LAST_ERR=/tmp/tmp.GTSkbFMbVw ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.sKx0WuUJRm ++ cat /tmp/tmp.GTSkbFMbVw ++ rm /tmp/tmp.sKx0WuUJRm /tmp/tmp.GTSkbFMbVw ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 21 -ge 120 ]] + echo -n . .+ sleep 1 + retry=22 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dWWMyUQxJP +++ mktemp ++ local LAST_ERR=/tmp/tmp.ty36NgcGt5 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.dWWMyUQxJP ++ cat /tmp/tmp.ty36NgcGt5 ++ rm /tmp/tmp.dWWMyUQxJP /tmp/tmp.ty36NgcGt5 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 22 -ge 120 ]] + echo -n . .+ sleep 1 + retry=23 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.reWfr777PZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.7yKCmfrZOb ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.reWfr777PZ ++ cat /tmp/tmp.7yKCmfrZOb ++ rm /tmp/tmp.reWfr777PZ /tmp/tmp.7yKCmfrZOb ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 23 -ge 120 ]] + echo -n . .+ sleep 1 + retry=24 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lbaer6tAgs +++ mktemp ++ local LAST_ERR=/tmp/tmp.OFc7UnQASk ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.lbaer6tAgs ++ cat /tmp/tmp.OFc7UnQASk ++ rm /tmp/tmp.lbaer6tAgs /tmp/tmp.OFc7UnQASk ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 24 -ge 120 ]] + echo -n . .+ sleep 1 + retry=25 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nPGO0Zxu01 +++ mktemp ++ local LAST_ERR=/tmp/tmp.9F1z761cI4 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.nPGO0Zxu01 ++ cat /tmp/tmp.9F1z761cI4 ++ rm /tmp/tmp.nPGO0Zxu01 /tmp/tmp.9F1z761cI4 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 25 -ge 120 ]] + echo -n . .+ sleep 1 + retry=26 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YDt80cxqGi +++ mktemp ++ local LAST_ERR=/tmp/tmp.hlLVREhfQ4 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.YDt80cxqGi ++ cat /tmp/tmp.hlLVREhfQ4 ++ rm /tmp/tmp.YDt80cxqGi /tmp/tmp.hlLVREhfQ4 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 26 -ge 120 ]] + echo -n . .+ sleep 1 + retry=27 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.V7o9MZaBSt +++ mktemp ++ local LAST_ERR=/tmp/tmp.kz4aOXhq0k ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.V7o9MZaBSt ++ cat /tmp/tmp.kz4aOXhq0k ++ rm /tmp/tmp.V7o9MZaBSt /tmp/tmp.kz4aOXhq0k ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 27 -ge 120 ]] + echo -n . .+ sleep 1 + retry=28 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9ON4gE0jOn +++ mktemp ++ local LAST_ERR=/tmp/tmp.71wAqmxwMA ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9ON4gE0jOn ++ cat /tmp/tmp.71wAqmxwMA ++ rm /tmp/tmp.9ON4gE0jOn /tmp/tmp.71wAqmxwMA ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 28 -ge 120 ]] + echo -n . .+ sleep 1 + retry=29 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RTh1Um9hfJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.Bjo9MN0mrD ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.RTh1Um9hfJ ++ cat /tmp/tmp.Bjo9MN0mrD ++ rm /tmp/tmp.RTh1Um9hfJ /tmp/tmp.Bjo9MN0mrD ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 29 -ge 120 ]] + echo -n . .+ sleep 1 + retry=30 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QQR5A3p8vX +++ mktemp ++ local LAST_ERR=/tmp/tmp.QE3gZrHuc6 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.QQR5A3p8vX ++ cat /tmp/tmp.QE3gZrHuc6 ++ rm /tmp/tmp.QQR5A3p8vX /tmp/tmp.QE3gZrHuc6 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 30 -ge 120 ]] + echo -n . .+ sleep 1 + retry=31 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nKkgArK7u3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.q7WFmUYaOA ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.nKkgArK7u3 ++ cat /tmp/tmp.q7WFmUYaOA ++ rm /tmp/tmp.nKkgArK7u3 /tmp/tmp.q7WFmUYaOA ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 31 -ge 120 ]] + echo -n . .+ sleep 1 + retry=32 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ko5BlHfo1v +++ mktemp ++ local LAST_ERR=/tmp/tmp.wQgrhsfP37 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Ko5BlHfo1v ++ cat /tmp/tmp.wQgrhsfP37 ++ rm /tmp/tmp.Ko5BlHfo1v /tmp/tmp.wQgrhsfP37 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 32 -ge 120 ]] + echo -n . .+ sleep 1 + retry=33 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KnklrqgKJG +++ mktemp ++ local LAST_ERR=/tmp/tmp.VT9mqoHH0w ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.KnklrqgKJG ++ cat /tmp/tmp.VT9mqoHH0w ++ rm /tmp/tmp.KnklrqgKJG /tmp/tmp.VT9mqoHH0w ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 33 -ge 120 ]] + echo -n . .+ sleep 1 + retry=34 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Nx0mVP12vn +++ mktemp ++ local LAST_ERR=/tmp/tmp.nLHB9uV9SV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Nx0mVP12vn ++ cat /tmp/tmp.nLHB9uV9SV ++ rm /tmp/tmp.Nx0mVP12vn /tmp/tmp.nLHB9uV9SV ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 34 -ge 120 ]] + echo -n . .+ sleep 1 + retry=35 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4dNDMwZ0gm +++ mktemp ++ local LAST_ERR=/tmp/tmp.fW7VjqovoT ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4dNDMwZ0gm ++ cat /tmp/tmp.fW7VjqovoT ++ rm /tmp/tmp.4dNDMwZ0gm /tmp/tmp.fW7VjqovoT ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 35 -ge 120 ]] + echo -n . .+ sleep 1 + retry=36 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gOYIKGcpBo +++ mktemp ++ local LAST_ERR=/tmp/tmp.Z9tlB49sMd ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.gOYIKGcpBo ++ cat /tmp/tmp.Z9tlB49sMd ++ rm /tmp/tmp.gOYIKGcpBo /tmp/tmp.Z9tlB49sMd ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 36 -ge 120 ]] + echo -n . .+ sleep 1 + retry=37 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LfySDFjlRM +++ mktemp ++ local LAST_ERR=/tmp/tmp.r1KiAXrHL4 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.LfySDFjlRM ++ cat /tmp/tmp.r1KiAXrHL4 ++ rm /tmp/tmp.LfySDFjlRM /tmp/tmp.r1KiAXrHL4 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 37 -ge 120 ]] + echo -n . .+ sleep 1 + retry=38 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xP6xzgRsrx +++ mktemp ++ local LAST_ERR=/tmp/tmp.CYmquSOgWI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xP6xzgRsrx ++ cat /tmp/tmp.CYmquSOgWI ++ rm /tmp/tmp.xP6xzgRsrx /tmp/tmp.CYmquSOgWI ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 38 -ge 120 ]] + echo -n . .+ sleep 1 + retry=39 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.V3ekt2SQjp +++ mktemp ++ local LAST_ERR=/tmp/tmp.KvFndDdHVx ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.V3ekt2SQjp ++ cat /tmp/tmp.KvFndDdHVx ++ rm /tmp/tmp.V3ekt2SQjp /tmp/tmp.KvFndDdHVx ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 39 -ge 120 ]] + echo -n . .+ sleep 1 + retry=40 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5F0vWrzxXA +++ mktemp ++ local LAST_ERR=/tmp/tmp.DaLt45UBrU ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5F0vWrzxXA ++ cat /tmp/tmp.DaLt45UBrU ++ rm /tmp/tmp.5F0vWrzxXA /tmp/tmp.DaLt45UBrU ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 40 -ge 120 ]] + echo -n . .+ sleep 1 + retry=41 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ucg77fOlXC +++ mktemp ++ local LAST_ERR=/tmp/tmp.SubiEHefe9 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Ucg77fOlXC ++ cat /tmp/tmp.SubiEHefe9 ++ rm /tmp/tmp.Ucg77fOlXC /tmp/tmp.SubiEHefe9 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 41 -ge 120 ]] + echo -n . .+ sleep 1 + retry=42 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XWPjfzHnOW +++ mktemp ++ local LAST_ERR=/tmp/tmp.mIQkw7Yhvz ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.XWPjfzHnOW ++ cat /tmp/tmp.mIQkw7Yhvz ++ rm /tmp/tmp.XWPjfzHnOW /tmp/tmp.mIQkw7Yhvz ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 42 -ge 120 ]] + echo -n . .+ sleep 1 + retry=43 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LayJx2rdzE +++ mktemp ++ local LAST_ERR=/tmp/tmp.h3PabwpSQg ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.LayJx2rdzE ++ cat /tmp/tmp.h3PabwpSQg ++ rm /tmp/tmp.LayJx2rdzE /tmp/tmp.h3PabwpSQg ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 43 -ge 120 ]] + echo -n . .+ sleep 1 + retry=44 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.aQpT8RhhZ8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Lxl7grGNAJ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.aQpT8RhhZ8 ++ cat /tmp/tmp.Lxl7grGNAJ ++ rm /tmp/tmp.aQpT8RhhZ8 /tmp/tmp.Lxl7grGNAJ ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 44 -ge 120 ]] + echo -n . .+ sleep 1 + retry=45 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.m8w3SMTPVW +++ mktemp ++ local LAST_ERR=/tmp/tmp.p0Sl4l2FMw ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.m8w3SMTPVW ++ cat /tmp/tmp.p0Sl4l2FMw ++ rm /tmp/tmp.m8w3SMTPVW /tmp/tmp.p0Sl4l2FMw ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 45 -ge 120 ]] + echo -n . .+ sleep 1 + retry=46 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.j7AWkn0Pgk +++ mktemp ++ local LAST_ERR=/tmp/tmp.y5UZHIcR6R ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.j7AWkn0Pgk ++ cat /tmp/tmp.y5UZHIcR6R ++ rm /tmp/tmp.j7AWkn0Pgk /tmp/tmp.y5UZHIcR6R ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 46 -ge 120 ]] + echo -n . .+ sleep 1 + retry=47 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ED5D4sc2Om +++ mktemp ++ local LAST_ERR=/tmp/tmp.neg1C4dF9v ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ED5D4sc2Om ++ cat /tmp/tmp.neg1C4dF9v ++ rm /tmp/tmp.ED5D4sc2Om /tmp/tmp.neg1C4dF9v ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 47 -ge 120 ]] + echo -n . .+ sleep 1 + retry=48 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EbkpNTdBUc +++ mktemp ++ local LAST_ERR=/tmp/tmp.jLuQtww7yR ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.EbkpNTdBUc ++ cat /tmp/tmp.jLuQtww7yR ++ rm /tmp/tmp.EbkpNTdBUc /tmp/tmp.jLuQtww7yR ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 48 -ge 120 ]] + echo -n . .+ sleep 1 + retry=49 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PblbvMk9fR +++ mktemp ++ local LAST_ERR=/tmp/tmp.kcF7JbvY7q ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.PblbvMk9fR ++ cat /tmp/tmp.kcF7JbvY7q ++ rm /tmp/tmp.PblbvMk9fR /tmp/tmp.kcF7JbvY7q ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 49 -ge 120 ]] + echo -n . .+ sleep 1 + retry=50 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ml7ooc79RM +++ mktemp ++ local LAST_ERR=/tmp/tmp.6BorkNo7Iq ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ml7ooc79RM ++ cat /tmp/tmp.6BorkNo7Iq ++ rm /tmp/tmp.ml7ooc79RM /tmp/tmp.6BorkNo7Iq ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 50 -ge 120 ]] + echo -n . .+ sleep 1 + retry=51 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.j1bqJQFLiz +++ mktemp ++ local LAST_ERR=/tmp/tmp.WELszFNJR8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.j1bqJQFLiz ++ cat /tmp/tmp.WELszFNJR8 ++ rm /tmp/tmp.j1bqJQFLiz /tmp/tmp.WELszFNJR8 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 51 -ge 120 ]] + echo -n . .+ sleep 1 + retry=52 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OzISFlg4Iz +++ mktemp ++ local LAST_ERR=/tmp/tmp.zmmfyq8xxc ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.OzISFlg4Iz ++ cat /tmp/tmp.zmmfyq8xxc ++ rm /tmp/tmp.OzISFlg4Iz /tmp/tmp.zmmfyq8xxc ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 52 -ge 120 ]] + echo -n . .+ sleep 1 + retry=53 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Oy1GPZPJ6S +++ mktemp ++ local LAST_ERR=/tmp/tmp.TRAnbSIBVj ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Oy1GPZPJ6S ++ cat /tmp/tmp.TRAnbSIBVj ++ rm /tmp/tmp.Oy1GPZPJ6S /tmp/tmp.TRAnbSIBVj ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 53 -ge 120 ]] + echo -n . .+ sleep 1 + retry=54 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Qo5XIrvh3g +++ mktemp ++ local LAST_ERR=/tmp/tmp.IgZJelut1A ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Qo5XIrvh3g ++ cat /tmp/tmp.IgZJelut1A ++ rm /tmp/tmp.Qo5XIrvh3g /tmp/tmp.IgZJelut1A ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 54 -ge 120 ]] + echo -n . .+ sleep 1 + retry=55 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zQ16MyQGu0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.gh5u94rber ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.zQ16MyQGu0 ++ cat /tmp/tmp.gh5u94rber ++ rm /tmp/tmp.zQ16MyQGu0 /tmp/tmp.gh5u94rber ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 55 -ge 120 ]] + echo -n . .+ sleep 1 + retry=56 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Z8sXyjUyRt +++ mktemp ++ local LAST_ERR=/tmp/tmp.ykTl2I6mS9 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Z8sXyjUyRt ++ cat /tmp/tmp.ykTl2I6mS9 ++ rm /tmp/tmp.Z8sXyjUyRt /tmp/tmp.ykTl2I6mS9 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 56 -ge 120 ]] + echo -n . .+ sleep 1 + retry=57 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MJ0qSglQ8G +++ mktemp ++ local LAST_ERR=/tmp/tmp.7sudM8YV6B ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.MJ0qSglQ8G ++ cat /tmp/tmp.7sudM8YV6B ++ rm /tmp/tmp.MJ0qSglQ8G /tmp/tmp.7sudM8YV6B ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 57 -ge 120 ]] + echo -n . .+ sleep 1 + retry=58 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JRoesuvdbd +++ mktemp ++ local LAST_ERR=/tmp/tmp.4j30d18n74 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.JRoesuvdbd ++ cat /tmp/tmp.4j30d18n74 ++ rm /tmp/tmp.JRoesuvdbd /tmp/tmp.4j30d18n74 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 58 -ge 120 ]] + echo -n . .+ sleep 1 + retry=59 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.z2nOZxfIhz +++ mktemp ++ local LAST_ERR=/tmp/tmp.0iGV3JOD3I ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.z2nOZxfIhz ++ cat /tmp/tmp.0iGV3JOD3I ++ rm /tmp/tmp.z2nOZxfIhz /tmp/tmp.0iGV3JOD3I ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 59 -ge 120 ]] + echo -n . .+ sleep 1 + retry=60 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GlyRGwPJkj +++ mktemp ++ local LAST_ERR=/tmp/tmp.iScCMuWOl9 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.GlyRGwPJkj ++ cat /tmp/tmp.iScCMuWOl9 ++ rm /tmp/tmp.GlyRGwPJkj /tmp/tmp.iScCMuWOl9 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 60 -ge 120 ]] + echo -n . .+ sleep 1 + retry=61 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Bfy7L5Ypym +++ mktemp ++ local LAST_ERR=/tmp/tmp.y5zPqK9pXr ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Bfy7L5Ypym ++ cat /tmp/tmp.y5zPqK9pXr ++ rm /tmp/tmp.Bfy7L5Ypym /tmp/tmp.y5zPqK9pXr ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 61 -ge 120 ]] + echo -n . .+ sleep 1 + retry=62 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6raQB39Aiy +++ mktemp ++ local LAST_ERR=/tmp/tmp.LbSc4MjqVa ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.6raQB39Aiy ++ cat /tmp/tmp.LbSc4MjqVa ++ rm /tmp/tmp.6raQB39Aiy /tmp/tmp.LbSc4MjqVa ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 62 -ge 120 ]] + echo -n . .+ sleep 1 + retry=63 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mDM316kU6q +++ mktemp ++ local LAST_ERR=/tmp/tmp.Z726QqLI16 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.mDM316kU6q ++ cat /tmp/tmp.Z726QqLI16 ++ rm /tmp/tmp.mDM316kU6q /tmp/tmp.Z726QqLI16 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 63 -ge 120 ]] + echo -n . .+ sleep 1 + retry=64 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WM4X1gWO5R +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZIcL98rgr5 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.WM4X1gWO5R ++ cat /tmp/tmp.ZIcL98rgr5 ++ rm /tmp/tmp.WM4X1gWO5R /tmp/tmp.ZIcL98rgr5 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 64 -ge 120 ]] + echo -n . .+ sleep 1 + retry=65 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HfeIBzMCTv +++ mktemp ++ local LAST_ERR=/tmp/tmp.ArjcY638ub ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.HfeIBzMCTv ++ cat /tmp/tmp.ArjcY638ub ++ rm /tmp/tmp.HfeIBzMCTv /tmp/tmp.ArjcY638ub ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 65 -ge 120 ]] + echo -n . .+ sleep 1 + retry=66 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6CRNbX9YHk +++ mktemp ++ local LAST_ERR=/tmp/tmp.a0C404xEAk ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.6CRNbX9YHk ++ cat /tmp/tmp.a0C404xEAk ++ rm /tmp/tmp.6CRNbX9YHk /tmp/tmp.a0C404xEAk ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 66 -ge 120 ]] + echo -n . .+ sleep 1 + retry=67 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.myiarr8ZZF +++ mktemp ++ local LAST_ERR=/tmp/tmp.tnnvhdZ29e ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.myiarr8ZZF ++ cat /tmp/tmp.tnnvhdZ29e ++ rm /tmp/tmp.myiarr8ZZF /tmp/tmp.tnnvhdZ29e ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 67 -ge 120 ]] + echo -n . .+ sleep 1 + retry=68 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rRF1hAZ6gf +++ mktemp ++ local LAST_ERR=/tmp/tmp.8gg5eGUVjy ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.rRF1hAZ6gf ++ cat /tmp/tmp.8gg5eGUVjy ++ rm /tmp/tmp.rRF1hAZ6gf /tmp/tmp.8gg5eGUVjy ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 68 -ge 120 ]] + echo -n . .+ sleep 1 + retry=69 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Oy7xVAwCMr +++ mktemp ++ local LAST_ERR=/tmp/tmp.6NBCyP18c8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Oy7xVAwCMr ++ cat /tmp/tmp.6NBCyP18c8 ++ rm /tmp/tmp.Oy7xVAwCMr /tmp/tmp.6NBCyP18c8 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 69 -ge 120 ]] + echo -n . .+ sleep 1 + retry=70 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sHAsARKZaU +++ mktemp ++ local LAST_ERR=/tmp/tmp.Zs9qXGqd4V ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.sHAsARKZaU ++ cat /tmp/tmp.Zs9qXGqd4V ++ rm /tmp/tmp.sHAsARKZaU /tmp/tmp.Zs9qXGqd4V ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 70 -ge 120 ]] + echo -n . .+ sleep 1 + retry=71 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.B0zGneMje1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.LWfeZmpq25 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.B0zGneMje1 ++ cat /tmp/tmp.LWfeZmpq25 ++ rm /tmp/tmp.B0zGneMje1 /tmp/tmp.LWfeZmpq25 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 71 -ge 120 ]] + echo -n . .+ sleep 1 + retry=72 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.CFyYIaoK0d +++ mktemp ++ local LAST_ERR=/tmp/tmp.coqyc98vwV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.CFyYIaoK0d ++ cat /tmp/tmp.coqyc98vwV ++ rm /tmp/tmp.CFyYIaoK0d /tmp/tmp.coqyc98vwV ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 72 -ge 120 ]] + echo -n . .+ sleep 1 + retry=73 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jZwwaj8cVh +++ mktemp ++ local LAST_ERR=/tmp/tmp.VoYS9Prmt4 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jZwwaj8cVh ++ cat /tmp/tmp.VoYS9Prmt4 ++ rm /tmp/tmp.jZwwaj8cVh /tmp/tmp.VoYS9Prmt4 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 73 -ge 120 ]] + echo -n . .+ sleep 1 + retry=74 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OkhitVb4G5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.qhGwUTfR8b ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.OkhitVb4G5 ++ cat /tmp/tmp.qhGwUTfR8b ++ rm /tmp/tmp.OkhitVb4G5 /tmp/tmp.qhGwUTfR8b ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 74 -ge 120 ]] + echo -n . .+ sleep 1 + retry=75 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Iw97JKe4WH +++ mktemp ++ local LAST_ERR=/tmp/tmp.mrFdmscWKM ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Iw97JKe4WH ++ cat /tmp/tmp.mrFdmscWKM ++ rm /tmp/tmp.Iw97JKe4WH /tmp/tmp.mrFdmscWKM ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 75 -ge 120 ]] + echo -n . .+ sleep 1 + retry=76 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ns0AOsEU6a +++ mktemp ++ local LAST_ERR=/tmp/tmp.FMzSRt5iDa ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ns0AOsEU6a ++ cat /tmp/tmp.FMzSRt5iDa ++ rm /tmp/tmp.ns0AOsEU6a /tmp/tmp.FMzSRt5iDa ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 76 -ge 120 ]] + echo -n . .+ sleep 1 + retry=77 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9yEvKbWycy +++ mktemp ++ local LAST_ERR=/tmp/tmp.8Jv45Vlgll ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9yEvKbWycy ++ cat /tmp/tmp.8Jv45Vlgll ++ rm /tmp/tmp.9yEvKbWycy /tmp/tmp.8Jv45Vlgll ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 77 -ge 120 ]] + echo -n . .+ sleep 1 + retry=78 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wYnVl6aCJy +++ mktemp ++ local LAST_ERR=/tmp/tmp.qEyAYTSTRX ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.wYnVl6aCJy ++ cat /tmp/tmp.qEyAYTSTRX ++ rm /tmp/tmp.wYnVl6aCJy /tmp/tmp.qEyAYTSTRX ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 78 -ge 120 ]] + echo -n . .+ sleep 1 + retry=79 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RFRz9TlQOT +++ mktemp ++ local LAST_ERR=/tmp/tmp.OshGJRPYAI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.RFRz9TlQOT ++ cat /tmp/tmp.OshGJRPYAI ++ rm /tmp/tmp.RFRz9TlQOT /tmp/tmp.OshGJRPYAI ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 79 -ge 120 ]] + echo -n . .+ sleep 1 + retry=80 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Qth5SnmJaS +++ mktemp ++ local LAST_ERR=/tmp/tmp.mPRPjbXyBg ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Qth5SnmJaS ++ cat /tmp/tmp.mPRPjbXyBg ++ rm /tmp/tmp.Qth5SnmJaS /tmp/tmp.mPRPjbXyBg ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 80 -ge 120 ]] + echo -n . .+ sleep 1 + retry=81 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SFTBgQpKuD +++ mktemp ++ local LAST_ERR=/tmp/tmp.TetwyXV6bK ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.SFTBgQpKuD ++ cat /tmp/tmp.TetwyXV6bK ++ rm /tmp/tmp.SFTBgQpKuD /tmp/tmp.TetwyXV6bK ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 81 -ge 120 ]] + echo -n . .+ sleep 1 + retry=82 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cDlNM6dc2s +++ mktemp ++ local LAST_ERR=/tmp/tmp.yoQVpZ7Ehc ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.cDlNM6dc2s ++ cat /tmp/tmp.yoQVpZ7Ehc ++ rm /tmp/tmp.cDlNM6dc2s /tmp/tmp.yoQVpZ7Ehc ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 82 -ge 120 ]] + echo -n . .+ sleep 1 + retry=83 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DJh8RnRa15 +++ mktemp ++ local LAST_ERR=/tmp/tmp.U6Yk7ffNN7 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.DJh8RnRa15 ++ cat /tmp/tmp.U6Yk7ffNN7 ++ rm /tmp/tmp.DJh8RnRa15 /tmp/tmp.U6Yk7ffNN7 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 83 -ge 120 ]] + echo -n . .+ sleep 1 + retry=84 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1E06jJI5kg +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ss0uSn3Ija ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.1E06jJI5kg ++ cat /tmp/tmp.Ss0uSn3Ija ++ rm /tmp/tmp.1E06jJI5kg /tmp/tmp.Ss0uSn3Ija ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 84 -ge 120 ]] + echo -n . .+ sleep 1 + retry=85 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uDlsQHt81f +++ mktemp ++ local LAST_ERR=/tmp/tmp.PRxv9wrTB2 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.uDlsQHt81f ++ cat /tmp/tmp.PRxv9wrTB2 ++ rm /tmp/tmp.uDlsQHt81f /tmp/tmp.PRxv9wrTB2 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 85 -ge 120 ]] + echo -n . .+ sleep 1 + retry=86 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DMQa1228ht +++ mktemp ++ local LAST_ERR=/tmp/tmp.5jxS7Hl1uZ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.DMQa1228ht ++ cat /tmp/tmp.5jxS7Hl1uZ ++ rm /tmp/tmp.DMQa1228ht /tmp/tmp.5jxS7Hl1uZ ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 86 -ge 120 ]] + echo -n . .+ sleep 1 + retry=87 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IGimK8vAMy +++ mktemp ++ local LAST_ERR=/tmp/tmp.Nxo7xI09Ft ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.IGimK8vAMy ++ cat /tmp/tmp.Nxo7xI09Ft ++ rm /tmp/tmp.IGimK8vAMy /tmp/tmp.Nxo7xI09Ft ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 87 -ge 120 ]] + echo -n . .+ sleep 1 + retry=88 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IkYm8fUEy5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Z4NwlS2JfF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.IkYm8fUEy5 ++ cat /tmp/tmp.Z4NwlS2JfF ++ rm /tmp/tmp.IkYm8fUEy5 /tmp/tmp.Z4NwlS2JfF ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 88 -ge 120 ]] + echo -n . .+ sleep 1 + retry=89 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Iw2VA0DSXm +++ mktemp ++ local LAST_ERR=/tmp/tmp.OjiSFij6Kr ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Iw2VA0DSXm ++ cat /tmp/tmp.OjiSFij6Kr ++ rm /tmp/tmp.Iw2VA0DSXm /tmp/tmp.OjiSFij6Kr ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 89 -ge 120 ]] + echo -n . .+ sleep 1 + retry=90 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rDboeodUAm +++ mktemp ++ local LAST_ERR=/tmp/tmp.U5dGu0Dm3N ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.rDboeodUAm ++ cat /tmp/tmp.U5dGu0Dm3N ++ rm /tmp/tmp.rDboeodUAm /tmp/tmp.U5dGu0Dm3N ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 90 -ge 120 ]] + echo -n . .+ sleep 1 + retry=91 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FlaW8nBCuD +++ mktemp ++ local LAST_ERR=/tmp/tmp.qKlRqRnUu0 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.FlaW8nBCuD ++ cat /tmp/tmp.qKlRqRnUu0 ++ rm /tmp/tmp.FlaW8nBCuD /tmp/tmp.qKlRqRnUu0 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 91 -ge 120 ]] + echo -n . .+ sleep 1 + retry=92 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7xhrD2IAbr +++ mktemp ++ local LAST_ERR=/tmp/tmp.ITqi6J8VQI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.7xhrD2IAbr ++ cat /tmp/tmp.ITqi6J8VQI ++ rm /tmp/tmp.7xhrD2IAbr /tmp/tmp.ITqi6J8VQI ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 92 -ge 120 ]] + echo -n . .+ sleep 1 + retry=93 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.REvTQYvOdu +++ mktemp ++ local LAST_ERR=/tmp/tmp.YOEKiZ985N ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.REvTQYvOdu ++ cat /tmp/tmp.YOEKiZ985N ++ rm /tmp/tmp.REvTQYvOdu /tmp/tmp.YOEKiZ985N ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 93 -ge 120 ]] + echo -n . .+ sleep 1 + retry=94 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.y8K1jvGYtB +++ mktemp ++ local LAST_ERR=/tmp/tmp.gwNuVxCuKQ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.y8K1jvGYtB ++ cat /tmp/tmp.gwNuVxCuKQ ++ rm /tmp/tmp.y8K1jvGYtB /tmp/tmp.gwNuVxCuKQ ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 94 -ge 120 ]] + echo -n . .+ sleep 1 + retry=95 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7sN4HGxWN8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.QN19b3uiPZ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.7sN4HGxWN8 ++ cat /tmp/tmp.QN19b3uiPZ ++ rm /tmp/tmp.7sN4HGxWN8 /tmp/tmp.QN19b3uiPZ ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 95 -ge 120 ]] + echo -n . .+ sleep 1 + retry=96 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4RbIepqXhn +++ mktemp ++ local LAST_ERR=/tmp/tmp.qts62KvpvK ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4RbIepqXhn ++ cat /tmp/tmp.qts62KvpvK ++ rm /tmp/tmp.4RbIepqXhn /tmp/tmp.qts62KvpvK ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 96 -ge 120 ]] + echo -n . .+ sleep 1 + retry=97 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wAE2lmtqBO +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ay9BLt1buI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.wAE2lmtqBO ++ cat /tmp/tmp.Ay9BLt1buI ++ rm /tmp/tmp.wAE2lmtqBO /tmp/tmp.Ay9BLt1buI ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 97 -ge 120 ]] + echo -n . .+ sleep 1 + retry=98 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.b5F4Q39RjS +++ mktemp ++ local LAST_ERR=/tmp/tmp.rF0P1k05IK ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.b5F4Q39RjS ++ cat /tmp/tmp.rF0P1k05IK ++ rm /tmp/tmp.b5F4Q39RjS /tmp/tmp.rF0P1k05IK ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 98 -ge 120 ]] + echo -n . .+ sleep 1 + retry=99 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QfcZMzVtHW +++ mktemp ++ local LAST_ERR=/tmp/tmp.yO2SMB1RFe ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.QfcZMzVtHW ++ cat /tmp/tmp.yO2SMB1RFe ++ rm /tmp/tmp.QfcZMzVtHW /tmp/tmp.yO2SMB1RFe ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 99 -ge 120 ]] + echo -n . .+ sleep 1 + retry=100 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.do9wD9dJwb +++ mktemp ++ local LAST_ERR=/tmp/tmp.ptjbUbim4H ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.do9wD9dJwb ++ cat /tmp/tmp.ptjbUbim4H ++ rm /tmp/tmp.do9wD9dJwb /tmp/tmp.ptjbUbim4H ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 100 -ge 120 ]] + echo -n . .+ sleep 1 + retry=101 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FAu9OuhwfH +++ mktemp ++ local LAST_ERR=/tmp/tmp.kewPadULDK ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.FAu9OuhwfH ++ cat /tmp/tmp.kewPadULDK ++ rm /tmp/tmp.FAu9OuhwfH /tmp/tmp.kewPadULDK ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 101 -ge 120 ]] + echo -n . .+ sleep 1 + retry=102 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DCcsJ2l4U7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.fHSFVRmgIh ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.DCcsJ2l4U7 ++ cat /tmp/tmp.fHSFVRmgIh ++ rm /tmp/tmp.DCcsJ2l4U7 /tmp/tmp.fHSFVRmgIh ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 102 -ge 120 ]] + echo -n . .+ sleep 1 + retry=103 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vNtIgEWbuK +++ mktemp ++ local LAST_ERR=/tmp/tmp.7Y19Ob5Efc ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.vNtIgEWbuK ++ cat /tmp/tmp.7Y19Ob5Efc ++ rm /tmp/tmp.vNtIgEWbuK /tmp/tmp.7Y19Ob5Efc ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 103 -ge 120 ]] + echo -n . .+ sleep 1 + retry=104 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cuaTD5d2UR +++ mktemp ++ local LAST_ERR=/tmp/tmp.af7Cwfh9ae ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.cuaTD5d2UR ++ cat /tmp/tmp.af7Cwfh9ae ++ rm /tmp/tmp.cuaTD5d2UR /tmp/tmp.af7Cwfh9ae ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 104 -ge 120 ]] + echo -n . .+ sleep 1 + retry=105 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rrvqFdNWM2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.W6ad5PZvwA ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.rrvqFdNWM2 ++ cat /tmp/tmp.W6ad5PZvwA ++ rm /tmp/tmp.rrvqFdNWM2 /tmp/tmp.W6ad5PZvwA ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 105 -ge 120 ]] + echo -n . .+ sleep 1 + retry=106 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.c2tjk5sbrs +++ mktemp ++ local LAST_ERR=/tmp/tmp.orJtrBHEQV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.c2tjk5sbrs ++ cat /tmp/tmp.orJtrBHEQV ++ rm /tmp/tmp.c2tjk5sbrs /tmp/tmp.orJtrBHEQV ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 106 -ge 120 ]] + echo -n . .+ sleep 1 + retry=107 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AZJpII4qvt +++ mktemp ++ local LAST_ERR=/tmp/tmp.dsGkU9ItAa ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.AZJpII4qvt ++ cat /tmp/tmp.dsGkU9ItAa ++ rm /tmp/tmp.AZJpII4qvt /tmp/tmp.dsGkU9ItAa ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 107 -ge 120 ]] + echo -n . .+ sleep 1 + retry=108 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.E3R8SIbjAF +++ mktemp ++ local LAST_ERR=/tmp/tmp.v55g8aoYGv ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.E3R8SIbjAF ++ cat /tmp/tmp.v55g8aoYGv ++ rm /tmp/tmp.E3R8SIbjAF /tmp/tmp.v55g8aoYGv ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 108 -ge 120 ]] + echo -n . .+ sleep 1 + retry=109 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.150Q89CxK8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.pYSmtQy6OO ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.150Q89CxK8 ++ cat /tmp/tmp.pYSmtQy6OO ++ rm /tmp/tmp.150Q89CxK8 /tmp/tmp.pYSmtQy6OO ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 109 -ge 120 ]] + echo -n . .+ sleep 1 + retry=110 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XnYlyHUnqE +++ mktemp ++ local LAST_ERR=/tmp/tmp.U0cziOdQFc ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.XnYlyHUnqE ++ cat /tmp/tmp.U0cziOdQFc ++ rm /tmp/tmp.XnYlyHUnqE /tmp/tmp.U0cziOdQFc ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 110 -ge 120 ]] + echo -n . .+ sleep 1 + retry=111 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.i38nVJO3eU +++ mktemp ++ local LAST_ERR=/tmp/tmp.rcevlLOMWz ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.i38nVJO3eU ++ cat /tmp/tmp.rcevlLOMWz ++ rm /tmp/tmp.i38nVJO3eU /tmp/tmp.rcevlLOMWz ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 111 -ge 120 ]] + echo -n . .+ sleep 1 + retry=112 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NGQHPxYToL +++ mktemp ++ local LAST_ERR=/tmp/tmp.gAiJb0LqOd ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.NGQHPxYToL ++ cat /tmp/tmp.gAiJb0LqOd ++ rm /tmp/tmp.NGQHPxYToL /tmp/tmp.gAiJb0LqOd ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 112 -ge 120 ]] + echo -n . .+ sleep 1 + retry=113 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vCy0T7REAs +++ mktemp ++ local LAST_ERR=/tmp/tmp.80S3SFASP5 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.vCy0T7REAs ++ cat /tmp/tmp.80S3SFASP5 ++ rm /tmp/tmp.vCy0T7REAs /tmp/tmp.80S3SFASP5 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 113 -ge 120 ]] + echo -n . .+ sleep 1 + retry=114 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tdXiEjnN30 +++ mktemp ++ local LAST_ERR=/tmp/tmp.5pqNHGdcbi ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.tdXiEjnN30 ++ cat /tmp/tmp.5pqNHGdcbi ++ rm /tmp/tmp.tdXiEjnN30 /tmp/tmp.5pqNHGdcbi ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 114 -ge 120 ]] + echo -n . .+ sleep 1 + retry=115 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wQq6uZoupi +++ mktemp ++ local LAST_ERR=/tmp/tmp.chjcgHMll6 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.wQq6uZoupi ++ cat /tmp/tmp.chjcgHMll6 ++ rm /tmp/tmp.wQq6uZoupi /tmp/tmp.chjcgHMll6 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 115 -ge 120 ]] + echo -n . .+ sleep 1 + retry=116 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8BehBjlqcq +++ mktemp ++ local LAST_ERR=/tmp/tmp.yHpRoaISrJ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.8BehBjlqcq ++ cat /tmp/tmp.yHpRoaISrJ ++ rm /tmp/tmp.8BehBjlqcq /tmp/tmp.yHpRoaISrJ ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 116 -ge 120 ]] + echo -n . .+ sleep 1 + retry=117 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xkEBE59STp +++ mktemp ++ local LAST_ERR=/tmp/tmp.kaviskxWOT ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xkEBE59STp ++ cat /tmp/tmp.kaviskxWOT ++ rm /tmp/tmp.xkEBE59STp /tmp/tmp.kaviskxWOT ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 117 -ge 120 ]] + echo -n . .+ sleep 1 + retry=118 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.StMOeMUd2X +++ mktemp ++ local LAST_ERR=/tmp/tmp.jKjbS8YrJR ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.StMOeMUd2X ++ cat /tmp/tmp.jKjbS8YrJR ++ rm /tmp/tmp.StMOeMUd2X /tmp/tmp.jKjbS8YrJR ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 118 -ge 120 ]] + echo -n . .+ sleep 1 + retry=119 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5dCkHHAAEC +++ mktemp ++ local LAST_ERR=/tmp/tmp.DSieuruuk5 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5dCkHHAAEC ++ cat /tmp/tmp.DSieuruuk5 ++ rm /tmp/tmp.5dCkHHAAEC /tmp/tmp.DSieuruuk5 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 119 -ge 120 ]] + echo -n . .+ sleep 1 + retry=120 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gyVRJmhJwV +++ mktemp ++ local LAST_ERR=/tmp/tmp.thFH6zJSuN ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.gyVRJmhJwV ++ cat /tmp/tmp.thFH6zJSuN ++ rm /tmp/tmp.gyVRJmhJwV /tmp/tmp.thFH6zJSuN ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 120 -ge 120 ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-0 was not resized, max retries exceeded' persistentvolumeclaim/mongod-data-some-name-rs0-0 was not resized, max retries exceeded + return 1 + return 1 + echo 'Enabling PVC resize' Enabling PVC resize + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "add", "path": "/spec/enableVolumeExpansion", "value":true }]' ++ mktemp + local LAST_OUT=/tmp/tmp.p6WDzVQ5bd ++ mktemp + local LAST_ERR=/tmp/tmp.iZDg08RtJD + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "add", "path": "/spec/enableVolumeExpansion", "value":true }]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.p6WDzVQ5bd perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.iZDg08RtJD + rm /tmp/tmp.p6WDzVQ5bd /tmp/tmp.iZDg08RtJD + return 0 + sleep 10 + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5E9tz8tznk +++ mktemp ++ local LAST_ERR=/tmp/tmp.HtgV830M9x ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5E9tz8tznk ++ cat /tmp/tmp.HtgV830M9x ++ rm /tmp/tmp.5E9tz8tznk /tmp/tmp.HtgV830M9x ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 1 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9oNERXgnxl +++ mktemp ++ local LAST_ERR=/tmp/tmp.wC8n9lHgj9 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9oNERXgnxl ++ cat /tmp/tmp.wC8n9lHgj9 ++ rm /tmp/tmp.9oNERXgnxl /tmp/tmp.wC8n9lHgj9 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 2 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.z08uEA4ThL +++ mktemp ++ local LAST_ERR=/tmp/tmp.Fl84VkF80z ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.z08uEA4ThL ++ cat /tmp/tmp.Fl84VkF80z ++ rm /tmp/tmp.z08uEA4ThL /tmp/tmp.Fl84VkF80z ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 3 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Fcul076VXo +++ mktemp ++ local LAST_ERR=/tmp/tmp.kJxJJ2ocjt ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Fcul076VXo ++ cat /tmp/tmp.kJxJJ2ocjt ++ rm /tmp/tmp.Fcul076VXo /tmp/tmp.kJxJJ2ocjt ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 4 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tEgqwO8wFN +++ mktemp ++ local LAST_ERR=/tmp/tmp.OX4P3VDuRY ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.tEgqwO8wFN ++ cat /tmp/tmp.OX4P3VDuRY ++ rm /tmp/tmp.tEgqwO8wFN /tmp/tmp.OX4P3VDuRY ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 5 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.o9spC06niW +++ mktemp ++ local LAST_ERR=/tmp/tmp.a3ftToqlyQ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.o9spC06niW ++ cat /tmp/tmp.a3ftToqlyQ ++ rm /tmp/tmp.o9spC06niW /tmp/tmp.a3ftToqlyQ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 6 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nz2cSxxzva +++ mktemp ++ local LAST_ERR=/tmp/tmp.Nhu05jxqob ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.nz2cSxxzva ++ cat /tmp/tmp.Nhu05jxqob ++ rm /tmp/tmp.nz2cSxxzva /tmp/tmp.Nhu05jxqob ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo + wait_all_pvc_resize 2Gi + local expected_size=2Gi + local max_retry=120 + local sleep_time=5 ++ kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name +++ mktemp ++ local LAST_OUT=/tmp/tmp.HpgbTStcrC +++ mktemp ++ local LAST_ERR=/tmp/tmp.Z4SZkMVtjj ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pvc -l app.kubernetes.io/component=mongod -o name ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.HpgbTStcrC ++ cat /tmp/tmp.Z4SZkMVtjj ++ rm /tmp/tmp.HpgbTStcrC /tmp/tmp.Z4SZkMVtjj ++ return 0 + for pvc in '$(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name)' + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-0 2Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-0 + local expected_size=2Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wlaR0V0S8H +++ mktemp ++ local LAST_ERR=/tmp/tmp.77axNFaNnt ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.wlaR0V0S8H ++ cat /tmp/tmp.77axNFaNnt ++ rm /tmp/tmp.wlaR0V0S8H /tmp/tmp.77axNFaNnt ++ return 0 + [[ 2Gi == \2\G\i ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized + return 0 + for pvc in '$(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name)' + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-1 2Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-1 + local expected_size=2Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-1 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-1 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dBLhH0ZiBR +++ mktemp ++ local LAST_ERR=/tmp/tmp.rIHcxOCFGC ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.dBLhH0ZiBR ++ cat /tmp/tmp.rIHcxOCFGC ++ rm /tmp/tmp.dBLhH0ZiBR /tmp/tmp.rIHcxOCFGC ++ return 0 + [[ 2Gi == \2\G\i ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-1 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-1 was resized + return 0 + for pvc in '$(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name)' + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-2 2Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-2 + local expected_size=2Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-2 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-2 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ka8go2tA3c +++ mktemp ++ local LAST_ERR=/tmp/tmp.eNrgjff0kX ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Ka8go2tA3c ++ cat /tmp/tmp.eNrgjff0kX ++ rm /tmp/tmp.Ka8go2tA3c /tmp/tmp.eNrgjff0kX ++ return 0 + [[ 2Gi == \2\G\i ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-2 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-2 was resized + return 0 + return 0 + echo + [[ 0 == 1 ]] + [[ -n '' ]] + desc 'create resourcequota' + set +o xtrace ----------------------------------------------------------------------------------- create resourcequota ----------------------------------------------------------------------------------- + apply_resourcequota 7Gi + local quota=7Gi + local default_sc ++ get_default_storageclass ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fvuA1kpAai +++ mktemp ++ local LAST_ERR=/tmp/tmp.jGOeUSqQhX ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.fvuA1kpAai ++ cat /tmp/tmp.jGOeUSqQhX ++ rm /tmp/tmp.fvuA1kpAai /tmp/tmp.jGOeUSqQhX ++ return 0 + default_sc=standard-rwo + echo 'Applying resourcequota for default storageclass standard-rwo with quota 7Gi' Applying resourcequota for default storageclass standard-rwo with quota 7Gi + sed s/STORAGECLASS/standard-rwo/ + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/e2e-tests/pvc-resize/conf/resourcequota.yml + sed s/QUOTA/7Gi/ + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.FABdXNw2u6 ++ mktemp + local LAST_ERR=/tmp/tmp.8xCW9VfuRO + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.FABdXNw2u6 resourcequota/default-storage-quota created + cat /tmp/tmp.8xCW9VfuRO + rm /tmp/tmp.FABdXNw2u6 /tmp/tmp.8xCW9VfuRO + return 0 + patch_pvc_request some-name 3G + local cluster=some-name + local size=3G + echo 'Patching PVC request to 3G in some-name' Patching PVC request to 3G in some-name + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"3G"}]' ++ mktemp + local LAST_OUT=/tmp/tmp.NdPsYMAj36 ++ mktemp + local LAST_ERR=/tmp/tmp.3SmzxAmn4R + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"3G"}]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.NdPsYMAj36 perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.3SmzxAmn4R + rm /tmp/tmp.NdPsYMAj36 /tmp/tmp.3SmzxAmn4R + return 0 + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MK0j9h8KzG +++ mktemp ++ local LAST_ERR=/tmp/tmp.w7WRWLYF77 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.MK0j9h8KzG ++ cat /tmp/tmp.w7WRWLYF77 ++ rm /tmp/tmp.MK0j9h8KzG /tmp/tmp.w7WRWLYF77 ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo + echo + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-0 3Gi + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-0 + local expected_size=3Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gLRMe0aWnZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.HbNdHcPYWz ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.gLRMe0aWnZ ++ cat /tmp/tmp.HbNdHcPYWz ++ rm /tmp/tmp.gLRMe0aWnZ /tmp/tmp.HbNdHcPYWz ++ return 0 + [[ 2Gi == \3\G\i ]] + [[ 0 -ge 120 ]] + echo -n . .+ sleep 5 + retry=1 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.K47FDc23Il +++ mktemp ++ local LAST_ERR=/tmp/tmp.0oVwWG6EF6 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.K47FDc23Il ++ cat /tmp/tmp.0oVwWG6EF6 ++ rm /tmp/tmp.K47FDc23Il /tmp/tmp.0oVwWG6EF6 ++ return 0 + [[ 2Gi == \3\G\i ]] + [[ 1 -ge 120 ]] + echo -n . .+ sleep 5 + retry=2 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FjEhOi2cdV +++ mktemp ++ local LAST_ERR=/tmp/tmp.qfRxcJSNhi ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.FjEhOi2cdV ++ cat /tmp/tmp.qfRxcJSNhi ++ rm /tmp/tmp.FjEhOi2cdV /tmp/tmp.qfRxcJSNhi ++ return 0 + [[ 2Gi == \3\G\i ]] + [[ 2 -ge 120 ]] + echo -n . .+ sleep 5 + retry=3 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TafZsR4Ruw +++ mktemp ++ local LAST_ERR=/tmp/tmp.7LRdpiFzzJ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.TafZsR4Ruw ++ cat /tmp/tmp.7LRdpiFzzJ ++ rm /tmp/tmp.TafZsR4Ruw /tmp/tmp.7LRdpiFzzJ ++ return 0 + [[ 2Gi == \3\G\i ]] + [[ 3 -ge 120 ]] + echo -n . .+ sleep 5 + retry=4 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cJDVZmimL9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.1FLPIIZbYG ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.cJDVZmimL9 ++ cat /tmp/tmp.1FLPIIZbYG ++ rm /tmp/tmp.cJDVZmimL9 /tmp/tmp.1FLPIIZbYG ++ return 0 + [[ 2Gi == \3\G\i ]] + [[ 4 -ge 120 ]] + echo -n . .+ sleep 5 + retry=5 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KbydbbR3Us +++ mktemp ++ local LAST_ERR=/tmp/tmp.l1VZ5AOwxi ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.KbydbbR3Us ++ cat /tmp/tmp.l1VZ5AOwxi ++ rm /tmp/tmp.KbydbbR3Us /tmp/tmp.l1VZ5AOwxi ++ return 0 + [[ 2Gi == \3\G\i ]] + [[ 5 -ge 120 ]] + echo -n . .+ sleep 5 + retry=6 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UREr3GYY9G +++ mktemp ++ local LAST_ERR=/tmp/tmp.EVcIeDdosa ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.UREr3GYY9G ++ cat /tmp/tmp.EVcIeDdosa ++ rm /tmp/tmp.UREr3GYY9G /tmp/tmp.EVcIeDdosa ++ return 0 + [[ 3Gi == \3\G\i ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized + return 0 + apply_resourcequota 9Gi + local quota=9Gi + local default_sc ++ get_default_storageclass ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Zn4gt1jOm8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.K3zir0lp4t ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Zn4gt1jOm8 ++ cat /tmp/tmp.K3zir0lp4t ++ rm /tmp/tmp.Zn4gt1jOm8 /tmp/tmp.K3zir0lp4t ++ return 0 + default_sc=standard-rwo + echo 'Applying resourcequota for default storageclass standard-rwo with quota 9Gi' Applying resourcequota for default storageclass standard-rwo with quota 9Gi + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/e2e-tests/pvc-resize/conf/resourcequota.yml + sed s/QUOTA/9Gi/ + sed s/STORAGECLASS/standard-rwo/ + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.rFlrHdRyap ++ mktemp + local LAST_ERR=/tmp/tmp.23gKKQSJFD + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.rFlrHdRyap resourcequota/default-storage-quota configured + cat /tmp/tmp.23gKKQSJFD + rm /tmp/tmp.rFlrHdRyap /tmp/tmp.23gKKQSJFD + return 0 + patch_pvc_request some-name 3G + local cluster=some-name + local size=3G + echo 'Patching PVC request to 3G in some-name' Patching PVC request to 3G in some-name + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"3G"}]' ++ mktemp + local LAST_OUT=/tmp/tmp.WK7uc3S9Ps ++ mktemp + local LAST_ERR=/tmp/tmp.0TXFTA3dK2 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"3G"}]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.WK7uc3S9Ps perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.0TXFTA3dK2 + rm /tmp/tmp.WK7uc3S9Ps /tmp/tmp.0TXFTA3dK2 + return 0 + wait_cluster_consistency some-name 42 + local cluster_name=some-name + local wait_time=42 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Z29wjoeTIv +++ mktemp ++ local LAST_ERR=/tmp/tmp.oeK958ezeV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Z29wjoeTIv ++ cat /tmp/tmp.oeK958ezeV ++ rm /tmp/tmp.Z29wjoeTIv /tmp/tmp.oeK958ezeV ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 1 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NeXi0J2HJM +++ mktemp ++ local LAST_ERR=/tmp/tmp.rZJ92pBXKw ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.NeXi0J2HJM ++ cat /tmp/tmp.rZJ92pBXKw ++ rm /tmp/tmp.NeXi0J2HJM /tmp/tmp.rZJ92pBXKw ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 2 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Qz1Q6z18f8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.2Bn4myJ35a ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Qz1Q6z18f8 ++ cat /tmp/tmp.2Bn4myJ35a ++ rm /tmp/tmp.Qz1Q6z18f8 /tmp/tmp.2Bn4myJ35a ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 3 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KfzMkeUjtv +++ mktemp ++ local LAST_ERR=/tmp/tmp.G3JNAX0Wb3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.KfzMkeUjtv ++ cat /tmp/tmp.G3JNAX0Wb3 ++ rm /tmp/tmp.KfzMkeUjtv /tmp/tmp.G3JNAX0Wb3 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 4 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vTC7BXYMek +++ mktemp ++ local LAST_ERR=/tmp/tmp.JhfYWU2y2R ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.vTC7BXYMek ++ cat /tmp/tmp.JhfYWU2y2R ++ rm /tmp/tmp.vTC7BXYMek /tmp/tmp.JhfYWU2y2R ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 5 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WEUngCm1oO +++ mktemp ++ local LAST_ERR=/tmp/tmp.jiT6pnQeAF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.WEUngCm1oO ++ cat /tmp/tmp.jiT6pnQeAF ++ rm /tmp/tmp.WEUngCm1oO /tmp/tmp.jiT6pnQeAF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 6 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.k3cNk9dpWs +++ mktemp ++ local LAST_ERR=/tmp/tmp.UJoLQ3Og29 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.k3cNk9dpWs ++ cat /tmp/tmp.UJoLQ3Og29 ++ rm /tmp/tmp.k3cNk9dpWs /tmp/tmp.UJoLQ3Og29 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 7 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qcIb70JW4z +++ mktemp ++ local LAST_ERR=/tmp/tmp.KI04JlW8kI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.qcIb70JW4z ++ cat /tmp/tmp.KI04JlW8kI ++ rm /tmp/tmp.qcIb70JW4z /tmp/tmp.KI04JlW8kI ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo + echo + wait_all_pvc_resize 3Gi + local expected_size=3Gi + local max_retry=120 + local sleep_time=5 ++ kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name +++ mktemp ++ local LAST_OUT=/tmp/tmp.csK5v3T9Mo +++ mktemp ++ local LAST_ERR=/tmp/tmp.adEplleZkk ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pvc -l app.kubernetes.io/component=mongod -o name ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.csK5v3T9Mo ++ cat /tmp/tmp.adEplleZkk ++ rm /tmp/tmp.csK5v3T9Mo /tmp/tmp.adEplleZkk ++ return 0 + for pvc in '$(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name)' + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-0 3Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-0 + local expected_size=3Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BjnQNdhrZb +++ mktemp ++ local LAST_ERR=/tmp/tmp.mUHlwlgNNl ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.BjnQNdhrZb ++ cat /tmp/tmp.mUHlwlgNNl ++ rm /tmp/tmp.BjnQNdhrZb /tmp/tmp.mUHlwlgNNl ++ return 0 + [[ 3Gi == \3\G\i ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized + return 0 + for pvc in '$(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name)' + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-1 3Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-1 + local expected_size=3Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-1 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-1 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.joSdEd4MsH +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZqxdMuGAc9 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.joSdEd4MsH ++ cat /tmp/tmp.ZqxdMuGAc9 ++ rm /tmp/tmp.joSdEd4MsH /tmp/tmp.ZqxdMuGAc9 ++ return 0 + [[ 3Gi == \3\G\i ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-1 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-1 was resized + return 0 + for pvc in '$(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name)' + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-2 3Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-2 + local expected_size=3Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-2 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-2 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4XKNWFjVGb +++ mktemp ++ local LAST_ERR=/tmp/tmp.gDrBvXeMiS ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4XKNWFjVGb ++ cat /tmp/tmp.gDrBvXeMiS ++ rm /tmp/tmp.4XKNWFjVGb /tmp/tmp.gDrBvXeMiS ++ return 0 + [[ 3Gi == \3\G\i ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-2 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-2 was resized + return 0 + return 0 + desc 'test downscale' + set +o xtrace ----------------------------------------------------------------------------------- test downscale ----------------------------------------------------------------------------------- + patch_pvc_request some-name 1G + local cluster=some-name + local size=1G + echo 'Patching PVC request to 1G in some-name' Patching PVC request to 1G in some-name + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"1G"}]' ++ mktemp + local LAST_OUT=/tmp/tmp.zMwBF3K63i ++ mktemp + local LAST_ERR=/tmp/tmp.hkkoefaP6Y + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"1G"}]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.zMwBF3K63i perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.hkkoefaP6Y + rm /tmp/tmp.zMwBF3K63i /tmp/tmp.hkkoefaP6Y + return 0 + wait_cluster_status some-name error + local cluster=some-name + local expected=error + echo -n 'Waiting for psmdb/some-name status to be error' Waiting for psmdb/some-name status to be error++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.B15YCuDoXr +++ mktemp ++ local LAST_ERR=/tmp/tmp.QNwUmng1RR ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.B15YCuDoXr ++ cat /tmp/tmp.QNwUmng1RR ++ rm /tmp/tmp.B15YCuDoXr /tmp/tmp.QNwUmng1RR ++ return 0 + [[ error == \e\r\r\o\r ]] + echo + echo 'psmdb/some-name status is error' psmdb/some-name status is error + patch_pvc_request some-name 3G + local cluster=some-name + local size=3G + echo 'Patching PVC request to 3G in some-name' Patching PVC request to 3G in some-name + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"3G"}]' ++ mktemp + local LAST_OUT=/tmp/tmp.McPPpKVZ67 ++ mktemp + local LAST_ERR=/tmp/tmp.nyo8bK0TqR + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"3G"}]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.McPPpKVZ67 perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.nyo8bK0TqR + rm /tmp/tmp.McPPpKVZ67 /tmp/tmp.nyo8bK0TqR + return 0 + wait_cluster_status some-name ready + local cluster=some-name + local expected=ready + echo -n 'Waiting for psmdb/some-name status to be ready' Waiting for psmdb/some-name status to be ready++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.w5ka131Qpe +++ mktemp ++ local LAST_ERR=/tmp/tmp.rlvOLscbeq ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.w5ka131Qpe ++ cat /tmp/tmp.rlvOLscbeq ++ rm /tmp/tmp.w5ka131Qpe /tmp/tmp.rlvOLscbeq ++ return 0 + [[ error == \r\e\a\d\y ]] + [[ 7 -ge 60 ]] + echo -n . .+ sleep 5 + retry=8 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jwIwaOLb7U +++ mktemp ++ local LAST_ERR=/tmp/tmp.U15wgaCqA2 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jwIwaOLb7U ++ cat /tmp/tmp.U15wgaCqA2 ++ rm /tmp/tmp.jwIwaOLb7U /tmp/tmp.U15wgaCqA2 ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo + echo 'psmdb/some-name status is ready' psmdb/some-name status is ready + destroy pvc-resize-5656 + local namespace=pvc-resize-5656 + local ignore_logs=true + [[ 0 == 1 ]] + desc 'destroy cluster/operator and all other resources' + set +o xtrace ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + '[' true == false ']' + delete_backups + desc 'Delete psmdb-backup' + set +o xtrace ----------------------------------------------------------------------------------- Delete psmdb-backup ----------------------------------------------------------------------------------- ++ kubectl_bin get psmdb-backup --no-headers ++ wc -l +++ mktemp ++ local LAST_OUT=/tmp/tmp.NLvW9hBDlh +++ mktemp ++ local LAST_ERR=/tmp/tmp.3ss14b6Trf ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb-backup --no-headers ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.NLvW9hBDlh ++ cat /tmp/tmp.3ss14b6Trf No resources found in pvc-resize-5656 namespace. ++ rm /tmp/tmp.NLvW9hBDlh /tmp/tmp.3ss14b6Trf ++ return 0 + '[' 0 '!=' 0 ']' + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.e88QeTv5je ++ mktemp + local LAST_ERR=/tmp/tmp.7q004ULaQL + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.e88QeTv5je customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.7q004ULaQL + rm /tmp/tmp.e88QeTv5je /tmp/tmp.7q004ULaQL + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/deploy/crd.yaml ++ grep -v '\-\-\-' + for crd_name in '$(yq eval '\''.metadata.name'\'' "${src_dir}/deploy/crd.yaml" | grep -v '\''\-\-\-'\'')' + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.GDHORezMb7 ++ mktemp + local LAST_ERR=/tmp/tmp.xRJUvYnpFf + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.GDHORezMb7 + cat /tmp/tmp.xRJUvYnpFf + rm /tmp/tmp.GDHORezMb7 /tmp/tmp.xRJUvYnpFf + return 0 + for crd_name in '$(yq eval '\''.metadata.name'\'' "${src_dir}/deploy/crd.yaml" | grep -v '\''\-\-\-'\'')' + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.0wMhWetrMM ++ mktemp + local LAST_ERR=/tmp/tmp.V0hgyG3JYX + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.0wMhWetrMM + cat /tmp/tmp.V0hgyG3JYX + rm /tmp/tmp.0wMhWetrMM /tmp/tmp.V0hgyG3JYX + return 0 + for crd_name in '$(yq eval '\''.metadata.name'\'' "${src_dir}/deploy/crd.yaml" | grep -v '\''\-\-\-'\'')' + grep -v NAMESPACE + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch perconaservermongodbs.psmdb.percona.com -n pvc-resize-5656 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaservermongodb.psmdb.percona.com/some-name patched + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.CcaXBqUT48 ++ mktemp + local LAST_ERR=/tmp/tmp.9xMapHdlHy + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.CcaXBqUT48 customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com condition met + cat /tmp/tmp.9xMapHdlHy + rm /tmp/tmp.CcaXBqUT48 /tmp/tmp.9xMapHdlHy + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.DhJJL0lMC2 ++ mktemp + local LAST_ERR=/tmp/tmp.U46eZ6OJfv + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1926/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.DhJJL0lMC2 clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.U46eZ6OJfv + rm /tmp/tmp.DhJJL0lMC2 /tmp/tmp.U46eZ6OJfv + return 0 + destroy_cert_manager + kubectl_bin delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.3laUQ3eRrs ++ mktemp + local LAST_ERR=/tmp/tmp.UcF90WQOFZ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.3laUQ3eRrs + cat /tmp/tmp.UcF90WQOFZ Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.3laUQ3eRrs + cat /tmp/tmp.UcF90WQOFZ Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 4 + for i in '$(seq 0 2)' + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.3laUQ3eRrs + cat /tmp/tmp.UcF90WQOFZ Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 8 + cat /tmp/tmp.3laUQ3eRrs + cat /tmp/tmp.UcF90WQOFZ Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + rm /tmp/tmp.3laUQ3eRrs /tmp/tmp.UcF90WQOFZ + return 1 + true + '[' -n '' ']' + '[' -n psmdb-operator ']' + kubectl_bin delete --grace-period=0 --force=true namespace pvc-resize-5656 + rm -rf /tmp/tmp.kcQwVgU6Ei ++ mktemp + kubectl_bin delete --grace-period=0 --force=true namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.Etsi8SlbTM ++ mktemp + desc 'test passed' + set +o xtrace ----------------------------------------------------------------------------------- test passed ----------------------------------------------------------------------------------- + local LAST_OUT=/tmp/tmp.fMIJiRTuez + local LAST_ERR=/tmp/tmp.p2jzg8xmje + local exit_status=0 + local timeout=4 ++ mktemp ++ seq 0 2 + local LAST_ERR=/tmp/tmp.05wHvx5LDp + local exit_status=0 + local timeout=4 + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace pvc-resize-5656 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace psmdb-operator