Log: /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/e2e-tests/logs/pvc-resize.log Warning: version difference between client (1.35) and server (1.31) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.35) and server (1.31) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.35) and server (1.31) exceeds the supported minor version skew of +/-1 + '[' 0 == 1 ']' + ensure_default_sc_allows_expansion + local default_sc ++ get_default_storageclass ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rKizFLF4a4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.vxhfSDssp7 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.rKizFLF4a4 ++ cat /tmp/tmp.vxhfSDssp7 ++ rm /tmp/tmp.rKizFLF4a4 /tmp/tmp.vxhfSDssp7 ++ return 0 + default_sc=standard-rwo + echo 'Checking if default storageclass standard-rwo allows volume expansion' Checking if default storageclass standard-rwo allows volume expansion + local allowVolumeExpansion ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.name=="standard-rwo")].allowVolumeExpansion}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SOpwPstigF +++ mktemp ++ local LAST_ERR=/tmp/tmp.yrilQrSerY ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.name=="standard-rwo")].allowVolumeExpansion}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.SOpwPstigF ++ cat /tmp/tmp.yrilQrSerY ++ rm /tmp/tmp.SOpwPstigF /tmp/tmp.yrilQrSerY ++ return 0 + allowVolumeExpansion=true + [[ true != \t\r\u\e ]] + create_infra pvc-resize-26375 + local ns=pvc-resize-26375 + [[ 1 == 1 ]] + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.lcZz39CPEk ++ mktemp + local LAST_ERR=/tmp/tmp.cfFWbxsQ2k + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.lcZz39CPEk customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.cfFWbxsQ2k + rm /tmp/tmp.lcZz39CPEk /tmp/tmp.cfFWbxsQ2k + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/deploy/crd.yaml ++ grep -v '\-\-\-' grep: warning: stray \ before - grep: warning: stray \ before - + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.vwyHbKE6dQ ++ mktemp + local LAST_ERR=/tmp/tmp.dKMajWQCAz + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.vwyHbKE6dQ + cat /tmp/tmp.dKMajWQCAz + rm /tmp/tmp.vwyHbKE6dQ /tmp/tmp.dKMajWQCAz + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.KGTO8Whn4F ++ mktemp + local LAST_ERR=/tmp/tmp.inY5K92cHS + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.KGTO8Whn4F + cat /tmp/tmp.inY5K92cHS + rm /tmp/tmp.KGTO8Whn4F /tmp/tmp.inY5K92cHS + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' No resources found + kubectl patch perconaservermongodbs.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbs" + : + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.IQG9VMnlWF ++ mktemp + local LAST_ERR=/tmp/tmp.ZDuO52RW3H + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.IQG9VMnlWF + cat /tmp/tmp.ZDuO52RW3H + rm /tmp/tmp.IQG9VMnlWF /tmp/tmp.ZDuO52RW3H + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.cNU2mjywtX ++ mktemp + local LAST_ERR=/tmp/tmp.KGEvNuBwCQ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.cNU2mjywtX clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.KGEvNuBwCQ + rm /tmp/tmp.cNU2mjywtX /tmp/tmp.KGEvNuBwCQ + return 0 + check_crd_for_deletion PR-2213-482da7af + local git_tag=PR-2213-482da7af ++ curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/PR-2213-482da7af/deploy/crd.yaml ++ yq eval .metadata.name ++ /usr/sbin/sed s/---//g ++ /usr/sbin/sed ':a;N;$!ba;s/\n/ /g' + for crd_name in $(curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/${git_tag}/deploy/crd.yaml | yq eval '.metadata.name' | $sed 's/---//g' | $sed ':a;N;$!ba;s/\n/ /g') ++ kubectl_bin get crd/null -o 'jsonpath={.status.conditions[-1].type}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.K8EyW8Z2iV +++ mktemp ++ local LAST_ERR=/tmp/tmp.kjaETn43iU ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.K8EyW8Z2iV ++ cat /tmp/tmp.kjaETn43iU Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 0 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.K8EyW8Z2iV ++ cat /tmp/tmp.kjaETn43iU Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 4 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.K8EyW8Z2iV ++ cat /tmp/tmp.kjaETn43iU Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 8 ++ cat /tmp/tmp.K8EyW8Z2iV ++ cat /tmp/tmp.kjaETn43iU Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ rm /tmp/tmp.K8EyW8Z2iV /tmp/tmp.kjaETn43iU ++ return 1 + [[ '' == \T\e\r\m\i\n\a\t\i\n\g ]] + '[' -n psmdb-operator ']' + create_namespace psmdb-operator + local namespace=psmdb-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep validate-auth + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + grep -E -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' + awk '{print$1}' + '[' -n '' ']' + desc 'cleaned up old namespaces psmdb-operator' + xargs kubectl delete ns + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace psmdb-operator --ignore-not-found ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.MqHT5liAC1 + local LAST_OUT=/tmp/tmp.y3PL3vKkHQ ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.CxKLSo4pna + local exit_status=0 + local timeout=4 ++ seq 0 2 + local LAST_ERR=/tmp/tmp.asQqOswLk7 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get ns + for i in $(seq 0 2) + set +e + kubectl delete namespace psmdb-operator --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.y3PL3vKkHQ + cat /tmp/tmp.CxKLSo4pna + rm /tmp/tmp.y3PL3vKkHQ /tmp/tmp.CxKLSo4pna + return 0 namespace "pvc-resize-18478" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.MqHT5liAC1 namespace "psmdb-operator" deleted + cat /tmp/tmp.asQqOswLk7 + rm /tmp/tmp.MqHT5liAC1 /tmp/tmp.asQqOswLk7 + return 0 + kubectl_bin wait --for=delete namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.sML39heCby ++ mktemp + local LAST_ERR=/tmp/tmp.qKWmTwHvWN + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.sML39heCby + cat /tmp/tmp.qKWmTwHvWN + rm /tmp/tmp.sML39heCby /tmp/tmp.qKWmTwHvWN + return 0 + desc 'create namespace psmdb-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.Dt9EHLmZio ++ mktemp + local LAST_ERR=/tmp/tmp.OgrOhZfv6v + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Dt9EHLmZio namespace/psmdb-operator created + cat /tmp/tmp.OgrOhZfv6v + rm /tmp/tmp.Dt9EHLmZio /tmp/tmp.OgrOhZfv6v + return 0 + set_kube_ctx psmdb-operator + local namespace=psmdb-operator ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.4Q3va6IQi3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.ENq8zYLVVF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4Q3va6IQi3 ++ cat /tmp/tmp.ENq8zYLVVF ++ rm /tmp/tmp.4Q3va6IQi3 /tmp/tmp.ENq8zYLVVF ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2213-482da7af-11-cluster3 --namespace=psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.dr99G5vTQv ++ mktemp + local LAST_ERR=/tmp/tmp.4QKeinvgm2 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2213-482da7af-11-cluster3 --namespace=psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.dr99G5vTQv Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2213-482da7af-11-cluster3" modified. + cat /tmp/tmp.4QKeinvgm2 + rm /tmp/tmp.dr99G5vTQv /tmp/tmp.4QKeinvgm2 + return 0 + deploy_operator + desc 'start PSMDB operator: perconalab/percona-server-mongodb-operator:PR-2213-482da7af' + set +o xtrace ----------------------------------------------------------------------------------- start PSMDB operator: perconalab/percona-server-mongodb-operator:PR-2213-482da7af ----------------------------------------------------------------------------------- + local cr_file + '[' -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/e2e-tests/pvc-resize/conf/crd.yaml ']' + cr_file=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/deploy/crd.yaml + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.c0G4jf2rRV ++ mktemp + local LAST_ERR=/tmp/tmp.vgJRvojuuV + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.c0G4jf2rRV customresourcedefinition.apiextensions.k8s.io/perconaservermongodbbackups.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbrestores.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com serverside-applied + cat /tmp/tmp.vgJRvojuuV + rm /tmp/tmp.c0G4jf2rRV /tmp/tmp.vgJRvojuuV + return 0 + '[' -n psmdb-operator ']' + apply_rbac cw-rbac + local operator_namespace=psmdb-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: psmdb-operator^' + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.pGYIIbxMOp ++ mktemp + local LAST_ERR=/tmp/tmp.EW03NvVF5l + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.pGYIIbxMOp clusterrole.rbac.authorization.k8s.io/percona-server-mongodb-operator created serviceaccount/percona-server-mongodb-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-server-mongodb-operator created + cat /tmp/tmp.EW03NvVF5l + rm /tmp/tmp.pGYIIbxMOp /tmp/tmp.EW03NvVF5l + return 0 + yq eval ' (.spec.template.spec.containers[].image = "perconalab/percona-server-mongodb-operator:PR-2213-482da7af") | ((.. | select(.[] == "DISABLE_TELEMETRY")) |= .value="true") | ((.. | select(.[] == "LOG_LEVEL")) |= .value="DEBUG")' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/deploy/cw-operator.yaml + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.klmEV3sXmE ++ mktemp + local LAST_ERR=/tmp/tmp.xvC61HPVKC + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.klmEV3sXmE deployment.apps/percona-server-mongodb-operator created + cat /tmp/tmp.xvC61HPVKC + rm /tmp/tmp.klmEV3sXmE /tmp/tmp.xvC61HPVKC + return 0 + sleep 20 ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.4NstVyZstx +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ex8OPipA8w ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4NstVyZstx ++ cat /tmp/tmp.Ex8OPipA8w ++ rm /tmp/tmp.4NstVyZstx /tmp/tmp.Ex8OPipA8w ++ return 0 + wait_operator_pod percona-server-mongodb-operator-68bf6755c6-4tmtb + local pod=percona-server-mongodb-operator-68bf6755c6-4tmtb + set +o xtrace waiting for pod/percona-server-mongodb-operator-68bf6755c6-4tmtb to be ready.OK + echo 'Print operator info from log' Print operator info from log + grep 'Manager starting up' ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.wHWPJR4EqL +++ mktemp ++ local LAST_ERR=/tmp/tmp.Af5CE1nhFS ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.wHWPJR4EqL ++ cat /tmp/tmp.Af5CE1nhFS ++ rm /tmp/tmp.wHWPJR4EqL /tmp/tmp.Af5CE1nhFS ++ return 0 + kubectl_bin logs -n psmdb-operator percona-server-mongodb-operator-68bf6755c6-4tmtb ++ mktemp + local LAST_OUT=/tmp/tmp.oa6FNkmhYY ++ mktemp + local LAST_ERR=/tmp/tmp.5m6mIkQEMJ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl logs -n psmdb-operator percona-server-mongodb-operator-68bf6755c6-4tmtb + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.oa6FNkmhYY + cat /tmp/tmp.5m6mIkQEMJ + rm /tmp/tmp.oa6FNkmhYY /tmp/tmp.5m6mIkQEMJ + return 0 2026-01-23T12:50:54.150Z INFO setup Manager starting up {"gitCommit": "482da7afc76cb5e5983e0dafeee205d21b673bb3", "gitBranch": "PR-2213-482da7af", "buildTime": "", "goVersion": "go1.25.6", "os": "linux", "arch": "amd64"} + create_namespace pvc-resize-26375 + local namespace=pvc-resize-26375 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + grep -E -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' + awk '{print$1}' ++ mktemp + '[' -n '' ']' + desc 'cleaned up old namespaces pvc-resize-26375' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pvc-resize-26375 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pvc-resize-26375 --ignore-not-found + xargs kubectl delete ns + local LAST_OUT=/tmp/tmp.n6pPhhVmTL ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.5T9GBd8yMa + local exit_status=0 + local timeout=4 + local LAST_OUT=/tmp/tmp.s5b9p8wes0 ++ seq 0 2 ++ mktemp + for i in $(seq 0 2) + set +e + kubectl get ns + local LAST_ERR=/tmp/tmp.JxGupATJIq + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete namespace pvc-resize-26375 --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.n6pPhhVmTL + cat /tmp/tmp.5T9GBd8yMa + rm /tmp/tmp.n6pPhhVmTL /tmp/tmp.5T9GBd8yMa + return 0 error: resource(s) were provided, but no name was specified + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.s5b9p8wes0 + cat /tmp/tmp.JxGupATJIq + rm /tmp/tmp.s5b9p8wes0 /tmp/tmp.JxGupATJIq + return 0 + kubectl_bin wait --for=delete namespace pvc-resize-26375 ++ mktemp + local LAST_OUT=/tmp/tmp.MOpJvHYvAp ++ mktemp + local LAST_ERR=/tmp/tmp.fXQIPQRrRb + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete namespace pvc-resize-26375 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.MOpJvHYvAp + cat /tmp/tmp.fXQIPQRrRb + rm /tmp/tmp.MOpJvHYvAp /tmp/tmp.fXQIPQRrRb + return 0 + desc 'create namespace pvc-resize-26375' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pvc-resize-26375 ----------------------------------------------------------------------------------- + kubectl_bin create namespace pvc-resize-26375 ++ mktemp + local LAST_OUT=/tmp/tmp.uGSbe06E3O ++ mktemp + local LAST_ERR=/tmp/tmp.TWrZf6TGtx + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace pvc-resize-26375 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.uGSbe06E3O namespace/pvc-resize-26375 created + cat /tmp/tmp.TWrZf6TGtx + rm /tmp/tmp.uGSbe06E3O /tmp/tmp.TWrZf6TGtx + return 0 + set_kube_ctx pvc-resize-26375 + local namespace=pvc-resize-26375 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZLIcFAfXZf +++ mktemp ++ local LAST_ERR=/tmp/tmp.YxCJ6mfx3y ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ZLIcFAfXZf ++ cat /tmp/tmp.YxCJ6mfx3y ++ rm /tmp/tmp.ZLIcFAfXZf /tmp/tmp.YxCJ6mfx3y ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2213-482da7af-11-cluster3 --namespace=pvc-resize-26375 ++ mktemp + local LAST_OUT=/tmp/tmp.KC0t1LLspx ++ mktemp + local LAST_ERR=/tmp/tmp.ibtRdsKhJU + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2213-482da7af-11-cluster3 --namespace=pvc-resize-26375 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.KC0t1LLspx Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2213-482da7af-11-cluster3" modified. + cat /tmp/tmp.ibtRdsKhJU + rm /tmp/tmp.KC0t1LLspx /tmp/tmp.ibtRdsKhJU + return 0 + desc 'create secrets and psmdb client' + set +o xtrace ----------------------------------------------------------------------------------- create secrets and psmdb client ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/e2e-tests/conf/secrets.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.kgFsdX2iGq ++ mktemp + local LAST_ERR=/tmp/tmp.K98LkP17ib + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/e2e-tests/conf/secrets.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/e2e-tests/conf/client.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.kgFsdX2iGq secret/some-users created deployment.apps/psmdb-client created + cat /tmp/tmp.K98LkP17ib + rm /tmp/tmp.kgFsdX2iGq /tmp/tmp.K98LkP17ib + return 0 + desc 'create PSMDB cluster' + set +o xtrace ----------------------------------------------------------------------------------- create PSMDB cluster ----------------------------------------------------------------------------------- + cluster=some-name + '[' 0 == 1 ']' + spinup_psmdb some-name-rs0 /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/e2e-tests/pvc-resize/conf/some-name.yml + local cluster=some-name-rs0 + local config=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/e2e-tests/pvc-resize/conf/some-name.yml + local size=3 + desc 'create first PSMDB cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PSMDB cluster ----------------------------------------------------------------------------------- + apply_cluster /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/e2e-tests/pvc-resize/conf/some-name.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/e2e-tests/pvc-resize/conf/some-name.yml + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/e2e-tests/pvc-resize/conf/some-name.yml + kubectl_bin apply -f - + yq eval '(.spec | select(has("pmm"))).pmm.image = "percona/pmm-client:2.44.1-1"' + yq eval '(.spec | select(.image == null)).image = "perconalab/percona-server-mongodb-operator:main-mongod8.0"' ++ mktemp + yq eval '(.spec | select(has("initImage"))).initImage = "perconalab/percona-server-mongodb-operator:PR-2213-482da7af"' + /usr/sbin/sed -e s/NAME_SPACE/pvc-resize-26375/g + yq eval '(.spec | select(has("backup"))).backup.image = "perconalab/percona-server-mongodb-operator:main-backup"' + yq eval '.spec.upgradeOptions.apply="Never"' + local LAST_OUT=/tmp/tmp.tYHWdBYmCA ++ mktemp + local LAST_ERR=/tmp/tmp.f32xXxNKSH + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.tYHWdBYmCA perconaservermongodb.psmdb.percona.com/some-name created + cat /tmp/tmp.f32xXxNKSH + rm /tmp/tmp.tYHWdBYmCA /tmp/tmp.f32xXxNKSH + return 0 + desc 'check if Pod is started' + set +o xtrace ----------------------------------------------------------------------------------- check if Pod is started ----------------------------------------------------------------------------------- + wait_for_running some-name-rs0 3 + local name=some-name-rs0 + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=rs0 + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-rs0-0 + local pod=some-name-rs0-0 + set +o xtrace waiting for pod/some-name-rs0-0 to be ready.....................OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-rs0-1 + local pod=some-name-rs0-1 + set +o xtrace waiting for pod/some-name-rs0-1 to be ready............OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BAxCw95DgO +++ mktemp ++ local LAST_ERR=/tmp/tmp.NPNTKkGzMR ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.BAxCw95DgO ++ cat /tmp/tmp.NPNTKkGzMR ++ rm /tmp/tmp.BAxCw95DgO /tmp/tmp.NPNTKkGzMR ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-rs0-2 + local pod=some-name-rs0-2 + set +o xtrace waiting for pod/some-name-rs0-2 to be ready.............OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bVdnBXFESn +++ mktemp ++ local LAST_ERR=/tmp/tmp.iCTZ1RbeIE ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.bVdnBXFESn ++ cat /tmp/tmp.iCTZ1RbeIE ++ rm /tmp/tmp.bVdnBXFESn /tmp/tmp.iCTZ1RbeIE ++ return 0 + [[ '' == \t\r\u\e ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8vByAGKvh3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.bytIDYuF5w ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.8vByAGKvh3 ++ cat /tmp/tmp.bytIDYuF5w ++ rm /tmp/tmp.8vByAGKvh3 /tmp/tmp.bytIDYuF5w ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness + sleep 20 + compare_kubectl statefulset/some-name-rs0 + local resource=statefulset/some-name-rs0 + local postfix= + local skip_generation_check= + local expected_result=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/e2e-tests/pvc-resize/compare/statefulset_some-name-rs0.yml + local new_result=/tmp/tmp.qmg2Iuu1GB/statefulset_some-name-rs0.yml + '[' -n '' -a -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/e2e-tests/pvc-resize/compare/statefulset_some-name-rs0-oc.yml ']' + kubectl_bin get -o yaml statefulset/some-name-rs0 + yq eval ' del(.metadata.ownerReferences[].apiVersion) | del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "NAMESPACE")) | del(.metadata.selfLink) | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.status) | (.. | select(tag == "!!str")) |= sub("pvc-resize-26375", "NAME_SPACE") | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "extensions/v1beta1")) = "apps/v1" | (.. | select(. == "batch/v1beta1")) = "batch/v1" ' - ++ mktemp + local LAST_OUT=/tmp/tmp.Neo6yYSoqi ++ mktemp + local LAST_ERR=/tmp/tmp.AQv7Z0tOCr + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get -o yaml statefulset/some-name-rs0 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Neo6yYSoqi + cat /tmp/tmp.AQv7Z0tOCr + rm /tmp/tmp.Neo6yYSoqi /tmp/tmp.AQv7Z0tOCr + return 0 + yq -i eval 'del(.spec.persistentVolumeClaimRetentionPolicy)' /tmp/tmp.qmg2Iuu1GB/statefulset_some-name-rs0.yml + version_gt 1.22 ++ echo '1.31 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + yq -i eval 'del(.spec.internalTrafficPolicy)' /tmp/tmp.qmg2Iuu1GB/statefulset_some-name-rs0.yml + yq -i eval 'del(.spec.allocateLoadBalancerNodePorts)' /tmp/tmp.qmg2Iuu1GB/statefulset_some-name-rs0.yml + [[ /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/e2e-tests/pvc-resize/compare/statefulset_some-name-rs0.yml == */cronjob* ]] + '[' -n '' ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/e2e-tests/pvc-resize/compare/statefulset_some-name-rs0.yml /tmp/tmp.qmg2Iuu1GB/statefulset_some-name-rs0.yml + log 'compare_kubectl: statefulset/some-name-rs0 OK' + set +o xtrace [2026-01-23T12:53:48+0000] compare_kubectl: statefulset/some-name-rs0 OK + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + run_mongo 'db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})' userAdmin:userAdmin123456@some-name-rs0.pvc-resize-26375 + local 'command=db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})' + local uri=userAdmin:userAdmin123456@some-name-rs0.pvc-resize-26375 + local driver=mongodb+srv + local suffix=.svc.cluster.local + local mongo_flag= + local replica_set=rs0 + [[ userAdmin:userAdmin123456@some-name-rs0.pvc-resize-26375 == *cfg* ]] ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AGrxzHCiiY +++ mktemp ++ local LAST_ERR=/tmp/tmp.l9RJzebnqk ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.AGrxzHCiiY ++ cat /tmp/tmp.l9RJzebnqk ++ rm /tmp/tmp.AGrxzHCiiY /tmp/tmp.l9RJzebnqk ++ return 0 + local client_container=psmdb-client-696897d69b-mv54g + kubectl_bin exec psmdb-client-696897d69b-mv54g -- bash -c 'printf '\''db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})\n'\'' | mongo mongodb+srv://userAdmin:userAdmin123456@some-name-rs0.pvc-resize-26375.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.CfyIAjBbqp ++ mktemp + local LAST_ERR=/tmp/tmp.I5ZopP6aMH + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-696897d69b-mv54g -- bash -c 'printf '\''db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})\n'\'' | mongo mongodb+srv://userAdmin:userAdmin123456@some-name-rs0.pvc-resize-26375.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.CfyIAjBbqp Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-rs0-1.some-name-rs0.pvc-resize-26375.svc.cluster.local:27017,some-name-rs0-2.some-name-rs0.pvc-resize-26375.svc.cluster.local:27017,some-name-rs0-0.some-name-rs0.pvc-resize-26375.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb&replicaSet=rs0&ssl=false Implicit session: session { "id" : UUID("a062f14f-a9ae-4f5d-b7cf-91695ae5cee3") } Percona Server for MongoDB server version: v8.0.17-6 WARNING: shell and server versions do not match Successfully added user: { "user" : "myApp", "roles" : [ { "db" : "myApp", "role" : "readWrite" } ] } bye + cat /tmp/tmp.I5ZopP6aMH + rm /tmp/tmp.CfyIAjBbqp /tmp/tmp.I5ZopP6aMH + return 0 + run_mongo 'use myApp\n db.test.insert({ x: 100500 })' myApp:myPass@some-name-rs0.pvc-resize-26375 + local 'command=use myApp\n db.test.insert({ x: 100500 })' + local uri=myApp:myPass@some-name-rs0.pvc-resize-26375 + local driver=mongodb+srv + local suffix=.svc.cluster.local + local mongo_flag= + local replica_set=rs0 + [[ myApp:myPass@some-name-rs0.pvc-resize-26375 == *cfg* ]] ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8mvFKsmKLO +++ mktemp ++ local LAST_ERR=/tmp/tmp.rz8tBhVviN ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.8mvFKsmKLO ++ cat /tmp/tmp.rz8tBhVviN ++ rm /tmp/tmp.8mvFKsmKLO /tmp/tmp.rz8tBhVviN ++ return 0 + local client_container=psmdb-client-696897d69b-mv54g + kubectl_bin exec psmdb-client-696897d69b-mv54g -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb+srv://myApp:myPass@some-name-rs0.pvc-resize-26375.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.stvlnxDJ6Z ++ mktemp + local LAST_ERR=/tmp/tmp.I2gN5FWoZi + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-696897d69b-mv54g -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb+srv://myApp:myPass@some-name-rs0.pvc-resize-26375.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.stvlnxDJ6Z Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-rs0-1.some-name-rs0.pvc-resize-26375.svc.cluster.local:27017,some-name-rs0-2.some-name-rs0.pvc-resize-26375.svc.cluster.local:27017,some-name-rs0-0.some-name-rs0.pvc-resize-26375.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb&replicaSet=rs0&ssl=false Implicit session: session { "id" : UUID("73ed9f69-57e8-4277-8de5-b5877bea55b4") } Percona Server for MongoDB server version: v8.0.17-6 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.I2gN5FWoZi + rm /tmp/tmp.stvlnxDJ6Z /tmp/tmp.I2gN5FWoZi + return 0 + patch_pvc_request some-name 2G + local cluster=some-name + local size=2G + echo 'Patching PVC request to 2G in some-name' Patching PVC request to 2G in some-name + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"2G"}]' ++ mktemp + local LAST_OUT=/tmp/tmp.p9t8lBBCxt ++ mktemp + local LAST_ERR=/tmp/tmp.YdLo8iguSG + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"2G"}]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.p9t8lBBCxt perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.YdLo8iguSG + rm /tmp/tmp.p9t8lBBCxt /tmp/tmp.YdLo8iguSG + return 0 + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fwEaBYEgF5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.aVbwQVjA5w ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.fwEaBYEgF5 ++ cat /tmp/tmp.aVbwQVjA5w ++ rm /tmp/tmp.fwEaBYEgF5 /tmp/tmp.aVbwQVjA5w ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo .OK .OK + echo + wait_all_pvc_resize 2Gi 120 1 + local expected_size=2Gi + local max_retry=120 + local sleep_time=1 ++ kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name +++ mktemp ++ local LAST_OUT=/tmp/tmp.uUMlIHIh5J +++ mktemp ++ local LAST_ERR=/tmp/tmp.m2uMiitPja ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc -l app.kubernetes.io/component=mongod -o name ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.uUMlIHIh5J ++ cat /tmp/tmp.m2uMiitPja ++ rm /tmp/tmp.uUMlIHIh5J /tmp/tmp.m2uMiitPja ++ return 0 + for pvc in $(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name) + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-0 2Gi 120 1 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-0 + local expected_size=2Gi + local max_retry=120 + local sleep_time=1 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jz58BX69az +++ mktemp ++ local LAST_ERR=/tmp/tmp.gDspPjtvyb ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jz58BX69az ++ cat /tmp/tmp.gDspPjtvyb ++ rm /tmp/tmp.jz58BX69az /tmp/tmp.gDspPjtvyb ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 0 -ge 120 ]] + echo -n . .+ sleep 1 + retry=1 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4Acta2MZmt +++ mktemp ++ local LAST_ERR=/tmp/tmp.E1Hg9IgbN1 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4Acta2MZmt ++ cat /tmp/tmp.E1Hg9IgbN1 ++ rm /tmp/tmp.4Acta2MZmt /tmp/tmp.E1Hg9IgbN1 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 1 -ge 120 ]] + echo -n . .+ sleep 1 + retry=2 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sj09hYdTU2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.8rIwcAnrIn ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.sj09hYdTU2 ++ cat /tmp/tmp.8rIwcAnrIn ++ rm /tmp/tmp.sj09hYdTU2 /tmp/tmp.8rIwcAnrIn ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 2 -ge 120 ]] + echo -n . .+ sleep 1 + retry=3 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uWtdQ6p2yM +++ mktemp ++ local LAST_ERR=/tmp/tmp.Qtf4E7qcYd ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.uWtdQ6p2yM ++ cat /tmp/tmp.Qtf4E7qcYd ++ rm /tmp/tmp.uWtdQ6p2yM /tmp/tmp.Qtf4E7qcYd ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 3 -ge 120 ]] + echo -n . .+ sleep 1 + retry=4 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LDDyxOdslv +++ mktemp ++ local LAST_ERR=/tmp/tmp.OfUHZCU5Sd ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.LDDyxOdslv ++ cat /tmp/tmp.OfUHZCU5Sd ++ rm /tmp/tmp.LDDyxOdslv /tmp/tmp.OfUHZCU5Sd ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 4 -ge 120 ]] + echo -n . .+ sleep 1 + retry=5 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BMhkme6Gtn +++ mktemp ++ local LAST_ERR=/tmp/tmp.TNHBSX5z1m ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.BMhkme6Gtn ++ cat /tmp/tmp.TNHBSX5z1m ++ rm /tmp/tmp.BMhkme6Gtn /tmp/tmp.TNHBSX5z1m ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 5 -ge 120 ]] + echo -n . .+ sleep 1 + retry=6 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.c8wm9sbgiI +++ mktemp ++ local LAST_ERR=/tmp/tmp.HK961815mj ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.c8wm9sbgiI ++ cat /tmp/tmp.HK961815mj ++ rm /tmp/tmp.c8wm9sbgiI /tmp/tmp.HK961815mj ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 6 -ge 120 ]] + echo -n . .+ sleep 1 + retry=7 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7yk8QX4TH9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.dyBLiISEKa ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.7yk8QX4TH9 ++ cat /tmp/tmp.dyBLiISEKa ++ rm /tmp/tmp.7yk8QX4TH9 /tmp/tmp.dyBLiISEKa ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 7 -ge 120 ]] + echo -n . .+ sleep 1 + retry=8 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0AnbVmBi5N +++ mktemp ++ local LAST_ERR=/tmp/tmp.WwTf18TxXI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.0AnbVmBi5N ++ cat /tmp/tmp.WwTf18TxXI ++ rm /tmp/tmp.0AnbVmBi5N /tmp/tmp.WwTf18TxXI ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 8 -ge 120 ]] + echo -n . .+ sleep 1 + retry=9 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yu9l6vYADy +++ mktemp ++ local LAST_ERR=/tmp/tmp.Sjew0v9wlk ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.yu9l6vYADy ++ cat /tmp/tmp.Sjew0v9wlk ++ rm /tmp/tmp.yu9l6vYADy /tmp/tmp.Sjew0v9wlk ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 9 -ge 120 ]] + echo -n . .+ sleep 1 + retry=10 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.q8HmAO34p4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.41mTJgrfnZ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.q8HmAO34p4 ++ cat /tmp/tmp.41mTJgrfnZ ++ rm /tmp/tmp.q8HmAO34p4 /tmp/tmp.41mTJgrfnZ ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 10 -ge 120 ]] + echo -n . .+ sleep 1 + retry=11 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rqfvilhVfk +++ mktemp ++ local LAST_ERR=/tmp/tmp.yRRkIVEZ7k ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.rqfvilhVfk ++ cat /tmp/tmp.yRRkIVEZ7k ++ rm /tmp/tmp.rqfvilhVfk /tmp/tmp.yRRkIVEZ7k ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 11 -ge 120 ]] + echo -n . .+ sleep 1 + retry=12 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jpo1ueltUU +++ mktemp ++ local LAST_ERR=/tmp/tmp.suQ4Q0XXOE ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jpo1ueltUU ++ cat /tmp/tmp.suQ4Q0XXOE ++ rm /tmp/tmp.jpo1ueltUU /tmp/tmp.suQ4Q0XXOE ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 12 -ge 120 ]] + echo -n . .+ sleep 1 + retry=13 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.aChuNJbkz8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.0Zn7a1Kc0v ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.aChuNJbkz8 ++ cat /tmp/tmp.0Zn7a1Kc0v ++ rm /tmp/tmp.aChuNJbkz8 /tmp/tmp.0Zn7a1Kc0v ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 13 -ge 120 ]] + echo -n . .+ sleep 1 + retry=14 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tB0KJRArxY +++ mktemp ++ local LAST_ERR=/tmp/tmp.ewXvuitIuZ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.tB0KJRArxY ++ cat /tmp/tmp.ewXvuitIuZ ++ rm /tmp/tmp.tB0KJRArxY /tmp/tmp.ewXvuitIuZ ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 14 -ge 120 ]] + echo -n . .+ sleep 1 + retry=15 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.f6Klf9pnFz +++ mktemp ++ local LAST_ERR=/tmp/tmp.PkWm47Ojjs ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.f6Klf9pnFz ++ cat /tmp/tmp.PkWm47Ojjs ++ rm /tmp/tmp.f6Klf9pnFz /tmp/tmp.PkWm47Ojjs ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 15 -ge 120 ]] + echo -n . .+ sleep 1 + retry=16 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BofiTUcC64 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Tv1O1xFI3d ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.BofiTUcC64 ++ cat /tmp/tmp.Tv1O1xFI3d ++ rm /tmp/tmp.BofiTUcC64 /tmp/tmp.Tv1O1xFI3d ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 16 -ge 120 ]] + echo -n . .+ sleep 1 + retry=17 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rIjGNosyLT +++ mktemp ++ local LAST_ERR=/tmp/tmp.QwPjUAZWwA ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.rIjGNosyLT ++ cat /tmp/tmp.QwPjUAZWwA ++ rm /tmp/tmp.rIjGNosyLT /tmp/tmp.QwPjUAZWwA ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 17 -ge 120 ]] + echo -n . .+ sleep 1 + retry=18 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.H7HIUlKhLV +++ mktemp ++ local LAST_ERR=/tmp/tmp.RMPUapuSob ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.H7HIUlKhLV ++ cat /tmp/tmp.RMPUapuSob ++ rm /tmp/tmp.H7HIUlKhLV /tmp/tmp.RMPUapuSob ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 18 -ge 120 ]] + echo -n . .+ sleep 1 + retry=19 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oM7OREQ0II +++ mktemp ++ local LAST_ERR=/tmp/tmp.Tm6Qy6GNFt ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.oM7OREQ0II ++ cat /tmp/tmp.Tm6Qy6GNFt ++ rm /tmp/tmp.oM7OREQ0II /tmp/tmp.Tm6Qy6GNFt ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 19 -ge 120 ]] + echo -n . .+ sleep 1 + retry=20 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lH0nYSsJX3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.eaHDJkE82Z ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.lH0nYSsJX3 ++ cat /tmp/tmp.eaHDJkE82Z ++ rm /tmp/tmp.lH0nYSsJX3 /tmp/tmp.eaHDJkE82Z ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 20 -ge 120 ]] + echo -n . .+ sleep 1 + retry=21 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rXTkSd8cUf +++ mktemp ++ local LAST_ERR=/tmp/tmp.1r08NYkDpI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.rXTkSd8cUf ++ cat /tmp/tmp.1r08NYkDpI ++ rm /tmp/tmp.rXTkSd8cUf /tmp/tmp.1r08NYkDpI ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 21 -ge 120 ]] + echo -n . .+ sleep 1 + retry=22 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.CSeomcD6ds +++ mktemp ++ local LAST_ERR=/tmp/tmp.sOviowIsTa ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.CSeomcD6ds ++ cat /tmp/tmp.sOviowIsTa ++ rm /tmp/tmp.CSeomcD6ds /tmp/tmp.sOviowIsTa ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 22 -ge 120 ]] + echo -n . .+ sleep 1 + retry=23 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vFEGxTpsCf +++ mktemp ++ local LAST_ERR=/tmp/tmp.0vKzsFRuXK ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.vFEGxTpsCf ++ cat /tmp/tmp.0vKzsFRuXK ++ rm /tmp/tmp.vFEGxTpsCf /tmp/tmp.0vKzsFRuXK ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 23 -ge 120 ]] + echo -n . .+ sleep 1 + retry=24 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GgSzkDTG7s +++ mktemp ++ local LAST_ERR=/tmp/tmp.ak5zufO3gL ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.GgSzkDTG7s ++ cat /tmp/tmp.ak5zufO3gL ++ rm /tmp/tmp.GgSzkDTG7s /tmp/tmp.ak5zufO3gL ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 24 -ge 120 ]] + echo -n . .+ sleep 1 + retry=25 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zhmcUalrLm +++ mktemp ++ local LAST_ERR=/tmp/tmp.u3aazsqUGZ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.zhmcUalrLm ++ cat /tmp/tmp.u3aazsqUGZ ++ rm /tmp/tmp.zhmcUalrLm /tmp/tmp.u3aazsqUGZ ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 25 -ge 120 ]] + echo -n . .+ sleep 1 + retry=26 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ec30UeCWqR +++ mktemp ++ local LAST_ERR=/tmp/tmp.tAO6lIRM2H ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ec30UeCWqR ++ cat /tmp/tmp.tAO6lIRM2H ++ rm /tmp/tmp.ec30UeCWqR /tmp/tmp.tAO6lIRM2H ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 26 -ge 120 ]] + echo -n . .+ sleep 1 + retry=27 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Wtv31Do5oy +++ mktemp ++ local LAST_ERR=/tmp/tmp.Y5Sevl931d ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Wtv31Do5oy ++ cat /tmp/tmp.Y5Sevl931d ++ rm /tmp/tmp.Wtv31Do5oy /tmp/tmp.Y5Sevl931d ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 27 -ge 120 ]] + echo -n . .+ sleep 1 + retry=28 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.POnSMGUNG7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.gwfe7OmzdE ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.POnSMGUNG7 ++ cat /tmp/tmp.gwfe7OmzdE ++ rm /tmp/tmp.POnSMGUNG7 /tmp/tmp.gwfe7OmzdE ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 28 -ge 120 ]] + echo -n . .+ sleep 1 + retry=29 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zpPjjY5q89 +++ mktemp ++ local LAST_ERR=/tmp/tmp.WCYry6pT0C ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.zpPjjY5q89 ++ cat /tmp/tmp.WCYry6pT0C ++ rm /tmp/tmp.zpPjjY5q89 /tmp/tmp.WCYry6pT0C ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 29 -ge 120 ]] + echo -n . .+ sleep 1 + retry=30 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hIhtNwRZ80 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Bd81oX5bNa ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.hIhtNwRZ80 ++ cat /tmp/tmp.Bd81oX5bNa ++ rm /tmp/tmp.hIhtNwRZ80 /tmp/tmp.Bd81oX5bNa ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 30 -ge 120 ]] + echo -n . .+ sleep 1 + retry=31 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LhwYd6ihdO +++ mktemp ++ local LAST_ERR=/tmp/tmp.hFha8HuX2c ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.LhwYd6ihdO ++ cat /tmp/tmp.hFha8HuX2c ++ rm /tmp/tmp.LhwYd6ihdO /tmp/tmp.hFha8HuX2c ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 31 -ge 120 ]] + echo -n . .+ sleep 1 + retry=32 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KdRxT4RKiO +++ mktemp ++ local LAST_ERR=/tmp/tmp.D9gMjIJkOm ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.KdRxT4RKiO ++ cat /tmp/tmp.D9gMjIJkOm ++ rm /tmp/tmp.KdRxT4RKiO /tmp/tmp.D9gMjIJkOm ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 32 -ge 120 ]] + echo -n . .+ sleep 1 + retry=33 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.R1o9WGQ3vB +++ mktemp ++ local LAST_ERR=/tmp/tmp.I7axoS8EFA ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.R1o9WGQ3vB ++ cat /tmp/tmp.I7axoS8EFA ++ rm /tmp/tmp.R1o9WGQ3vB /tmp/tmp.I7axoS8EFA ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 33 -ge 120 ]] + echo -n . .+ sleep 1 + retry=34 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BiRQQGt2Na +++ mktemp ++ local LAST_ERR=/tmp/tmp.b77lPIqQGP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.BiRQQGt2Na ++ cat /tmp/tmp.b77lPIqQGP ++ rm /tmp/tmp.BiRQQGt2Na /tmp/tmp.b77lPIqQGP ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 34 -ge 120 ]] + echo -n . .+ sleep 1 + retry=35 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4giEP1Kjko +++ mktemp ++ local LAST_ERR=/tmp/tmp.lTxjCVT9iE ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4giEP1Kjko ++ cat /tmp/tmp.lTxjCVT9iE ++ rm /tmp/tmp.4giEP1Kjko /tmp/tmp.lTxjCVT9iE ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 35 -ge 120 ]] + echo -n . .+ sleep 1 + retry=36 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wPpb0lpI3w +++ mktemp ++ local LAST_ERR=/tmp/tmp.HO0MmVg2xy ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.wPpb0lpI3w ++ cat /tmp/tmp.HO0MmVg2xy ++ rm /tmp/tmp.wPpb0lpI3w /tmp/tmp.HO0MmVg2xy ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 36 -ge 120 ]] + echo -n . .+ sleep 1 + retry=37 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IoPs6pOs8T +++ mktemp ++ local LAST_ERR=/tmp/tmp.SjxLfdHvLr ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.IoPs6pOs8T ++ cat /tmp/tmp.SjxLfdHvLr ++ rm /tmp/tmp.IoPs6pOs8T /tmp/tmp.SjxLfdHvLr ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 37 -ge 120 ]] + echo -n . .+ sleep 1 + retry=38 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4UqH3kHNf4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.0D0duRxPiW ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4UqH3kHNf4 ++ cat /tmp/tmp.0D0duRxPiW ++ rm /tmp/tmp.4UqH3kHNf4 /tmp/tmp.0D0duRxPiW ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 38 -ge 120 ]] + echo -n . .+ sleep 1 + retry=39 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ekJRnD1Q40 +++ mktemp ++ local LAST_ERR=/tmp/tmp.2O5TjrHtYn ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ekJRnD1Q40 ++ cat /tmp/tmp.2O5TjrHtYn ++ rm /tmp/tmp.ekJRnD1Q40 /tmp/tmp.2O5TjrHtYn ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 39 -ge 120 ]] + echo -n . .+ sleep 1 + retry=40 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tjZ8IcVG9W +++ mktemp ++ local LAST_ERR=/tmp/tmp.XRNJ70VaHF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.tjZ8IcVG9W ++ cat /tmp/tmp.XRNJ70VaHF ++ rm /tmp/tmp.tjZ8IcVG9W /tmp/tmp.XRNJ70VaHF ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 40 -ge 120 ]] + echo -n . .+ sleep 1 + retry=41 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XgyrTFMXjm +++ mktemp ++ local LAST_ERR=/tmp/tmp.rLqo8itsi8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.XgyrTFMXjm ++ cat /tmp/tmp.rLqo8itsi8 ++ rm /tmp/tmp.XgyrTFMXjm /tmp/tmp.rLqo8itsi8 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 41 -ge 120 ]] + echo -n . .+ sleep 1 + retry=42 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0ohdBaPHsH +++ mktemp ++ local LAST_ERR=/tmp/tmp.BLiSROwG5a ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.0ohdBaPHsH ++ cat /tmp/tmp.BLiSROwG5a ++ rm /tmp/tmp.0ohdBaPHsH /tmp/tmp.BLiSROwG5a ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 42 -ge 120 ]] + echo -n . .+ sleep 1 + retry=43 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zLbWooqpUM +++ mktemp ++ local LAST_ERR=/tmp/tmp.qExlpuLk7H ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.zLbWooqpUM ++ cat /tmp/tmp.qExlpuLk7H ++ rm /tmp/tmp.zLbWooqpUM /tmp/tmp.qExlpuLk7H ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 43 -ge 120 ]] + echo -n . .+ sleep 1 + retry=44 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XhacS9EFXo +++ mktemp ++ local LAST_ERR=/tmp/tmp.H5yn5Zctof ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.XhacS9EFXo ++ cat /tmp/tmp.H5yn5Zctof ++ rm /tmp/tmp.XhacS9EFXo /tmp/tmp.H5yn5Zctof ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 44 -ge 120 ]] + echo -n . .+ sleep 1 + retry=45 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gbRS5HArml +++ mktemp ++ local LAST_ERR=/tmp/tmp.fSH736czW0 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.gbRS5HArml ++ cat /tmp/tmp.fSH736czW0 ++ rm /tmp/tmp.gbRS5HArml /tmp/tmp.fSH736czW0 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 45 -ge 120 ]] + echo -n . .+ sleep 1 + retry=46 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sirAyJOa9E +++ mktemp ++ local LAST_ERR=/tmp/tmp.U1hfGgVdnS ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.sirAyJOa9E ++ cat /tmp/tmp.U1hfGgVdnS ++ rm /tmp/tmp.sirAyJOa9E /tmp/tmp.U1hfGgVdnS ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 46 -ge 120 ]] + echo -n . .+ sleep 1 + retry=47 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YOY8knt7Yz +++ mktemp ++ local LAST_ERR=/tmp/tmp.tG7VZYHSiv ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.YOY8knt7Yz ++ cat /tmp/tmp.tG7VZYHSiv ++ rm /tmp/tmp.YOY8knt7Yz /tmp/tmp.tG7VZYHSiv ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 47 -ge 120 ]] + echo -n . .+ sleep 1 + retry=48 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bL2NnD7BSS +++ mktemp ++ local LAST_ERR=/tmp/tmp.oDvz7DvGvc ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.bL2NnD7BSS ++ cat /tmp/tmp.oDvz7DvGvc ++ rm /tmp/tmp.bL2NnD7BSS /tmp/tmp.oDvz7DvGvc ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 48 -ge 120 ]] + echo -n . .+ sleep 1 + retry=49 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TLrSytJb5u +++ mktemp ++ local LAST_ERR=/tmp/tmp.wtspFt1M7c ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.TLrSytJb5u ++ cat /tmp/tmp.wtspFt1M7c ++ rm /tmp/tmp.TLrSytJb5u /tmp/tmp.wtspFt1M7c ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 49 -ge 120 ]] + echo -n . .+ sleep 1 + retry=50 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ATTLAwi2Zt +++ mktemp ++ local LAST_ERR=/tmp/tmp.b2l2kefPIZ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ATTLAwi2Zt ++ cat /tmp/tmp.b2l2kefPIZ ++ rm /tmp/tmp.ATTLAwi2Zt /tmp/tmp.b2l2kefPIZ ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 50 -ge 120 ]] + echo -n . .+ sleep 1 + retry=51 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jYyTDz7b1g +++ mktemp ++ local LAST_ERR=/tmp/tmp.ByX5aA9i5X ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jYyTDz7b1g ++ cat /tmp/tmp.ByX5aA9i5X ++ rm /tmp/tmp.jYyTDz7b1g /tmp/tmp.ByX5aA9i5X ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 51 -ge 120 ]] + echo -n . .+ sleep 1 + retry=52 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xVKZlz08Km +++ mktemp ++ local LAST_ERR=/tmp/tmp.stxtBVFlrr ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xVKZlz08Km ++ cat /tmp/tmp.stxtBVFlrr ++ rm /tmp/tmp.xVKZlz08Km /tmp/tmp.stxtBVFlrr ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 52 -ge 120 ]] + echo -n . .+ sleep 1 + retry=53 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sDjw5BF3Jv +++ mktemp ++ local LAST_ERR=/tmp/tmp.L05InRnC1b ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.sDjw5BF3Jv ++ cat /tmp/tmp.L05InRnC1b ++ rm /tmp/tmp.sDjw5BF3Jv /tmp/tmp.L05InRnC1b ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 53 -ge 120 ]] + echo -n . .+ sleep 1 + retry=54 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KADXwmc3TH +++ mktemp ++ local LAST_ERR=/tmp/tmp.sLdpAZ1g0F ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.KADXwmc3TH ++ cat /tmp/tmp.sLdpAZ1g0F ++ rm /tmp/tmp.KADXwmc3TH /tmp/tmp.sLdpAZ1g0F ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 54 -ge 120 ]] + echo -n . .+ sleep 1 + retry=55 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qFkJdaRrkA +++ mktemp ++ local LAST_ERR=/tmp/tmp.fR5aVR6SyC ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.qFkJdaRrkA ++ cat /tmp/tmp.fR5aVR6SyC ++ rm /tmp/tmp.qFkJdaRrkA /tmp/tmp.fR5aVR6SyC ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 55 -ge 120 ]] + echo -n . .+ sleep 1 + retry=56 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Lj41KWxH95 +++ mktemp ++ local LAST_ERR=/tmp/tmp.CZEwvwFrPO ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Lj41KWxH95 ++ cat /tmp/tmp.CZEwvwFrPO ++ rm /tmp/tmp.Lj41KWxH95 /tmp/tmp.CZEwvwFrPO ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 56 -ge 120 ]] + echo -n . .+ sleep 1 + retry=57 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HdOlibLcqh +++ mktemp ++ local LAST_ERR=/tmp/tmp.x6eADqiB3f ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.HdOlibLcqh ++ cat /tmp/tmp.x6eADqiB3f ++ rm /tmp/tmp.HdOlibLcqh /tmp/tmp.x6eADqiB3f ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 57 -ge 120 ]] + echo -n . .+ sleep 1 + retry=58 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KwxM6FDds2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.1DCRq1yJHY ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.KwxM6FDds2 ++ cat /tmp/tmp.1DCRq1yJHY ++ rm /tmp/tmp.KwxM6FDds2 /tmp/tmp.1DCRq1yJHY ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 58 -ge 120 ]] + echo -n . .+ sleep 1 + retry=59 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9lGz9pPoiM +++ mktemp ++ local LAST_ERR=/tmp/tmp.PhLm24jn6V ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9lGz9pPoiM ++ cat /tmp/tmp.PhLm24jn6V ++ rm /tmp/tmp.9lGz9pPoiM /tmp/tmp.PhLm24jn6V ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 59 -ge 120 ]] + echo -n . .+ sleep 1 + retry=60 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SKEzj0o3EC +++ mktemp ++ local LAST_ERR=/tmp/tmp.WdZUMPpagl ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.SKEzj0o3EC ++ cat /tmp/tmp.WdZUMPpagl ++ rm /tmp/tmp.SKEzj0o3EC /tmp/tmp.WdZUMPpagl ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 60 -ge 120 ]] + echo -n . .+ sleep 1 + retry=61 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Em1o9og5Ce +++ mktemp ++ local LAST_ERR=/tmp/tmp.8UBY20Ak9c ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Em1o9og5Ce ++ cat /tmp/tmp.8UBY20Ak9c ++ rm /tmp/tmp.Em1o9og5Ce /tmp/tmp.8UBY20Ak9c ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 61 -ge 120 ]] + echo -n . .+ sleep 1 + retry=62 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OP0T5kg3Pu +++ mktemp ++ local LAST_ERR=/tmp/tmp.uxm5huQYVK ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.OP0T5kg3Pu ++ cat /tmp/tmp.uxm5huQYVK ++ rm /tmp/tmp.OP0T5kg3Pu /tmp/tmp.uxm5huQYVK ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 62 -ge 120 ]] + echo -n . .+ sleep 1 + retry=63 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kk5OOTyMTq +++ mktemp ++ local LAST_ERR=/tmp/tmp.CUwXi27ipK ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.kk5OOTyMTq ++ cat /tmp/tmp.CUwXi27ipK ++ rm /tmp/tmp.kk5OOTyMTq /tmp/tmp.CUwXi27ipK ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 63 -ge 120 ]] + echo -n . .+ sleep 1 + retry=64 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zixn4E1yl7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.nAmS3G7COf ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.zixn4E1yl7 ++ cat /tmp/tmp.nAmS3G7COf ++ rm /tmp/tmp.zixn4E1yl7 /tmp/tmp.nAmS3G7COf ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 64 -ge 120 ]] + echo -n . .+ sleep 1 + retry=65 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vBDkIMTgsd +++ mktemp ++ local LAST_ERR=/tmp/tmp.sPY1QrxDni ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.vBDkIMTgsd ++ cat /tmp/tmp.sPY1QrxDni ++ rm /tmp/tmp.vBDkIMTgsd /tmp/tmp.sPY1QrxDni ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 65 -ge 120 ]] + echo -n . .+ sleep 1 + retry=66 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tp1lP18OBz +++ mktemp ++ local LAST_ERR=/tmp/tmp.qgW7SKcAzF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.tp1lP18OBz ++ cat /tmp/tmp.qgW7SKcAzF ++ rm /tmp/tmp.tp1lP18OBz /tmp/tmp.qgW7SKcAzF ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 66 -ge 120 ]] + echo -n . .+ sleep 1 + retry=67 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Pp0MoBUaVV +++ mktemp ++ local LAST_ERR=/tmp/tmp.DYZ4sBIUlW ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Pp0MoBUaVV ++ cat /tmp/tmp.DYZ4sBIUlW ++ rm /tmp/tmp.Pp0MoBUaVV /tmp/tmp.DYZ4sBIUlW ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 67 -ge 120 ]] + echo -n . .+ sleep 1 + retry=68 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BUy4gHIbu4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.qjBpxaMblD ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.BUy4gHIbu4 ++ cat /tmp/tmp.qjBpxaMblD ++ rm /tmp/tmp.BUy4gHIbu4 /tmp/tmp.qjBpxaMblD ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 68 -ge 120 ]] + echo -n . .+ sleep 1 + retry=69 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FgnJO4wGRM +++ mktemp ++ local LAST_ERR=/tmp/tmp.qiabPtnYFh ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.FgnJO4wGRM ++ cat /tmp/tmp.qiabPtnYFh ++ rm /tmp/tmp.FgnJO4wGRM /tmp/tmp.qiabPtnYFh ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 69 -ge 120 ]] + echo -n . .+ sleep 1 + retry=70 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gaKQeS9D5U +++ mktemp ++ local LAST_ERR=/tmp/tmp.7XlZwkcNqj ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.gaKQeS9D5U ++ cat /tmp/tmp.7XlZwkcNqj ++ rm /tmp/tmp.gaKQeS9D5U /tmp/tmp.7XlZwkcNqj ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 70 -ge 120 ]] + echo -n . .+ sleep 1 + retry=71 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kErk09LbiD +++ mktemp ++ local LAST_ERR=/tmp/tmp.qjbwvDfhKm ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.kErk09LbiD ++ cat /tmp/tmp.qjbwvDfhKm ++ rm /tmp/tmp.kErk09LbiD /tmp/tmp.qjbwvDfhKm ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 71 -ge 120 ]] + echo -n . .+ sleep 1 + retry=72 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Pr67MnZoMt +++ mktemp ++ local LAST_ERR=/tmp/tmp.88AjDIMz55 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Pr67MnZoMt ++ cat /tmp/tmp.88AjDIMz55 ++ rm /tmp/tmp.Pr67MnZoMt /tmp/tmp.88AjDIMz55 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 72 -ge 120 ]] + echo -n . .+ sleep 1 + retry=73 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Bk8Y9HRYlr +++ mktemp ++ local LAST_ERR=/tmp/tmp.shLDV7ueqg ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Bk8Y9HRYlr ++ cat /tmp/tmp.shLDV7ueqg ++ rm /tmp/tmp.Bk8Y9HRYlr /tmp/tmp.shLDV7ueqg ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 73 -ge 120 ]] + echo -n . .+ sleep 1 + retry=74 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4mkll6s3dp +++ mktemp ++ local LAST_ERR=/tmp/tmp.lciVY4xoYX ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4mkll6s3dp ++ cat /tmp/tmp.lciVY4xoYX ++ rm /tmp/tmp.4mkll6s3dp /tmp/tmp.lciVY4xoYX ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 74 -ge 120 ]] + echo -n . .+ sleep 1 + retry=75 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Y4mUphDmKb +++ mktemp ++ local LAST_ERR=/tmp/tmp.kvdQiV1tOS ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Y4mUphDmKb ++ cat /tmp/tmp.kvdQiV1tOS ++ rm /tmp/tmp.Y4mUphDmKb /tmp/tmp.kvdQiV1tOS ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 75 -ge 120 ]] + echo -n . .+ sleep 1 + retry=76 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ue6sG3umcv +++ mktemp ++ local LAST_ERR=/tmp/tmp.cfukV8d5eP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ue6sG3umcv ++ cat /tmp/tmp.cfukV8d5eP ++ rm /tmp/tmp.ue6sG3umcv /tmp/tmp.cfukV8d5eP ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 76 -ge 120 ]] + echo -n . .+ sleep 1 + retry=77 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TpGEMs9lVh +++ mktemp ++ local LAST_ERR=/tmp/tmp.udkN4TIPsY ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.TpGEMs9lVh ++ cat /tmp/tmp.udkN4TIPsY ++ rm /tmp/tmp.TpGEMs9lVh /tmp/tmp.udkN4TIPsY ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 77 -ge 120 ]] + echo -n . .+ sleep 1 + retry=78 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.H2IpF1H2on +++ mktemp ++ local LAST_ERR=/tmp/tmp.8NlOXDP1aH ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.H2IpF1H2on ++ cat /tmp/tmp.8NlOXDP1aH ++ rm /tmp/tmp.H2IpF1H2on /tmp/tmp.8NlOXDP1aH ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 78 -ge 120 ]] + echo -n . .+ sleep 1 + retry=79 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UuMcX9NiPp +++ mktemp ++ local LAST_ERR=/tmp/tmp.wmpzuLooDQ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.UuMcX9NiPp ++ cat /tmp/tmp.wmpzuLooDQ ++ rm /tmp/tmp.UuMcX9NiPp /tmp/tmp.wmpzuLooDQ ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 79 -ge 120 ]] + echo -n . .+ sleep 1 + retry=80 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cRxwB0gtp9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.9vKlZR4OuL ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.cRxwB0gtp9 ++ cat /tmp/tmp.9vKlZR4OuL ++ rm /tmp/tmp.cRxwB0gtp9 /tmp/tmp.9vKlZR4OuL ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 80 -ge 120 ]] + echo -n . .+ sleep 1 + retry=81 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.05rKQGRmHk +++ mktemp ++ local LAST_ERR=/tmp/tmp.BIeYjIhFvt ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.05rKQGRmHk ++ cat /tmp/tmp.BIeYjIhFvt ++ rm /tmp/tmp.05rKQGRmHk /tmp/tmp.BIeYjIhFvt ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 81 -ge 120 ]] + echo -n . .+ sleep 1 + retry=82 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.atv5uKdiWE +++ mktemp ++ local LAST_ERR=/tmp/tmp.2r0Fu3CMFJ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.atv5uKdiWE ++ cat /tmp/tmp.2r0Fu3CMFJ ++ rm /tmp/tmp.atv5uKdiWE /tmp/tmp.2r0Fu3CMFJ ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 82 -ge 120 ]] + echo -n . .+ sleep 1 + retry=83 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Jb6cleae3I +++ mktemp ++ local LAST_ERR=/tmp/tmp.KdjfsIi1Vp ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Jb6cleae3I ++ cat /tmp/tmp.KdjfsIi1Vp ++ rm /tmp/tmp.Jb6cleae3I /tmp/tmp.KdjfsIi1Vp ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 83 -ge 120 ]] + echo -n . .+ sleep 1 + retry=84 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.foRboaRWsL +++ mktemp ++ local LAST_ERR=/tmp/tmp.VqxwshMejo ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.foRboaRWsL ++ cat /tmp/tmp.VqxwshMejo ++ rm /tmp/tmp.foRboaRWsL /tmp/tmp.VqxwshMejo ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 84 -ge 120 ]] + echo -n . .+ sleep 1 + retry=85 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WqEKevHI80 +++ mktemp ++ local LAST_ERR=/tmp/tmp.XDCOnYaLfB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.WqEKevHI80 ++ cat /tmp/tmp.XDCOnYaLfB ++ rm /tmp/tmp.WqEKevHI80 /tmp/tmp.XDCOnYaLfB ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 85 -ge 120 ]] + echo -n . .+ sleep 1 + retry=86 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MBSkxEuGHG +++ mktemp ++ local LAST_ERR=/tmp/tmp.bQy9Z2zZMg ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.MBSkxEuGHG ++ cat /tmp/tmp.bQy9Z2zZMg ++ rm /tmp/tmp.MBSkxEuGHG /tmp/tmp.bQy9Z2zZMg ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 86 -ge 120 ]] + echo -n . .+ sleep 1 + retry=87 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lyFPwB7Eqz +++ mktemp ++ local LAST_ERR=/tmp/tmp.yoGaJdFcgj ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.lyFPwB7Eqz ++ cat /tmp/tmp.yoGaJdFcgj ++ rm /tmp/tmp.lyFPwB7Eqz /tmp/tmp.yoGaJdFcgj ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 87 -ge 120 ]] + echo -n . .+ sleep 1 + retry=88 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pmqKvRednW +++ mktemp ++ local LAST_ERR=/tmp/tmp.hpJBTX4SP6 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.pmqKvRednW ++ cat /tmp/tmp.hpJBTX4SP6 ++ rm /tmp/tmp.pmqKvRednW /tmp/tmp.hpJBTX4SP6 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 88 -ge 120 ]] + echo -n . .+ sleep 1 + retry=89 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nB8319tASU +++ mktemp ++ local LAST_ERR=/tmp/tmp.J1GA5CY2SJ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.nB8319tASU ++ cat /tmp/tmp.J1GA5CY2SJ ++ rm /tmp/tmp.nB8319tASU /tmp/tmp.J1GA5CY2SJ ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 89 -ge 120 ]] + echo -n . .+ sleep 1 + retry=90 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XPPCqoy7MA +++ mktemp ++ local LAST_ERR=/tmp/tmp.yyPYYn4KmS ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.XPPCqoy7MA ++ cat /tmp/tmp.yyPYYn4KmS ++ rm /tmp/tmp.XPPCqoy7MA /tmp/tmp.yyPYYn4KmS ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 90 -ge 120 ]] + echo -n . .+ sleep 1 + retry=91 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ILiCslDbJ9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.82cjhTpBzQ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ILiCslDbJ9 ++ cat /tmp/tmp.82cjhTpBzQ ++ rm /tmp/tmp.ILiCslDbJ9 /tmp/tmp.82cjhTpBzQ ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 91 -ge 120 ]] + echo -n . .+ sleep 1 + retry=92 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mMwuRvivxG +++ mktemp ++ local LAST_ERR=/tmp/tmp.8VNRq8VnYJ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.mMwuRvivxG ++ cat /tmp/tmp.8VNRq8VnYJ ++ rm /tmp/tmp.mMwuRvivxG /tmp/tmp.8VNRq8VnYJ ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 92 -ge 120 ]] + echo -n . .+ sleep 1 + retry=93 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wTIFXzHurs +++ mktemp ++ local LAST_ERR=/tmp/tmp.bX9Uzcz0CP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.wTIFXzHurs ++ cat /tmp/tmp.bX9Uzcz0CP ++ rm /tmp/tmp.wTIFXzHurs /tmp/tmp.bX9Uzcz0CP ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 93 -ge 120 ]] + echo -n . .+ sleep 1 + retry=94 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.o7rzDghlp5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.1rTfeYSewc ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.o7rzDghlp5 ++ cat /tmp/tmp.1rTfeYSewc ++ rm /tmp/tmp.o7rzDghlp5 /tmp/tmp.1rTfeYSewc ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 94 -ge 120 ]] + echo -n . .+ sleep 1 + retry=95 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FC43mkVgZd +++ mktemp ++ local LAST_ERR=/tmp/tmp.bJGAmTj5Ym ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.FC43mkVgZd ++ cat /tmp/tmp.bJGAmTj5Ym ++ rm /tmp/tmp.FC43mkVgZd /tmp/tmp.bJGAmTj5Ym ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 95 -ge 120 ]] + echo -n . .+ sleep 1 + retry=96 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uygwLsYhZ8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.U4tithg5ow ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.uygwLsYhZ8 ++ cat /tmp/tmp.U4tithg5ow ++ rm /tmp/tmp.uygwLsYhZ8 /tmp/tmp.U4tithg5ow ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 96 -ge 120 ]] + echo -n . .+ sleep 1 + retry=97 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.P4de0Zaj4S +++ mktemp ++ local LAST_ERR=/tmp/tmp.UOjLaWirnX ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.P4de0Zaj4S ++ cat /tmp/tmp.UOjLaWirnX ++ rm /tmp/tmp.P4de0Zaj4S /tmp/tmp.UOjLaWirnX ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 97 -ge 120 ]] + echo -n . .+ sleep 1 + retry=98 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2AaprTON3C +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZtjOZjNjRq ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.2AaprTON3C ++ cat /tmp/tmp.ZtjOZjNjRq ++ rm /tmp/tmp.2AaprTON3C /tmp/tmp.ZtjOZjNjRq ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 98 -ge 120 ]] + echo -n . .+ sleep 1 + retry=99 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.o9gLIZNc6J +++ mktemp ++ local LAST_ERR=/tmp/tmp.cGu7h7o3QJ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.o9gLIZNc6J ++ cat /tmp/tmp.cGu7h7o3QJ ++ rm /tmp/tmp.o9gLIZNc6J /tmp/tmp.cGu7h7o3QJ ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 99 -ge 120 ]] + echo -n . .+ sleep 1 + retry=100 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ikRtejoqOi +++ mktemp ++ local LAST_ERR=/tmp/tmp.EqEcODzBME ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ikRtejoqOi ++ cat /tmp/tmp.EqEcODzBME ++ rm /tmp/tmp.ikRtejoqOi /tmp/tmp.EqEcODzBME ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 100 -ge 120 ]] + echo -n . .+ sleep 1 + retry=101 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9OHHDCL7lL +++ mktemp ++ local LAST_ERR=/tmp/tmp.rhRy2Igffp ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9OHHDCL7lL ++ cat /tmp/tmp.rhRy2Igffp ++ rm /tmp/tmp.9OHHDCL7lL /tmp/tmp.rhRy2Igffp ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 101 -ge 120 ]] + echo -n . .+ sleep 1 + retry=102 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DUfKTlkT1g +++ mktemp ++ local LAST_ERR=/tmp/tmp.eYMrOJutCd ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.DUfKTlkT1g ++ cat /tmp/tmp.eYMrOJutCd ++ rm /tmp/tmp.DUfKTlkT1g /tmp/tmp.eYMrOJutCd ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 102 -ge 120 ]] + echo -n . .+ sleep 1 + retry=103 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.m4APaaVrxI +++ mktemp ++ local LAST_ERR=/tmp/tmp.Mf1NU67ski ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.m4APaaVrxI ++ cat /tmp/tmp.Mf1NU67ski ++ rm /tmp/tmp.m4APaaVrxI /tmp/tmp.Mf1NU67ski ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 103 -ge 120 ]] + echo -n . .+ sleep 1 + retry=104 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.iJr87omYmf +++ mktemp ++ local LAST_ERR=/tmp/tmp.yU8fhKXw8i ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.iJr87omYmf ++ cat /tmp/tmp.yU8fhKXw8i ++ rm /tmp/tmp.iJr87omYmf /tmp/tmp.yU8fhKXw8i ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 104 -ge 120 ]] + echo -n . .+ sleep 1 + retry=105 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VgS93z3YEO +++ mktemp ++ local LAST_ERR=/tmp/tmp.TqNGx1yTmn ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.VgS93z3YEO ++ cat /tmp/tmp.TqNGx1yTmn ++ rm /tmp/tmp.VgS93z3YEO /tmp/tmp.TqNGx1yTmn ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 105 -ge 120 ]] + echo -n . .+ sleep 1 + retry=106 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.f8rEUwouKW +++ mktemp ++ local LAST_ERR=/tmp/tmp.WH2jYTCUO8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.f8rEUwouKW ++ cat /tmp/tmp.WH2jYTCUO8 ++ rm /tmp/tmp.f8rEUwouKW /tmp/tmp.WH2jYTCUO8 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 106 -ge 120 ]] + echo -n . .+ sleep 1 + retry=107 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.H4mHWTS3dN +++ mktemp ++ local LAST_ERR=/tmp/tmp.HUWVleTzfv ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.H4mHWTS3dN ++ cat /tmp/tmp.HUWVleTzfv ++ rm /tmp/tmp.H4mHWTS3dN /tmp/tmp.HUWVleTzfv ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 107 -ge 120 ]] + echo -n . .+ sleep 1 + retry=108 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1ieNXDKHxa +++ mktemp ++ local LAST_ERR=/tmp/tmp.BgUct5K4Tg ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.1ieNXDKHxa ++ cat /tmp/tmp.BgUct5K4Tg ++ rm /tmp/tmp.1ieNXDKHxa /tmp/tmp.BgUct5K4Tg ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 108 -ge 120 ]] + echo -n . .+ sleep 1 + retry=109 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.srBMZgwsBe +++ mktemp ++ local LAST_ERR=/tmp/tmp.DseAc9uIMs ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.srBMZgwsBe ++ cat /tmp/tmp.DseAc9uIMs ++ rm /tmp/tmp.srBMZgwsBe /tmp/tmp.DseAc9uIMs ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 109 -ge 120 ]] + echo -n . .+ sleep 1 + retry=110 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jddgGRxBwp +++ mktemp ++ local LAST_ERR=/tmp/tmp.98Ei7xSxJj ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jddgGRxBwp ++ cat /tmp/tmp.98Ei7xSxJj ++ rm /tmp/tmp.jddgGRxBwp /tmp/tmp.98Ei7xSxJj ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 110 -ge 120 ]] + echo -n . .+ sleep 1 + retry=111 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KYmVUDwSFH +++ mktemp ++ local LAST_ERR=/tmp/tmp.ynDV8fQQnQ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.KYmVUDwSFH ++ cat /tmp/tmp.ynDV8fQQnQ ++ rm /tmp/tmp.KYmVUDwSFH /tmp/tmp.ynDV8fQQnQ ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 111 -ge 120 ]] + echo -n . .+ sleep 1 + retry=112 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.R4vkXYtl1I +++ mktemp ++ local LAST_ERR=/tmp/tmp.noGRW3sVNO ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.R4vkXYtl1I ++ cat /tmp/tmp.noGRW3sVNO ++ rm /tmp/tmp.R4vkXYtl1I /tmp/tmp.noGRW3sVNO ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 112 -ge 120 ]] + echo -n . .+ sleep 1 + retry=113 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.N3mw3WfF0l +++ mktemp ++ local LAST_ERR=/tmp/tmp.pL5FpVZgDy ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.N3mw3WfF0l ++ cat /tmp/tmp.pL5FpVZgDy ++ rm /tmp/tmp.N3mw3WfF0l /tmp/tmp.pL5FpVZgDy ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 113 -ge 120 ]] + echo -n . .+ sleep 1 + retry=114 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BfChKi7Xyi +++ mktemp ++ local LAST_ERR=/tmp/tmp.mxYaENO5pP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.BfChKi7Xyi ++ cat /tmp/tmp.mxYaENO5pP ++ rm /tmp/tmp.BfChKi7Xyi /tmp/tmp.mxYaENO5pP ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 114 -ge 120 ]] + echo -n . .+ sleep 1 + retry=115 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gl0Y2FvCHc +++ mktemp ++ local LAST_ERR=/tmp/tmp.XnFlZKBEjG ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.gl0Y2FvCHc ++ cat /tmp/tmp.XnFlZKBEjG ++ rm /tmp/tmp.gl0Y2FvCHc /tmp/tmp.XnFlZKBEjG ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 115 -ge 120 ]] + echo -n . .+ sleep 1 + retry=116 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fBVYr5ipff +++ mktemp ++ local LAST_ERR=/tmp/tmp.i7s2qoSiIU ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.fBVYr5ipff ++ cat /tmp/tmp.i7s2qoSiIU ++ rm /tmp/tmp.fBVYr5ipff /tmp/tmp.i7s2qoSiIU ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 116 -ge 120 ]] + echo -n . .+ sleep 1 + retry=117 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Cw39Jx2ENA +++ mktemp ++ local LAST_ERR=/tmp/tmp.E5V4TuYNPx ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Cw39Jx2ENA ++ cat /tmp/tmp.E5V4TuYNPx ++ rm /tmp/tmp.Cw39Jx2ENA /tmp/tmp.E5V4TuYNPx ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 117 -ge 120 ]] + echo -n . .+ sleep 1 + retry=118 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OGA5wi3Emv +++ mktemp ++ local LAST_ERR=/tmp/tmp.fNhAT9YyTZ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.OGA5wi3Emv ++ cat /tmp/tmp.fNhAT9YyTZ ++ rm /tmp/tmp.OGA5wi3Emv /tmp/tmp.fNhAT9YyTZ ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 118 -ge 120 ]] + echo -n . .+ sleep 1 + retry=119 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Px9d8p1hmx +++ mktemp ++ local LAST_ERR=/tmp/tmp.D0Za90ua4u ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Px9d8p1hmx ++ cat /tmp/tmp.D0Za90ua4u ++ rm /tmp/tmp.Px9d8p1hmx /tmp/tmp.D0Za90ua4u ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 119 -ge 120 ]] + echo -n . .+ sleep 1 + retry=120 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jSOWQirNwU +++ mktemp ++ local LAST_ERR=/tmp/tmp.lFVhPK2MyK ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jSOWQirNwU ++ cat /tmp/tmp.lFVhPK2MyK ++ rm /tmp/tmp.jSOWQirNwU /tmp/tmp.lFVhPK2MyK ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 120 -ge 120 ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-0 was not resized, max retries exceeded' persistentvolumeclaim/mongod-data-some-name-rs0-0 was not resized, max retries exceeded + return 1 + return 1 + echo 'Enabling PVC resize' Enabling PVC resize + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "add", "path": "/spec/storageScaling", "value":{"enableVolumeScaling":true} }]' ++ mktemp + local LAST_OUT=/tmp/tmp.CaETPp5qIo ++ mktemp + local LAST_ERR=/tmp/tmp.ZuEqnrraOn + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "add", "path": "/spec/storageScaling", "value":{"enableVolumeScaling":true} }]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.CaETPp5qIo perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.ZuEqnrraOn + rm /tmp/tmp.CaETPp5qIo /tmp/tmp.ZuEqnrraOn + return 0 + sleep 10 + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QbRgNhtcDu +++ mktemp ++ local LAST_ERR=/tmp/tmp.bzrv6GDleZ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.QbRgNhtcDu ++ cat /tmp/tmp.bzrv6GDleZ ++ rm /tmp/tmp.QbRgNhtcDu /tmp/tmp.bzrv6GDleZ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 1 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ICXRbHqWRY +++ mktemp ++ local LAST_ERR=/tmp/tmp.tWCHrV9rxy ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ICXRbHqWRY ++ cat /tmp/tmp.tWCHrV9rxy ++ rm /tmp/tmp.ICXRbHqWRY /tmp/tmp.tWCHrV9rxy ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 2 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.89aioZ7FH9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.jIkLBb8ES2 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.89aioZ7FH9 ++ cat /tmp/tmp.jIkLBb8ES2 ++ rm /tmp/tmp.89aioZ7FH9 /tmp/tmp.jIkLBb8ES2 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 3 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uq7MdBi2VC +++ mktemp ++ local LAST_ERR=/tmp/tmp.tbMXOaZy2g ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.uq7MdBi2VC ++ cat /tmp/tmp.tbMXOaZy2g ++ rm /tmp/tmp.uq7MdBi2VC /tmp/tmp.tbMXOaZy2g ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 4 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sIyvTSNun6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.mGKS7Ktx7F ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.sIyvTSNun6 ++ cat /tmp/tmp.mGKS7Ktx7F ++ rm /tmp/tmp.sIyvTSNun6 /tmp/tmp.mGKS7Ktx7F ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 5 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.D3HUB2jFaC +++ mktemp ++ local LAST_ERR=/tmp/tmp.oFLlw1ROVO ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.D3HUB2jFaC ++ cat /tmp/tmp.oFLlw1ROVO ++ rm /tmp/tmp.D3HUB2jFaC /tmp/tmp.oFLlw1ROVO ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 6 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IMl46d5ky2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.XgRV33e6g9 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.IMl46d5ky2 ++ cat /tmp/tmp.XgRV33e6g9 ++ rm /tmp/tmp.IMl46d5ky2 /tmp/tmp.XgRV33e6g9 ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo .OK .OK + wait_all_pvc_resize 2Gi + local expected_size=2Gi + local max_retry=120 + local sleep_time=5 ++ kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name +++ mktemp ++ local LAST_OUT=/tmp/tmp.oLiYdcKcZN +++ mktemp ++ local LAST_ERR=/tmp/tmp.yf17H1Irx9 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc -l app.kubernetes.io/component=mongod -o name ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.oLiYdcKcZN ++ cat /tmp/tmp.yf17H1Irx9 ++ rm /tmp/tmp.oLiYdcKcZN /tmp/tmp.yf17H1Irx9 ++ return 0 + for pvc in $(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name) + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-0 2Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-0 + local expected_size=2Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lxfj60pCOj +++ mktemp ++ local LAST_ERR=/tmp/tmp.2LiEpESJyA ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.lxfj60pCOj ++ cat /tmp/tmp.2LiEpESJyA ++ rm /tmp/tmp.lxfj60pCOj /tmp/tmp.2LiEpESJyA ++ return 0 + [[ 2Gi == \2\G\i ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized + return 0 + for pvc in $(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name) + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-1 2Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-1 + local expected_size=2Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-1 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-1 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.P2tvkM2lvs +++ mktemp ++ local LAST_ERR=/tmp/tmp.PxOf7Zm3hZ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.P2tvkM2lvs ++ cat /tmp/tmp.PxOf7Zm3hZ ++ rm /tmp/tmp.P2tvkM2lvs /tmp/tmp.PxOf7Zm3hZ ++ return 0 + [[ 2Gi == \2\G\i ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-1 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-1 was resized + return 0 + for pvc in $(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name) + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-2 2Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-2 + local expected_size=2Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-2 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-2 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1cB6Kz75xh +++ mktemp ++ local LAST_ERR=/tmp/tmp.JKtQodTJbX ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.1cB6Kz75xh ++ cat /tmp/tmp.JKtQodTJbX ++ rm /tmp/tmp.1cB6Kz75xh /tmp/tmp.JKtQodTJbX ++ return 0 + [[ 2Gi == \2\G\i ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-2 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-2 was resized + return 0 + return 0 + echo + [[ 0 == 1 ]] + [[ -n '' ]] + desc 'create resourcequota' + set +o xtrace ----------------------------------------------------------------------------------- create resourcequota ----------------------------------------------------------------------------------- + apply_resourcequota 7Gi + local quota=7Gi + local default_sc ++ get_default_storageclass ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QeUS7afVHk +++ mktemp ++ local LAST_ERR=/tmp/tmp.8LZg3m1MQo ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.QeUS7afVHk ++ cat /tmp/tmp.8LZg3m1MQo ++ rm /tmp/tmp.QeUS7afVHk /tmp/tmp.8LZg3m1MQo ++ return 0 + default_sc=standard-rwo + echo 'Applying resourcequota for default storageclass standard-rwo with quota 7Gi' Applying resourcequota for default storageclass standard-rwo with quota 7Gi + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/e2e-tests/pvc-resize/conf/resourcequota.yml + sed s/STORAGECLASS/standard-rwo/ + sed s/QUOTA/7Gi/ + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.fX0zg9UeNr ++ mktemp + local LAST_ERR=/tmp/tmp.FRGL4Hbr3B + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.fX0zg9UeNr resourcequota/default-storage-quota created + cat /tmp/tmp.FRGL4Hbr3B + rm /tmp/tmp.fX0zg9UeNr /tmp/tmp.FRGL4Hbr3B + return 0 + patch_pvc_request some-name 3G + local cluster=some-name + local size=3G + echo 'Patching PVC request to 3G in some-name' Patching PVC request to 3G in some-name + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"3G"}]' ++ mktemp + local LAST_OUT=/tmp/tmp.yRNMcfe7MB ++ mktemp + local LAST_ERR=/tmp/tmp.zuUOIonQJE + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"3G"}]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.yRNMcfe7MB perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.zuUOIonQJE + rm /tmp/tmp.yRNMcfe7MB /tmp/tmp.zuUOIonQJE + return 0 + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ai5qFGaik8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.veiVG0WdPJ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ai5qFGaik8 ++ cat /tmp/tmp.veiVG0WdPJ ++ rm /tmp/tmp.ai5qFGaik8 /tmp/tmp.veiVG0WdPJ ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo .OK .OK + echo + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-0 3Gi + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-0 + local expected_size=3Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YgFymdaVPl +++ mktemp ++ local LAST_ERR=/tmp/tmp.03QhaMoF4v ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.YgFymdaVPl ++ cat /tmp/tmp.03QhaMoF4v ++ rm /tmp/tmp.YgFymdaVPl /tmp/tmp.03QhaMoF4v ++ return 0 + [[ 2Gi == \3\G\i ]] + [[ 0 -ge 120 ]] + echo -n . .+ sleep 5 + retry=1 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EWPUmUPHUK +++ mktemp ++ local LAST_ERR=/tmp/tmp.9kV8KCMPcJ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.EWPUmUPHUK ++ cat /tmp/tmp.9kV8KCMPcJ ++ rm /tmp/tmp.EWPUmUPHUK /tmp/tmp.9kV8KCMPcJ ++ return 0 + [[ 2Gi == \3\G\i ]] + [[ 1 -ge 120 ]] + echo -n . .+ sleep 5 + retry=2 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jhXE6gDpIE +++ mktemp ++ local LAST_ERR=/tmp/tmp.SQa0P7S2ed ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jhXE6gDpIE ++ cat /tmp/tmp.SQa0P7S2ed ++ rm /tmp/tmp.jhXE6gDpIE /tmp/tmp.SQa0P7S2ed ++ return 0 + [[ 2Gi == \3\G\i ]] + [[ 2 -ge 120 ]] + echo -n . .+ sleep 5 + retry=3 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.39lj7GzpwZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.eLU32Wurf0 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.39lj7GzpwZ ++ cat /tmp/tmp.eLU32Wurf0 ++ rm /tmp/tmp.39lj7GzpwZ /tmp/tmp.eLU32Wurf0 ++ return 0 + [[ 2Gi == \3\G\i ]] + [[ 3 -ge 120 ]] + echo -n . .+ sleep 5 + retry=4 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.aaUbhuUJgN +++ mktemp ++ local LAST_ERR=/tmp/tmp.NoR7aa6P6U ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.aaUbhuUJgN ++ cat /tmp/tmp.NoR7aa6P6U ++ rm /tmp/tmp.aaUbhuUJgN /tmp/tmp.NoR7aa6P6U ++ return 0 + [[ 2Gi == \3\G\i ]] + [[ 4 -ge 120 ]] + echo -n . .+ sleep 5 + retry=5 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.R6LLVxk74T +++ mktemp ++ local LAST_ERR=/tmp/tmp.BHSHAVIGqP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.R6LLVxk74T ++ cat /tmp/tmp.BHSHAVIGqP ++ rm /tmp/tmp.R6LLVxk74T /tmp/tmp.BHSHAVIGqP ++ return 0 + [[ 2Gi == \3\G\i ]] + [[ 5 -ge 120 ]] + echo -n . .+ sleep 5 + retry=6 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bHvFuIV512 +++ mktemp ++ local LAST_ERR=/tmp/tmp.V5rlqpWUuF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.bHvFuIV512 ++ cat /tmp/tmp.V5rlqpWUuF ++ rm /tmp/tmp.bHvFuIV512 /tmp/tmp.V5rlqpWUuF ++ return 0 + [[ 2Gi == \3\G\i ]] + [[ 6 -ge 120 ]] + echo -n . .+ sleep 5 + retry=7 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ubcDG04Gq0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.ipGqFZONF2 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ubcDG04Gq0 ++ cat /tmp/tmp.ipGqFZONF2 ++ rm /tmp/tmp.ubcDG04Gq0 /tmp/tmp.ipGqFZONF2 ++ return 0 + [[ 2Gi == \3\G\i ]] + [[ 7 -ge 120 ]] + echo -n . .+ sleep 5 + retry=8 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.S163WXnrYe +++ mktemp ++ local LAST_ERR=/tmp/tmp.z2gucvFQV9 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.S163WXnrYe ++ cat /tmp/tmp.z2gucvFQV9 ++ rm /tmp/tmp.S163WXnrYe /tmp/tmp.z2gucvFQV9 ++ return 0 + [[ 2Gi == \3\G\i ]] + [[ 8 -ge 120 ]] + echo -n . .+ sleep 5 + retry=9 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ip5yjb2oXy +++ mktemp ++ local LAST_ERR=/tmp/tmp.dZEF1PTW9F ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ip5yjb2oXy ++ cat /tmp/tmp.dZEF1PTW9F ++ rm /tmp/tmp.ip5yjb2oXy /tmp/tmp.dZEF1PTW9F ++ return 0 + [[ 2Gi == \3\G\i ]] + [[ 9 -ge 120 ]] + echo -n . .+ sleep 5 + retry=10 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9MUrqeZRlf +++ mktemp ++ local LAST_ERR=/tmp/tmp.gEUOwrgK4U ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9MUrqeZRlf ++ cat /tmp/tmp.gEUOwrgK4U ++ rm /tmp/tmp.9MUrqeZRlf /tmp/tmp.gEUOwrgK4U ++ return 0 + [[ 2Gi == \3\G\i ]] + [[ 10 -ge 120 ]] + echo -n . .+ sleep 5 + retry=11 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qOjOdpty7S +++ mktemp ++ local LAST_ERR=/tmp/tmp.Qs50eLNaVR ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.qOjOdpty7S ++ cat /tmp/tmp.Qs50eLNaVR ++ rm /tmp/tmp.qOjOdpty7S /tmp/tmp.Qs50eLNaVR ++ return 0 + [[ 2Gi == \3\G\i ]] + [[ 11 -ge 120 ]] + echo -n . .+ sleep 5 + retry=12 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IO7xizdpnw +++ mktemp ++ local LAST_ERR=/tmp/tmp.ygPRctAt8d ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.IO7xizdpnw ++ cat /tmp/tmp.ygPRctAt8d ++ rm /tmp/tmp.IO7xizdpnw /tmp/tmp.ygPRctAt8d ++ return 0 + [[ 3Gi == \3\G\i ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized + return 0 + apply_resourcequota 9Gi + local quota=9Gi + local default_sc ++ get_default_storageclass ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BNJDRCxYtm +++ mktemp ++ local LAST_ERR=/tmp/tmp.tlIwdj3IuI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.BNJDRCxYtm ++ cat /tmp/tmp.tlIwdj3IuI ++ rm /tmp/tmp.BNJDRCxYtm /tmp/tmp.tlIwdj3IuI ++ return 0 + default_sc=standard-rwo + echo 'Applying resourcequota for default storageclass standard-rwo with quota 9Gi' Applying resourcequota for default storageclass standard-rwo with quota 9Gi + sed s/STORAGECLASS/standard-rwo/ + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/e2e-tests/pvc-resize/conf/resourcequota.yml + sed s/QUOTA/9Gi/ + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.IV89K0UytT ++ mktemp + local LAST_ERR=/tmp/tmp.f804HIob6C + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.IV89K0UytT resourcequota/default-storage-quota configured + cat /tmp/tmp.f804HIob6C + rm /tmp/tmp.IV89K0UytT /tmp/tmp.f804HIob6C + return 0 + patch_pvc_request some-name 3G + local cluster=some-name + local size=3G + echo 'Patching PVC request to 3G in some-name' Patching PVC request to 3G in some-name + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"3G"}]' ++ mktemp + local LAST_OUT=/tmp/tmp.ZcEjqrM5Qu ++ mktemp + local LAST_ERR=/tmp/tmp.b7z08UabE7 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"3G"}]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ZcEjqrM5Qu perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.b7z08UabE7 + rm /tmp/tmp.ZcEjqrM5Qu /tmp/tmp.b7z08UabE7 + return 0 + wait_cluster_consistency some-name 42 + local cluster_name=some-name + local wait_time=42 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QZxfB7ioJs +++ mktemp ++ local LAST_ERR=/tmp/tmp.JKDJXAIxij ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.QZxfB7ioJs ++ cat /tmp/tmp.JKDJXAIxij ++ rm /tmp/tmp.QZxfB7ioJs /tmp/tmp.JKDJXAIxij ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 1 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FClQ45KKxm +++ mktemp ++ local LAST_ERR=/tmp/tmp.Tnak4Fh0SN ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.FClQ45KKxm ++ cat /tmp/tmp.Tnak4Fh0SN ++ rm /tmp/tmp.FClQ45KKxm /tmp/tmp.Tnak4Fh0SN ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 2 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.S9JfCnxtFq +++ mktemp ++ local LAST_ERR=/tmp/tmp.RyvbkELNi8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.S9JfCnxtFq ++ cat /tmp/tmp.RyvbkELNi8 ++ rm /tmp/tmp.S9JfCnxtFq /tmp/tmp.RyvbkELNi8 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 3 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oziBZ2R0bJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.N35eXlJ54I ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.oziBZ2R0bJ ++ cat /tmp/tmp.N35eXlJ54I ++ rm /tmp/tmp.oziBZ2R0bJ /tmp/tmp.N35eXlJ54I ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 4 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.W7zpp6ICCE +++ mktemp ++ local LAST_ERR=/tmp/tmp.SvI9eNuLwc ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.W7zpp6ICCE ++ cat /tmp/tmp.SvI9eNuLwc ++ rm /tmp/tmp.W7zpp6ICCE /tmp/tmp.SvI9eNuLwc ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo .OK .OK + echo + wait_all_pvc_resize 3Gi + local expected_size=3Gi + local max_retry=120 + local sleep_time=5 ++ kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name +++ mktemp ++ local LAST_OUT=/tmp/tmp.WsbmfCbG4v +++ mktemp ++ local LAST_ERR=/tmp/tmp.c5jjJfZWEN ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc -l app.kubernetes.io/component=mongod -o name ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.WsbmfCbG4v ++ cat /tmp/tmp.c5jjJfZWEN ++ rm /tmp/tmp.WsbmfCbG4v /tmp/tmp.c5jjJfZWEN ++ return 0 + for pvc in $(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name) + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-0 3Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-0 + local expected_size=3Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9MH5spnAW8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.FSvGH8ROLX ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9MH5spnAW8 ++ cat /tmp/tmp.FSvGH8ROLX ++ rm /tmp/tmp.9MH5spnAW8 /tmp/tmp.FSvGH8ROLX ++ return 0 + [[ 3Gi == \3\G\i ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized + return 0 + for pvc in $(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name) + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-1 3Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-1 + local expected_size=3Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-1 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-1 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.b7apaNhAPt +++ mktemp ++ local LAST_ERR=/tmp/tmp.O5SYqG629q ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.b7apaNhAPt ++ cat /tmp/tmp.O5SYqG629q ++ rm /tmp/tmp.b7apaNhAPt /tmp/tmp.O5SYqG629q ++ return 0 + [[ 3Gi == \3\G\i ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-1 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-1 was resized + return 0 + for pvc in $(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name) + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-2 3Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-2 + local expected_size=3Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-2 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-2 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.F9dbwrV1JU +++ mktemp ++ local LAST_ERR=/tmp/tmp.wheRcH4REN ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.F9dbwrV1JU ++ cat /tmp/tmp.wheRcH4REN ++ rm /tmp/tmp.F9dbwrV1JU /tmp/tmp.wheRcH4REN ++ return 0 + [[ 3Gi == \3\G\i ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-2 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-2 was resized + return 0 + return 0 + desc 'test downscale' + set +o xtrace ----------------------------------------------------------------------------------- test downscale ----------------------------------------------------------------------------------- + patch_pvc_request some-name 1G + local cluster=some-name + local size=1G + echo 'Patching PVC request to 1G in some-name' Patching PVC request to 1G in some-name + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"1G"}]' ++ mktemp + local LAST_OUT=/tmp/tmp.nJ4d2PmWvR ++ mktemp + local LAST_ERR=/tmp/tmp.ztGGkZ7jft + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"1G"}]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.nJ4d2PmWvR perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.ztGGkZ7jft + rm /tmp/tmp.nJ4d2PmWvR /tmp/tmp.ztGGkZ7jft + return 0 + wait_cluster_status some-name error + local cluster=some-name + local expected=error + echo -n 'Waiting for psmdb/some-name status to be error' Waiting for psmdb/some-name status to be error++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RnhUqVqa6K +++ mktemp ++ local LAST_ERR=/tmp/tmp.LfLgLN0W1g ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.RnhUqVqa6K ++ cat /tmp/tmp.LfLgLN0W1g ++ rm /tmp/tmp.RnhUqVqa6K /tmp/tmp.LfLgLN0W1g ++ return 0 + [[ error == \e\r\r\o\r ]] + echo + echo 'psmdb/some-name status is error' psmdb/some-name status is error + patch_pvc_request some-name 3G + local cluster=some-name + local size=3G + echo 'Patching PVC request to 3G in some-name' Patching PVC request to 3G in some-name + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"3G"}]' ++ mktemp + local LAST_OUT=/tmp/tmp.CGQLlhj1xV ++ mktemp + local LAST_ERR=/tmp/tmp.bCyr0N44Hu + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"3G"}]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.CGQLlhj1xV perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.bCyr0N44Hu + rm /tmp/tmp.CGQLlhj1xV /tmp/tmp.bCyr0N44Hu + return 0 + wait_cluster_status some-name ready + local cluster=some-name + local expected=ready + echo -n 'Waiting for psmdb/some-name status to be ready' Waiting for psmdb/some-name status to be ready++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.haPVGeuUZb +++ mktemp ++ local LAST_ERR=/tmp/tmp.suqJKSNokr ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.haPVGeuUZb ++ cat /tmp/tmp.suqJKSNokr ++ rm /tmp/tmp.haPVGeuUZb /tmp/tmp.suqJKSNokr ++ return 0 + [[ error == \r\e\a\d\y ]] + [[ 4 -ge 60 ]] + echo -n . .+ sleep 5 + retry=5 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pV0uhs4TeF +++ mktemp ++ local LAST_ERR=/tmp/tmp.iNLUwqugEO ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.pV0uhs4TeF ++ cat /tmp/tmp.iNLUwqugEO ++ rm /tmp/tmp.pV0uhs4TeF /tmp/tmp.iNLUwqugEO ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo + echo 'psmdb/some-name status is ready' psmdb/some-name status is ready + desc 'test restartedAt annotation preservation during PVC resize' + set +o xtrace ----------------------------------------------------------------------------------- test restartedAt annotation preservation during PVC resize ----------------------------------------------------------------------------------- + echo 'delete resourcequota' delete resourcequota + kubectl_bin delete resourcequota default-storage-quota ++ mktemp + local LAST_OUT=/tmp/tmp.77dRCeJjM9 ++ mktemp + local LAST_ERR=/tmp/tmp.rIUgTRZfBm + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete resourcequota default-storage-quota + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.77dRCeJjM9 resourcequota "default-storage-quota" deleted from pvc-resize-26375 namespace + cat /tmp/tmp.rIUgTRZfBm + rm /tmp/tmp.77dRCeJjM9 /tmp/tmp.rIUgTRZfBm + return 0 + sts_name=some-name-rs0 + kubectl_bin rollout restart statefulset/some-name-rs0 ++ mktemp + local LAST_OUT=/tmp/tmp.gGmAVzSNHG ++ mktemp + local LAST_ERR=/tmp/tmp.wlM9zfeuFP + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl rollout restart statefulset/some-name-rs0 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.gGmAVzSNHG statefulset.apps/some-name-rs0 restarted + cat /tmp/tmp.wlM9zfeuFP + rm /tmp/tmp.gGmAVzSNHG /tmp/tmp.wlM9zfeuFP + return 0 + wait_for_running some-name-rs0 3 + local name=some-name-rs0 + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=rs0 + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-rs0-0 + local pod=some-name-rs0-0 + set +o xtrace waiting for pod/some-name-rs0-0 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-rs0-1 + local pod=some-name-rs0-1 + set +o xtrace waiting for pod/some-name-rs0-1 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tHke5qoHgC +++ mktemp ++ local LAST_ERR=/tmp/tmp.pOlXHTgcCu ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.tHke5qoHgC ++ cat /tmp/tmp.pOlXHTgcCu ++ rm /tmp/tmp.tHke5qoHgC /tmp/tmp.pOlXHTgcCu ++ return 0 + [[ false == \t\r\u\e ]] + wait_pod some-name-rs0-2 + local pod=some-name-rs0-2 + set +o xtrace waiting for pod/some-name-rs0-2 to be ready.OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PnlVAJ6Pj6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.arFFzaMtml ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.PnlVAJ6Pj6 ++ cat /tmp/tmp.arFFzaMtml ++ rm /tmp/tmp.PnlVAJ6Pj6 /tmp/tmp.arFFzaMtml ++ return 0 + [[ false == \t\r\u\e ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IatTdaRTIq +++ mktemp ++ local LAST_ERR=/tmp/tmp.9PgHVs6mtF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.IatTdaRTIq ++ cat /tmp/tmp.9PgHVs6mtF ++ rm /tmp/tmp.IatTdaRTIq /tmp/tmp.9PgHVs6mtF ++ return 0 + [[ false == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness........................................ ++ kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={.spec.template.metadata.annotations.kubectl\.kubernetes\.io/restartedAt}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jr2WWFySox +++ mktemp ++ local LAST_ERR=/tmp/tmp.2rXwVKQYbe ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get statefulset some-name-rs0 -o 'jsonpath={.spec.template.metadata.annotations.kubectl\.kubernetes\.io/restartedAt}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jr2WWFySox ++ cat /tmp/tmp.2rXwVKQYbe ++ rm /tmp/tmp.jr2WWFySox /tmp/tmp.2rXwVKQYbe ++ return 0 + restarted_at_before=2026-01-23T13:02:31Z + '[' -z 2026-01-23T13:02:31Z ']' + patch_pvc_request some-name 4G + local cluster=some-name + local size=4G + echo 'Patching PVC request to 4G in some-name' Patching PVC request to 4G in some-name + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"4G"}]' ++ mktemp + local LAST_OUT=/tmp/tmp.5BvMcgeFfh ++ mktemp + local LAST_ERR=/tmp/tmp.wA8RC4snCW + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"4G"}]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.5BvMcgeFfh perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.wA8RC4snCW + rm /tmp/tmp.5BvMcgeFfh /tmp/tmp.wA8RC4snCW + return 0 + wait_all_pvc_resize 4Gi + local expected_size=4Gi + local max_retry=120 + local sleep_time=5 ++ kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name +++ mktemp ++ local LAST_OUT=/tmp/tmp.UgMiXyD26E +++ mktemp ++ local LAST_ERR=/tmp/tmp.5EOUrVURuq ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc -l app.kubernetes.io/component=mongod -o name ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.UgMiXyD26E ++ cat /tmp/tmp.5EOUrVURuq ++ rm /tmp/tmp.UgMiXyD26E /tmp/tmp.5EOUrVURuq ++ return 0 + for pvc in $(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name) + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-0 4Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-0 + local expected_size=4Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ggqUwc552l +++ mktemp ++ local LAST_ERR=/tmp/tmp.IQqDPJS6dq ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ggqUwc552l ++ cat /tmp/tmp.IQqDPJS6dq ++ rm /tmp/tmp.ggqUwc552l /tmp/tmp.IQqDPJS6dq ++ return 0 + [[ 3Gi == \4\G\i ]] + [[ 0 -ge 120 ]] + echo -n . .+ sleep 5 + retry=1 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.V9T6wtZiaP +++ mktemp ++ local LAST_ERR=/tmp/tmp.1bhyLvz3FH ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.V9T6wtZiaP ++ cat /tmp/tmp.1bhyLvz3FH ++ rm /tmp/tmp.V9T6wtZiaP /tmp/tmp.1bhyLvz3FH ++ return 0 + [[ 3Gi == \4\G\i ]] + [[ 1 -ge 120 ]] + echo -n . .+ sleep 5 + retry=2 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.S2MielqcBG +++ mktemp ++ local LAST_ERR=/tmp/tmp.JdivBgHPKv ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.S2MielqcBG ++ cat /tmp/tmp.JdivBgHPKv ++ rm /tmp/tmp.S2MielqcBG /tmp/tmp.JdivBgHPKv ++ return 0 + [[ 3Gi == \4\G\i ]] + [[ 2 -ge 120 ]] + echo -n . .+ sleep 5 + retry=3 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xBXKrpFEpG +++ mktemp ++ local LAST_ERR=/tmp/tmp.ORKV4V6qqh ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xBXKrpFEpG ++ cat /tmp/tmp.ORKV4V6qqh ++ rm /tmp/tmp.xBXKrpFEpG /tmp/tmp.ORKV4V6qqh ++ return 0 + [[ 3Gi == \4\G\i ]] + [[ 3 -ge 120 ]] + echo -n . .+ sleep 5 + retry=4 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.S2aAwkFC8K +++ mktemp ++ local LAST_ERR=/tmp/tmp.5r17bvt4D3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.S2aAwkFC8K ++ cat /tmp/tmp.5r17bvt4D3 ++ rm /tmp/tmp.S2aAwkFC8K /tmp/tmp.5r17bvt4D3 ++ return 0 + [[ 3Gi == \4\G\i ]] + [[ 4 -ge 120 ]] + echo -n . .+ sleep 5 + retry=5 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.A9LFHqYbLN +++ mktemp ++ local LAST_ERR=/tmp/tmp.BSiVwnNy4l ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.A9LFHqYbLN ++ cat /tmp/tmp.BSiVwnNy4l ++ rm /tmp/tmp.A9LFHqYbLN /tmp/tmp.BSiVwnNy4l ++ return 0 + [[ 3Gi == \4\G\i ]] + [[ 5 -ge 120 ]] + echo -n . .+ sleep 5 + retry=6 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UN5mHcutf3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.1oLEKTMhjl ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.UN5mHcutf3 ++ cat /tmp/tmp.1oLEKTMhjl ++ rm /tmp/tmp.UN5mHcutf3 /tmp/tmp.1oLEKTMhjl ++ return 0 + [[ 3Gi == \4\G\i ]] + [[ 6 -ge 120 ]] + echo -n . .+ sleep 5 + retry=7 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6CoXRj3j34 +++ mktemp ++ local LAST_ERR=/tmp/tmp.wyxsKgyvhj ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.6CoXRj3j34 ++ cat /tmp/tmp.wyxsKgyvhj ++ rm /tmp/tmp.6CoXRj3j34 /tmp/tmp.wyxsKgyvhj ++ return 0 + [[ 3Gi == \4\G\i ]] + [[ 7 -ge 120 ]] + echo -n . .+ sleep 5 + retry=8 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gMAx9AjwBl +++ mktemp ++ local LAST_ERR=/tmp/tmp.aWDRGxuKA3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.gMAx9AjwBl ++ cat /tmp/tmp.aWDRGxuKA3 ++ rm /tmp/tmp.gMAx9AjwBl /tmp/tmp.aWDRGxuKA3 ++ return 0 + [[ 3Gi == \4\G\i ]] + [[ 8 -ge 120 ]] + echo -n . .+ sleep 5 + retry=9 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wF6glBT89x +++ mktemp ++ local LAST_ERR=/tmp/tmp.VCs64KE90s ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.wF6glBT89x ++ cat /tmp/tmp.VCs64KE90s ++ rm /tmp/tmp.wF6glBT89x /tmp/tmp.VCs64KE90s ++ return 0 + [[ 3Gi == \4\G\i ]] + [[ 9 -ge 120 ]] + echo -n . .+ sleep 5 + retry=10 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vM7iu4W3Qn +++ mktemp ++ local LAST_ERR=/tmp/tmp.i758xAR289 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.vM7iu4W3Qn ++ cat /tmp/tmp.i758xAR289 ++ rm /tmp/tmp.vM7iu4W3Qn /tmp/tmp.i758xAR289 ++ return 0 + [[ 3Gi == \4\G\i ]] + [[ 10 -ge 120 ]] + echo -n . .+ sleep 5 + retry=11 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5i9cVYI9ao +++ mktemp ++ local LAST_ERR=/tmp/tmp.Mwrk6OVWvL ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5i9cVYI9ao ++ cat /tmp/tmp.Mwrk6OVWvL ++ rm /tmp/tmp.5i9cVYI9ao /tmp/tmp.Mwrk6OVWvL ++ return 0 + [[ 3Gi == \4\G\i ]] + [[ 11 -ge 120 ]] + echo -n . .+ sleep 5 + retry=12 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1YO8UQJyun +++ mktemp ++ local LAST_ERR=/tmp/tmp.0u9TdXT2l0 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.1YO8UQJyun ++ cat /tmp/tmp.0u9TdXT2l0 ++ rm /tmp/tmp.1YO8UQJyun /tmp/tmp.0u9TdXT2l0 ++ return 0 + [[ 4Gi == \4\G\i ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized + return 0 + for pvc in $(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name) + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-1 4Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-1 + local expected_size=4Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-1 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-1 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ia4ATaE9NW +++ mktemp ++ local LAST_ERR=/tmp/tmp.GKRUe6MZW2 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ia4ATaE9NW ++ cat /tmp/tmp.GKRUe6MZW2 ++ rm /tmp/tmp.ia4ATaE9NW /tmp/tmp.GKRUe6MZW2 ++ return 0 + [[ 4Gi == \4\G\i ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-1 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-1 was resized + return 0 + for pvc in $(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name) + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-2 4Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-2 + local expected_size=4Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-2 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-2 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gJxajIX2dp +++ mktemp ++ local LAST_ERR=/tmp/tmp.5ysxMbio1b ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.gJxajIX2dp ++ cat /tmp/tmp.5ysxMbio1b ++ rm /tmp/tmp.gJxajIX2dp /tmp/tmp.5ysxMbio1b ++ return 0 + [[ 4Gi == \4\G\i ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-2 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-2 was resized + return 0 + return 0 + sleep_with_log 30 'wait for StatefulSet recreation after orphan delete during PVC resize' + local d=30 + local 'msg=wait for StatefulSet recreation after orphan delete during PVC resize' + log 'wait for StatefulSet recreation after orphan delete during PVC resize. sleeping for 30 seconds' + set +o xtrace [2026-01-23T13:05:34+0000] wait for StatefulSet recreation after orphan delete during PVC resize. sleeping for 30 seconds + sleep 30 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PjVLDHh0F2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.cWlIZTeJ7c ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.PjVLDHh0F2 ++ cat /tmp/tmp.cWlIZTeJ7c ++ rm /tmp/tmp.PjVLDHh0F2 /tmp/tmp.cWlIZTeJ7c ++ return 0 + cluster_status=ready ++ kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={.spec.template.metadata.annotations.kubectl\.kubernetes\.io/restartedAt}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nMwMxlFG59 +++ mktemp ++ local LAST_ERR=/tmp/tmp.bFEaM4NoFb ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get statefulset some-name-rs0 -o 'jsonpath={.spec.template.metadata.annotations.kubectl\.kubernetes\.io/restartedAt}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.nMwMxlFG59 ++ cat /tmp/tmp.bFEaM4NoFb ++ rm /tmp/tmp.nMwMxlFG59 /tmp/tmp.bFEaM4NoFb ++ return 0 + restarted_at_after=2026-01-23T13:02:31Z + [[ 2026-01-23T13:02:31Z != \2\0\2\6\-\0\1\-\2\3\T\1\3\:\0\2\:\3\1\Z ]] + [[ ready == \i\n\i\t\i\a\l\i\z\i\n\g ]] + wait_cluster_status some-name ready + local cluster=some-name + local expected=ready + echo -n 'Waiting for psmdb/some-name status to be ready' Waiting for psmdb/some-name status to be ready++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XsoFDBvZ35 +++ mktemp ++ local LAST_ERR=/tmp/tmp.SrQXQxpiKh ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.XsoFDBvZ35 ++ cat /tmp/tmp.SrQXQxpiKh ++ rm /tmp/tmp.XsoFDBvZ35 /tmp/tmp.SrQXQxpiKh ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo + echo 'psmdb/some-name status is ready' psmdb/some-name status is ready + wait_for_running some-name-rs0 3 + local name=some-name-rs0 + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=rs0 + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-rs0-0 + local pod=some-name-rs0-0 + set +o xtrace waiting for pod/some-name-rs0-0 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-rs0-1 + local pod=some-name-rs0-1 + set +o xtrace waiting for pod/some-name-rs0-1 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YLrLE7DkfW +++ mktemp ++ local LAST_ERR=/tmp/tmp.GdJpKXIn0Z ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.YLrLE7DkfW ++ cat /tmp/tmp.GdJpKXIn0Z ++ rm /tmp/tmp.YLrLE7DkfW /tmp/tmp.GdJpKXIn0Z ++ return 0 + [[ false == \t\r\u\e ]] + wait_pod some-name-rs0-2 + local pod=some-name-rs0-2 + set +o xtrace waiting for pod/some-name-rs0-2 to be ready.OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.D2NTOTp1C6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.DgxyTWZ3Vy ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.D2NTOTp1C6 ++ cat /tmp/tmp.DgxyTWZ3Vy ++ rm /tmp/tmp.D2NTOTp1C6 /tmp/tmp.DgxyTWZ3Vy ++ return 0 + [[ false == \t\r\u\e ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Zqx9NXn6jw +++ mktemp ++ local LAST_ERR=/tmp/tmp.e8PsrEsvSZ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Zqx9NXn6jw ++ cat /tmp/tmp.e8PsrEsvSZ ++ rm /tmp/tmp.Zqx9NXn6jw /tmp/tmp.e8PsrEsvSZ ++ return 0 + [[ false == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness + destroy pvc-resize-26375 + local namespace=pvc-resize-26375 + local ignore_logs=true + [[ 0 == 1 ]] + desc 'destroy cluster/operator and all other resources' + set +o xtrace ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + '[' true == false ']' + delete_backups + desc 'Delete psmdb-backup' + set +o xtrace ----------------------------------------------------------------------------------- Delete psmdb-backup ----------------------------------------------------------------------------------- ++ kubectl_bin get psmdb-backup --no-headers ++ wc -l +++ mktemp ++ local LAST_OUT=/tmp/tmp.HsrwqcvVAD +++ mktemp ++ local LAST_ERR=/tmp/tmp.vENxq5BNkb ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb-backup --no-headers ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.HsrwqcvVAD ++ cat /tmp/tmp.vENxq5BNkb No resources found in pvc-resize-26375 namespace. ++ rm /tmp/tmp.HsrwqcvVAD /tmp/tmp.vENxq5BNkb ++ return 0 + '[' 0 '!=' 0 ']' + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.Y0QKAXIumq ++ mktemp + local LAST_ERR=/tmp/tmp.wTkYeqcnJQ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Y0QKAXIumq customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.wTkYeqcnJQ + rm /tmp/tmp.Y0QKAXIumq /tmp/tmp.wTkYeqcnJQ + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/deploy/crd.yaml ++ grep -v '\-\-\-' grep: warning: stray \ before - grep: warning: stray \ before - + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + grep -v NAMESPACE error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.wY30FGD6LK ++ mktemp + local LAST_ERR=/tmp/tmp.zfWr3cHZka + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.wY30FGD6LK + cat /tmp/tmp.zfWr3cHZka + rm /tmp/tmp.wY30FGD6LK /tmp/tmp.zfWr3cHZka + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + grep -v NAMESPACE error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.6EQLpX9Zq7 ++ mktemp + local LAST_ERR=/tmp/tmp.4fJ68IZhT7 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.6EQLpX9Zq7 + cat /tmp/tmp.4fJ68IZhT7 + rm /tmp/tmp.6EQLpX9Zq7 /tmp/tmp.4fJ68IZhT7 + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + grep -v NAMESPACE + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch perconaservermongodbs.psmdb.percona.com -n pvc-resize-26375 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaservermongodb.psmdb.percona.com/some-name patched + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.CggztagSSd ++ mktemp + local LAST_ERR=/tmp/tmp.vMwnpA6LKK + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.CggztagSSd customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com condition met + cat /tmp/tmp.vMwnpA6LKK + rm /tmp/tmp.CggztagSSd /tmp/tmp.vMwnpA6LKK + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.3csxjOTTvJ ++ mktemp + local LAST_ERR=/tmp/tmp.EeMgT9kZsg + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2213/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.3csxjOTTvJ clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.EeMgT9kZsg + rm /tmp/tmp.3csxjOTTvJ /tmp/tmp.EeMgT9kZsg + return 0 + destroy_cert_manager + kubectl_bin delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.GPyPWobZKZ ++ mktemp + local LAST_ERR=/tmp/tmp.4axwvlruPW + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.GPyPWobZKZ + cat /tmp/tmp.4axwvlruPW Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 0 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.GPyPWobZKZ + cat /tmp/tmp.4axwvlruPW Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 4 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.GPyPWobZKZ + cat /tmp/tmp.4axwvlruPW Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 8 + cat /tmp/tmp.GPyPWobZKZ + cat /tmp/tmp.4axwvlruPW Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + rm /tmp/tmp.GPyPWobZKZ /tmp/tmp.4axwvlruPW + return 1 + true + '[' -n '' ']' + '[' -n psmdb-operator ']' + kubectl_bin delete --grace-period=0 --force=true namespace pvc-resize-26375 + rm -rf /tmp/tmp.qmg2Iuu1GB + kubectl_bin delete --grace-period=0 --force=true namespace psmdb-operator ++ mktemp ++ mktemp + desc 'test passed' + set +o xtrace ----------------------------------------------------------------------------------- test passed ----------------------------------------------------------------------------------- + local LAST_OUT=/tmp/tmp.WRSrFBZp67 ++ mktemp + local LAST_OUT=/tmp/tmp.bkfNaQOTzG ++ mktemp + local LAST_ERR=/tmp/tmp.6a45l194hH + local exit_status=0 + local timeout=4 ++ seq 0 2 + local LAST_ERR=/tmp/tmp.P3VmGtU82z + local exit_status=0 + local timeout=4 + for i in $(seq 0 2) + set +e + kubectl delete --grace-period=0 --force=true namespace psmdb-operator ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete --grace-period=0 --force=true namespace pvc-resize-26375