Log: /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/e2e-tests/logs/pvc-resize.log WARNING: version difference between client (1.33) and server (1.30) exceeds the supported minor version skew of +/-1 WARNING: version difference between client (1.33) and server (1.30) exceeds the supported minor version skew of +/-1 WARNING: version difference between client (1.33) and server (1.30) exceeds the supported minor version skew of +/-1 + '[' 0 == 1 ']' + ensure_default_sc_allows_expansion + local default_sc ++ get_default_storageclass ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6DsjG83Wjc +++ mktemp ++ local LAST_ERR=/tmp/tmp.wh2qIOS2tU ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.6DsjG83Wjc ++ cat /tmp/tmp.wh2qIOS2tU ++ rm /tmp/tmp.6DsjG83Wjc /tmp/tmp.wh2qIOS2tU ++ return 0 + default_sc=standard-rwo + echo 'Checking if default storageclass standard-rwo allows volume expansion' Checking if default storageclass standard-rwo allows volume expansion + local allowVolumeExpansion ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.name=="standard-rwo")].allowVolumeExpansion}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.az6KjYYR5R +++ mktemp ++ local LAST_ERR=/tmp/tmp.r9FXqniKzC ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.name=="standard-rwo")].allowVolumeExpansion}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.az6KjYYR5R ++ cat /tmp/tmp.r9FXqniKzC ++ rm /tmp/tmp.az6KjYYR5R /tmp/tmp.r9FXqniKzC ++ return 0 + allowVolumeExpansion=true + [[ true != \t\r\u\e ]] + create_infra pvc-resize-28274 + local ns=pvc-resize-28274 + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.xrtqsZlrPe ++ mktemp + local LAST_ERR=/tmp/tmp.tUkgai8wZg + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.xrtqsZlrPe customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.tUkgai8wZg + rm /tmp/tmp.xrtqsZlrPe /tmp/tmp.tUkgai8wZg + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/deploy/crd.yaml ++ grep -v '\-\-\-' + for crd_name in '$(yq eval '\''.metadata.name'\'' "${src_dir}/deploy/crd.yaml" | grep -v '\''\-\-\-'\'')' + grep -v NAMESPACE + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.7BcRoU31el ++ mktemp + local LAST_ERR=/tmp/tmp.edw9Urj57G + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.7BcRoU31el + cat /tmp/tmp.edw9Urj57G + rm /tmp/tmp.7BcRoU31el /tmp/tmp.edw9Urj57G + return 0 + for crd_name in '$(yq eval '\''.metadata.name'\'' "${src_dir}/deploy/crd.yaml" | grep -v '\''\-\-\-'\'')' + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + grep -v NAMESPACE + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.p5D0oKocHc ++ mktemp + local LAST_ERR=/tmp/tmp.36JiY9ijVq + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.p5D0oKocHc + cat /tmp/tmp.36JiY9ijVq + rm /tmp/tmp.p5D0oKocHc /tmp/tmp.36JiY9ijVq + return 0 + for crd_name in '$(yq eval '\''.metadata.name'\'' "${src_dir}/deploy/crd.yaml" | grep -v '\''\-\-\-'\'')' + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + grep -v NAMESPACE + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + kubectl patch perconaservermongodbs.psmdb.percona.com -n pvc-resize-29365 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaservermongodb.psmdb.percona.com/some-name patched + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.GRB15k91fr ++ mktemp + local LAST_ERR=/tmp/tmp.odSX3Dqv9L + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.GRB15k91fr customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com condition met + cat /tmp/tmp.odSX3Dqv9L + rm /tmp/tmp.GRB15k91fr /tmp/tmp.odSX3Dqv9L + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.AWJqtcvozY ++ mktemp + local LAST_ERR=/tmp/tmp.5S7SA8qSs2 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.AWJqtcvozY clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.5S7SA8qSs2 + rm /tmp/tmp.AWJqtcvozY /tmp/tmp.5S7SA8qSs2 + return 0 + check_crd_for_deletion PR-1904-a4e779d3 + local git_tag=PR-1904-a4e779d3 ++ yq eval .metadata.name ++ /usr/bin/sed s/---//g ++ curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/PR-1904-a4e779d3/deploy/crd.yaml ++ /usr/bin/sed ':a;N;$!ba;s/\n/ /g' + for crd_name in '$(curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/${git_tag}/deploy/crd.yaml | yq eval '\''.metadata.name'\'' | $sed '\''s/---//g'\'' | $sed '\'':a;N;$!ba;s/\n/ /g'\'')' ++ kubectl_bin get crd/null -o 'jsonpath={.status.conditions[-1].type}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tADVJKFC90 +++ mktemp ++ local LAST_ERR=/tmp/tmp.0kr4mcfqoh ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.tADVJKFC90 ++ cat /tmp/tmp.0kr4mcfqoh Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 0 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.tADVJKFC90 ++ cat /tmp/tmp.0kr4mcfqoh Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 4 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.tADVJKFC90 ++ cat /tmp/tmp.0kr4mcfqoh Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 8 ++ cat /tmp/tmp.tADVJKFC90 ++ cat /tmp/tmp.0kr4mcfqoh Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ rm /tmp/tmp.tADVJKFC90 /tmp/tmp.0kr4mcfqoh ++ return 1 + [[ '' == \T\e\r\m\i\n\a\t\i\n\g ]] + '[' -n psmdb-operator ']' + create_namespace psmdb-operator + local namespace=psmdb-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// ++ helm list --all-namespaces --filter chaos-mesh + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get clusterrolebinding + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + awk '{print$1}' + '[' -n '' ']' + desc 'cleaned up old namespaces psmdb-operator' + xargs kubectl delete ns + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace psmdb-operator --ignore-not-found ++ mktemp + egrep -v '^kube-|^default|Terminating|psmdb-operator|openshift|gke-mcs|^NAME' + local LAST_OUT=/tmp/tmp.v435A8uZvF ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.KFq4f8DN1M + local LAST_ERR=/tmp/tmp.vz8c1RXYBc ++ mktemp + local exit_status=0 + local timeout=4 + local LAST_ERR=/tmp/tmp.6SEu6kwtBZ + local exit_status=0 + local timeout=4 ++ seq 0 2 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace psmdb-operator --ignore-not-found + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.KFq4f8DN1M + cat /tmp/tmp.6SEu6kwtBZ + rm /tmp/tmp.KFq4f8DN1M /tmp/tmp.6SEu6kwtBZ + return 0 namespace "gke-managed-cim" deleted namespace "gke-managed-system" deleted namespace "gmp-public" deleted namespace "gmp-system" deleted namespace "pvc-resize-29365" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.v435A8uZvF namespace "psmdb-operator" deleted + cat /tmp/tmp.vz8c1RXYBc + rm /tmp/tmp.v435A8uZvF /tmp/tmp.vz8c1RXYBc + return 0 + kubectl_bin wait --for=delete namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.QR04n8oi8x ++ mktemp + local LAST_ERR=/tmp/tmp.HbWX6xufKH + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.QR04n8oi8x + cat /tmp/tmp.HbWX6xufKH + rm /tmp/tmp.QR04n8oi8x /tmp/tmp.HbWX6xufKH + return 0 + desc 'create namespace psmdb-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.izi7LfG3ZA ++ mktemp + local LAST_ERR=/tmp/tmp.HJc10eqNkY + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.izi7LfG3ZA namespace/psmdb-operator created + cat /tmp/tmp.HJc10eqNkY + rm /tmp/tmp.izi7LfG3ZA /tmp/tmp.HJc10eqNkY + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.sKour1if2M +++ mktemp ++ local LAST_ERR=/tmp/tmp.CcEP4asqLC ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.sKour1if2M ++ cat /tmp/tmp.CcEP4asqLC ++ rm /tmp/tmp.sKour1if2M /tmp/tmp.CcEP4asqLC ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-1904-a4e779d3-11-cluster5 --namespace=psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.ZWYykUiCRQ ++ mktemp + local LAST_ERR=/tmp/tmp.jnDKhRKN0x + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-1904-a4e779d3-11-cluster5 --namespace=psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ZWYykUiCRQ Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-1904-a4e779d3-11-cluster5" modified. + cat /tmp/tmp.jnDKhRKN0x + rm /tmp/tmp.ZWYykUiCRQ /tmp/tmp.jnDKhRKN0x + return 0 + deploy_operator + desc 'start PSMDB operator' + set +o xtrace ----------------------------------------------------------------------------------- start PSMDB operator ----------------------------------------------------------------------------------- + local cr_file + '[' -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/e2e-tests/pvc-resize/conf/crd.yaml ']' + cr_file=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/deploy/crd.yaml + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.lE8kfbLauY ++ mktemp + local LAST_ERR=/tmp/tmp.1ZRyIhvkSB + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.lE8kfbLauY customresourcedefinition.apiextensions.k8s.io/perconaservermongodbbackups.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbrestores.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com serverside-applied + cat /tmp/tmp.1ZRyIhvkSB + rm /tmp/tmp.lE8kfbLauY /tmp/tmp.1ZRyIhvkSB + return 0 + '[' -n psmdb-operator ']' + apply_rbac cw-rbac + local operator_namespace=psmdb-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: psmdb-operator^' + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.ISWgEhgdFu ++ mktemp + local LAST_ERR=/tmp/tmp.Bchcf8Gke5 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ISWgEhgdFu clusterrole.rbac.authorization.k8s.io/percona-server-mongodb-operator created serviceaccount/percona-server-mongodb-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-server-mongodb-operator created + cat /tmp/tmp.Bchcf8Gke5 + rm /tmp/tmp.ISWgEhgdFu /tmp/tmp.Bchcf8Gke5 + return 0 + kubectl_bin apply -f - ++ mktemp + yq eval ' (.spec.template.spec.containers[].image = "perconalab/percona-server-mongodb-operator:PR-1904-a4e779d3") | ((.. | select(.[] == "DISABLE_TELEMETRY")) |= .value="true") | ((.. | select(.[] == "LOG_LEVEL")) |= .value="DEBUG")' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/deploy/cw-operator.yaml + local LAST_OUT=/tmp/tmp.PvnNFE0loo ++ mktemp + local LAST_ERR=/tmp/tmp.VAanrgv6Ug + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.PvnNFE0loo deployment.apps/percona-server-mongodb-operator created + cat /tmp/tmp.VAanrgv6Ug + rm /tmp/tmp.PvnNFE0loo /tmp/tmp.VAanrgv6Ug + return 0 + sleep 2 ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.U1Lgra0zbJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.cP5d8AZR0v ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.U1Lgra0zbJ ++ cat /tmp/tmp.cP5d8AZR0v ++ rm /tmp/tmp.U1Lgra0zbJ /tmp/tmp.cP5d8AZR0v ++ return 0 + wait_pod percona-server-mongodb-operator-85c5b67d9d-hr22g + local pod=percona-server-mongodb-operator-85c5b67d9d-hr22g + set +o xtrace waiting for pod/percona-server-mongodb-operator-85c5b67d9d-hr22g to be ready.OK + echo 'Print operator info from log' Print operator info from log + grep 'Manager starting up' ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.y9vTW7DXoi +++ mktemp ++ local LAST_ERR=/tmp/tmp.6PDs4Uh7Jv ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.y9vTW7DXoi ++ cat /tmp/tmp.6PDs4Uh7Jv ++ rm /tmp/tmp.y9vTW7DXoi /tmp/tmp.6PDs4Uh7Jv ++ return 0 + kubectl_bin logs percona-server-mongodb-operator-85c5b67d9d-hr22g ++ mktemp + local LAST_OUT=/tmp/tmp.JLf2E1OvqC ++ mktemp + local LAST_ERR=/tmp/tmp.e22ju9knCL + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl logs percona-server-mongodb-operator-85c5b67d9d-hr22g + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.JLf2E1OvqC + cat /tmp/tmp.e22ju9knCL + rm /tmp/tmp.JLf2E1OvqC /tmp/tmp.e22ju9knCL + return 0 2025-05-14T17:17:43.402Z INFO setup Manager starting up {"gitCommit": "a4e779d35a323625e9652ede4a65877e643ad719", "gitBranch": "PR-1904-a4e779d3", "buildTime": "", "goVersion": "go1.24.3", "os": "linux", "arch": "amd64"} + create_namespace pvc-resize-28274 + local namespace=pvc-resize-28274 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get MutatingWebhookConfiguration + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ kubectl api-resources ++ awk '{print $1}' ++ awk '{print $1}' ++ grep chaos-mesh.org ++ kubectl get crd + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ kubectl get clusterrolebinding ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + '[' -n '' ']' + desc 'cleaned up old namespaces pvc-resize-28274' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pvc-resize-28274 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pvc-resize-28274 --ignore-not-found ++ mktemp + xargs kubectl delete ns + kubectl_bin get ns + awk '{print$1}' + egrep -v '^kube-|^default|Terminating|psmdb-operator|openshift|gke-mcs|^NAME' + local LAST_OUT=/tmp/tmp.XawgFQzlP8 ++ mktemp + local LAST_OUT=/tmp/tmp.m7LhuDglQu ++ mktemp + local LAST_ERR=/tmp/tmp.0FUGknXbdJ + local exit_status=0 + local timeout=4 ++ mktemp + local LAST_ERR=/tmp/tmp.FYojVyOC9x + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pvc-resize-28274 --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.XawgFQzlP8 + cat /tmp/tmp.FYojVyOC9x + rm /tmp/tmp.XawgFQzlP8 /tmp/tmp.FYojVyOC9x + return 0 + kubectl_bin wait --for=delete namespace pvc-resize-28274 ++ mktemp + local LAST_OUT=/tmp/tmp.4wxDZHqBwv ++ mktemp + local LAST_ERR=/tmp/tmp.lYm3pNQcN9 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete namespace pvc-resize-28274 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.m7LhuDglQu + cat /tmp/tmp.0FUGknXbdJ + rm /tmp/tmp.m7LhuDglQu /tmp/tmp.0FUGknXbdJ + return 0 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.4wxDZHqBwv + cat /tmp/tmp.lYm3pNQcN9 + rm /tmp/tmp.4wxDZHqBwv /tmp/tmp.lYm3pNQcN9 + return 0 + desc 'create namespace pvc-resize-28274' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pvc-resize-28274 ----------------------------------------------------------------------------------- + kubectl_bin create namespace pvc-resize-28274 ++ mktemp + local LAST_OUT=/tmp/tmp.UGKivedPpu ++ mktemp + local LAST_ERR=/tmp/tmp.mg7exSQ3y5 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pvc-resize-28274 namespace "gke-managed-cim" deleted namespace "gke-managed-system" deleted namespace "gmp-public" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.UGKivedPpu namespace/pvc-resize-28274 created + cat /tmp/tmp.mg7exSQ3y5 + rm /tmp/tmp.UGKivedPpu /tmp/tmp.mg7exSQ3y5 + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.QMZYxZY2YU +++ mktemp ++ local LAST_ERR=/tmp/tmp.smfgej8kdz ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context namespace "gmp-system" deleted ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.QMZYxZY2YU ++ cat /tmp/tmp.smfgej8kdz ++ rm /tmp/tmp.QMZYxZY2YU /tmp/tmp.smfgej8kdz ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-1904-a4e779d3-11-cluster5 --namespace=pvc-resize-28274 ++ mktemp + local LAST_OUT=/tmp/tmp.XAnWo0uky0 ++ mktemp + local LAST_ERR=/tmp/tmp.PQS1w5FNQR + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-1904-a4e779d3-11-cluster5 --namespace=pvc-resize-28274 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.XAnWo0uky0 Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-1904-a4e779d3-11-cluster5" modified. + cat /tmp/tmp.PQS1w5FNQR + rm /tmp/tmp.XAnWo0uky0 /tmp/tmp.PQS1w5FNQR + return 0 + desc 'create secrets and psmdb client' + set +o xtrace ----------------------------------------------------------------------------------- create secrets and psmdb client ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/e2e-tests/conf/secrets.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.SV4HltZMvU ++ mktemp + local LAST_ERR=/tmp/tmp.Hsk6WfVZwS + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/e2e-tests/conf/secrets.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/e2e-tests/conf/client.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.SV4HltZMvU secret/some-users created deployment.apps/psmdb-client created + cat /tmp/tmp.Hsk6WfVZwS + rm /tmp/tmp.SV4HltZMvU /tmp/tmp.Hsk6WfVZwS + return 0 + desc 'create PSMDB cluster' + set +o xtrace ----------------------------------------------------------------------------------- create PSMDB cluster ----------------------------------------------------------------------------------- + cluster=some-name + '[' 0 == 1 ']' + spinup_psmdb some-name-rs0 /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/e2e-tests/pvc-resize/conf/some-name.yml + local cluster=some-name-rs0 + local config=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/e2e-tests/pvc-resize/conf/some-name.yml + local size=3 + desc 'create first PSMDB cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PSMDB cluster ----------------------------------------------------------------------------------- + apply_cluster /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/e2e-tests/pvc-resize/conf/some-name.yml + '[' -z '' ']' + kubectl_bin apply -f - + cat_config /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/e2e-tests/pvc-resize/conf/some-name.yml + yq eval '.spec.upgradeOptions.apply="Never"' ++ mktemp + yq eval '(.spec | select(has("initImage"))).initImage = "perconalab/percona-server-mongodb-operator:PR-1904-a4e779d3"' + local LAST_OUT=/tmp/tmp.gLpDlLGNqh ++ mktemp + local LAST_ERR=/tmp/tmp.pPwc0CR1PS + local exit_status=0 + local timeout=4 + yq eval '(.spec | select(.image == null)).image = "perconalab/percona-server-mongodb-operator:main-mongod7.0"' + yq eval '(.spec | select(has("backup"))).backup.image = "perconalab/percona-server-mongodb-operator:main-backup"' + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/e2e-tests/pvc-resize/conf/some-name.yml + yq eval '(.spec | select(has("pmm"))).pmm.image = "perconalab/pmm-client:dev-latest"' ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.gLpDlLGNqh perconaservermongodb.psmdb.percona.com/some-name created + cat /tmp/tmp.pPwc0CR1PS + rm /tmp/tmp.gLpDlLGNqh /tmp/tmp.pPwc0CR1PS + return 0 + desc 'check if Pod is started' + set +o xtrace ----------------------------------------------------------------------------------- check if Pod is started ----------------------------------------------------------------------------------- + wait_for_running some-name-rs0 3 + local name=some-name-rs0 + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=rs0 + local cluster_name=some-name ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + [[ 0 -eq 2 ]] + wait_pod some-name-rs0-0 + local pod=some-name-rs0-0 + set +o xtrace waiting for pod/some-name-rs0-0 to be ready...........OK + for i in '$(seq 0 $last_pod)' + [[ 1 -eq 2 ]] + wait_pod some-name-rs0-1 + local pod=some-name-rs0-1 + set +o xtrace waiting for pod/some-name-rs0-1 to be ready...........OK + for i in '$(seq 0 $last_pod)' + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VT594fGmIp +++ mktemp ++ local LAST_ERR=/tmp/tmp.5jsfGuNaTt ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.VT594fGmIp ++ cat /tmp/tmp.5jsfGuNaTt ++ rm /tmp/tmp.VT594fGmIp /tmp/tmp.5jsfGuNaTt ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-rs0-2 + local pod=some-name-rs0-2 + set +o xtrace waiting for pod/some-name-rs0-2 to be ready..........OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].non_voting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.myHClLvLPp +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZTHxQW1Ky5 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].non_voting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.myHClLvLPp ++ cat /tmp/tmp.ZTHxQW1Ky5 ++ rm /tmp/tmp.myHClLvLPp /tmp/tmp.ZTHxQW1Ky5 ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness + sleep 20 + compare_kubectl statefulset/some-name-rs0 + local resource=statefulset/some-name-rs0 + local postfix= + local skip_generation_check= + local expected_result=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/e2e-tests/pvc-resize/compare/statefulset_some-name-rs0.yml + local new_result=/tmp/tmp.i92JTyYaWV/statefulset_some-name-rs0.yml + '[' -n '' -a -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/e2e-tests/pvc-resize/compare/statefulset_some-name-rs0-oc.yml ']' + kubectl_bin get -o yaml statefulset/some-name-rs0 + yq eval ' del(.metadata.ownerReferences[].apiVersion) | del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "NAMESPACE")) | del(.metadata.selfLink) | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.status) | (.. | select(tag == "!!str")) |= sub("pvc-resize-28274", "NAME_SPACE") | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "extensions/v1beta1")) = "apps/v1" | (.. | select(. == "batch/v1beta1")) = "batch/v1" ' - ++ mktemp + local LAST_OUT=/tmp/tmp.eZaBB9EglQ ++ mktemp + local LAST_ERR=/tmp/tmp.nSE0BSDUgR + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-rs0 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.eZaBB9EglQ + cat /tmp/tmp.nSE0BSDUgR + rm /tmp/tmp.eZaBB9EglQ /tmp/tmp.nSE0BSDUgR + return 0 + yq -i eval 'del(.spec.persistentVolumeClaimRetentionPolicy)' /tmp/tmp.i92JTyYaWV/statefulset_some-name-rs0.yml + version_gt 1.22 ++ echo '1.30 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + yq -i eval 'del(.spec.internalTrafficPolicy)' /tmp/tmp.i92JTyYaWV/statefulset_some-name-rs0.yml + yq -i eval 'del(.spec.allocateLoadBalancerNodePorts)' /tmp/tmp.i92JTyYaWV/statefulset_some-name-rs0.yml + [[ /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/e2e-tests/pvc-resize/compare/statefulset_some-name-rs0.yml == */cronjob* ]] + '[' -n '' ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/e2e-tests/pvc-resize/compare/statefulset_some-name-rs0.yml /tmp/tmp.i92JTyYaWV/statefulset_some-name-rs0.yml + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + run_mongo 'db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})' userAdmin:userAdmin123456@some-name-rs0.pvc-resize-28274 + local 'command=db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})' + local uri=userAdmin:userAdmin123456@some-name-rs0.pvc-resize-28274 + local driver=mongodb+srv + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9D1TwVykRZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.BwHYVwKwM8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9D1TwVykRZ ++ cat /tmp/tmp.BwHYVwKwM8 ++ rm /tmp/tmp.9D1TwVykRZ /tmp/tmp.BwHYVwKwM8 ++ return 0 + local client_container=psmdb-client-66f577db5f-pbbww + local mongo_flag= + [[ userAdmin:userAdmin123456@some-name-rs0.pvc-resize-28274 == *cfg* ]] + replica_set=rs0 + kubectl_bin exec psmdb-client-66f577db5f-pbbww -- bash -c 'printf '\''db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})\n'\'' | mongo mongodb+srv://userAdmin:userAdmin123456@some-name-rs0.pvc-resize-28274.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.kE9CHpi6J2 ++ mktemp + local LAST_ERR=/tmp/tmp.LPBi3y3pTW + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-66f577db5f-pbbww -- bash -c 'printf '\''db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})\n'\'' | mongo mongodb+srv://userAdmin:userAdmin123456@some-name-rs0.pvc-resize-28274.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.kE9CHpi6J2 Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-rs0-2.some-name-rs0.pvc-resize-28274.svc.cluster.local:27017,some-name-rs0-1.some-name-rs0.pvc-resize-28274.svc.cluster.local:27017,some-name-rs0-0.some-name-rs0.pvc-resize-28274.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb&replicaSet=rs0&ssl=false Implicit session: session { "id" : UUID("44be7709-3ddd-439a-b4a1-f0f7b590bd63") } Percona Server for MongoDB server version: v7.0.18-11 WARNING: shell and server versions do not match Successfully added user: { "user" : "myApp", "roles" : [ { "db" : "myApp", "role" : "readWrite" } ] } bye + cat /tmp/tmp.LPBi3y3pTW + rm /tmp/tmp.kE9CHpi6J2 /tmp/tmp.LPBi3y3pTW + return 0 + run_mongo 'use myApp\n db.test.insert({ x: 100500 })' myApp:myPass@some-name-rs0.pvc-resize-28274 + local 'command=use myApp\n db.test.insert({ x: 100500 })' + local uri=myApp:myPass@some-name-rs0.pvc-resize-28274 + local driver=mongodb+srv + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oFYVAkDoiB +++ mktemp ++ local LAST_ERR=/tmp/tmp.Q1I5BUeYnn ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.oFYVAkDoiB ++ cat /tmp/tmp.Q1I5BUeYnn ++ rm /tmp/tmp.oFYVAkDoiB /tmp/tmp.Q1I5BUeYnn ++ return 0 + local client_container=psmdb-client-66f577db5f-pbbww + local mongo_flag= + [[ myApp:myPass@some-name-rs0.pvc-resize-28274 == *cfg* ]] + replica_set=rs0 + kubectl_bin exec psmdb-client-66f577db5f-pbbww -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb+srv://myApp:myPass@some-name-rs0.pvc-resize-28274.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.CbRLGBhyUR ++ mktemp + local LAST_ERR=/tmp/tmp.VORDGlg7Y3 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-66f577db5f-pbbww -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb+srv://myApp:myPass@some-name-rs0.pvc-resize-28274.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.CbRLGBhyUR Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-rs0-1.some-name-rs0.pvc-resize-28274.svc.cluster.local:27017,some-name-rs0-0.some-name-rs0.pvc-resize-28274.svc.cluster.local:27017,some-name-rs0-2.some-name-rs0.pvc-resize-28274.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb&replicaSet=rs0&ssl=false Implicit session: session { "id" : UUID("db6a9e1f-44c3-415c-86b2-7f0de5131a18") } Percona Server for MongoDB server version: v7.0.18-11 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.VORDGlg7Y3 + rm /tmp/tmp.CbRLGBhyUR /tmp/tmp.VORDGlg7Y3 + return 0 + patch_pvc_request some-name 2G + local cluster=some-name + local size=2G + echo 'Patching PVC request to 2G in some-name' Patching PVC request to 2G in some-name + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"2G"}]' ++ mktemp + local LAST_OUT=/tmp/tmp.0tuqwvb81s ++ mktemp + local LAST_ERR=/tmp/tmp.B9wxBNQepI + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"2G"}]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.0tuqwvb81s perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.B9wxBNQepI + rm /tmp/tmp.0tuqwvb81s /tmp/tmp.B9wxBNQepI + return 0 + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vTBMYmn8GF +++ mktemp ++ local LAST_ERR=/tmp/tmp.kmpXLkrYpw ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.vTBMYmn8GF ++ cat /tmp/tmp.kmpXLkrYpw ++ rm /tmp/tmp.vTBMYmn8GF /tmp/tmp.kmpXLkrYpw ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo + echo + wait_all_pvc_resize 2Gi 120 1 + local expected_size=2Gi + local max_retry=120 + local sleep_time=1 ++ kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name +++ mktemp ++ local LAST_OUT=/tmp/tmp.fRPer2cMlF +++ mktemp ++ local LAST_ERR=/tmp/tmp.G6ybrXXQG9 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pvc -l app.kubernetes.io/component=mongod -o name ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.fRPer2cMlF ++ cat /tmp/tmp.G6ybrXXQG9 ++ rm /tmp/tmp.fRPer2cMlF /tmp/tmp.G6ybrXXQG9 ++ return 0 + for pvc in '$(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name)' + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-0 2Gi 120 1 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-0 + local expected_size=2Gi + local max_retry=120 + local sleep_time=1 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1ev7QD9Y5d +++ mktemp ++ local LAST_ERR=/tmp/tmp.bCwwp3cBCv ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.1ev7QD9Y5d ++ cat /tmp/tmp.bCwwp3cBCv ++ rm /tmp/tmp.1ev7QD9Y5d /tmp/tmp.bCwwp3cBCv ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 0 -ge 120 ]] + echo -n . .+ sleep 1 + retry=1 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KnKI5RHB5g +++ mktemp ++ local LAST_ERR=/tmp/tmp.qDcEKn6M3M ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.KnKI5RHB5g ++ cat /tmp/tmp.qDcEKn6M3M ++ rm /tmp/tmp.KnKI5RHB5g /tmp/tmp.qDcEKn6M3M ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 1 -ge 120 ]] + echo -n . .+ sleep 1 + retry=2 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EgPQ5Jvy9w +++ mktemp ++ local LAST_ERR=/tmp/tmp.Gbj9IqUySt ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.EgPQ5Jvy9w ++ cat /tmp/tmp.Gbj9IqUySt ++ rm /tmp/tmp.EgPQ5Jvy9w /tmp/tmp.Gbj9IqUySt ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 2 -ge 120 ]] + echo -n . .+ sleep 1 + retry=3 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gnprkTiU5H +++ mktemp ++ local LAST_ERR=/tmp/tmp.hsmR23q6Wc ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.gnprkTiU5H ++ cat /tmp/tmp.hsmR23q6Wc ++ rm /tmp/tmp.gnprkTiU5H /tmp/tmp.hsmR23q6Wc ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 3 -ge 120 ]] + echo -n . .+ sleep 1 + retry=4 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.663tafEbYc +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZYegvUWzIG ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.663tafEbYc ++ cat /tmp/tmp.ZYegvUWzIG ++ rm /tmp/tmp.663tafEbYc /tmp/tmp.ZYegvUWzIG ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 4 -ge 120 ]] + echo -n . .+ sleep 1 + retry=5 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.q6VlGvPmvf +++ mktemp ++ local LAST_ERR=/tmp/tmp.ntmYpVsIfs ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.q6VlGvPmvf ++ cat /tmp/tmp.ntmYpVsIfs ++ rm /tmp/tmp.q6VlGvPmvf /tmp/tmp.ntmYpVsIfs ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 5 -ge 120 ]] + echo -n . .+ sleep 1 + retry=6 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hgBWG3J8tV +++ mktemp ++ local LAST_ERR=/tmp/tmp.HxwGMa5Rxi ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.hgBWG3J8tV ++ cat /tmp/tmp.HxwGMa5Rxi ++ rm /tmp/tmp.hgBWG3J8tV /tmp/tmp.HxwGMa5Rxi ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 6 -ge 120 ]] + echo -n . .+ sleep 1 + retry=7 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JcHokurJIo +++ mktemp ++ local LAST_ERR=/tmp/tmp.LhHTwKEbnC ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.JcHokurJIo ++ cat /tmp/tmp.LhHTwKEbnC ++ rm /tmp/tmp.JcHokurJIo /tmp/tmp.LhHTwKEbnC ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 7 -ge 120 ]] + echo -n . .+ sleep 1 + retry=8 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jj3pZy8xco +++ mktemp ++ local LAST_ERR=/tmp/tmp.9yrnHsLNNe ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jj3pZy8xco ++ cat /tmp/tmp.9yrnHsLNNe ++ rm /tmp/tmp.jj3pZy8xco /tmp/tmp.9yrnHsLNNe ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 8 -ge 120 ]] + echo -n . .+ sleep 1 + retry=9 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uhkzTaHuWz +++ mktemp ++ local LAST_ERR=/tmp/tmp.AQkKOyazi9 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.uhkzTaHuWz ++ cat /tmp/tmp.AQkKOyazi9 ++ rm /tmp/tmp.uhkzTaHuWz /tmp/tmp.AQkKOyazi9 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 9 -ge 120 ]] + echo -n . .+ sleep 1 + retry=10 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ngE1yYqP57 +++ mktemp ++ local LAST_ERR=/tmp/tmp.KdFwaHVA2u ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ngE1yYqP57 ++ cat /tmp/tmp.KdFwaHVA2u ++ rm /tmp/tmp.ngE1yYqP57 /tmp/tmp.KdFwaHVA2u ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 10 -ge 120 ]] + echo -n . .+ sleep 1 + retry=11 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jvCN0Lv2jv +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZDsR6oBn6E ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jvCN0Lv2jv ++ cat /tmp/tmp.ZDsR6oBn6E ++ rm /tmp/tmp.jvCN0Lv2jv /tmp/tmp.ZDsR6oBn6E ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 11 -ge 120 ]] + echo -n . .+ sleep 1 + retry=12 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9TPllgvzSA +++ mktemp ++ local LAST_ERR=/tmp/tmp.RFisfhpqmc ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9TPllgvzSA ++ cat /tmp/tmp.RFisfhpqmc ++ rm /tmp/tmp.9TPllgvzSA /tmp/tmp.RFisfhpqmc ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 12 -ge 120 ]] + echo -n . .+ sleep 1 + retry=13 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.t1HU9HUNjM +++ mktemp ++ local LAST_ERR=/tmp/tmp.fhzF5tHz0k ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.t1HU9HUNjM ++ cat /tmp/tmp.fhzF5tHz0k ++ rm /tmp/tmp.t1HU9HUNjM /tmp/tmp.fhzF5tHz0k ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 13 -ge 120 ]] + echo -n . .+ sleep 1 + retry=14 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FVPSd1DNb8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.L5ZcVgnI4f ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.FVPSd1DNb8 ++ cat /tmp/tmp.L5ZcVgnI4f ++ rm /tmp/tmp.FVPSd1DNb8 /tmp/tmp.L5ZcVgnI4f ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 14 -ge 120 ]] + echo -n . .+ sleep 1 + retry=15 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4wx71usUWZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.oYprlKPQq6 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4wx71usUWZ ++ cat /tmp/tmp.oYprlKPQq6 ++ rm /tmp/tmp.4wx71usUWZ /tmp/tmp.oYprlKPQq6 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 15 -ge 120 ]] + echo -n . .+ sleep 1 + retry=16 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7xjIAVjqeo +++ mktemp ++ local LAST_ERR=/tmp/tmp.3P6fHcGXWh ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.7xjIAVjqeo ++ cat /tmp/tmp.3P6fHcGXWh ++ rm /tmp/tmp.7xjIAVjqeo /tmp/tmp.3P6fHcGXWh ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 16 -ge 120 ]] + echo -n . .+ sleep 1 + retry=17 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MLQ8E1j7AR +++ mktemp ++ local LAST_ERR=/tmp/tmp.1VAY1K1dYk ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.MLQ8E1j7AR ++ cat /tmp/tmp.1VAY1K1dYk ++ rm /tmp/tmp.MLQ8E1j7AR /tmp/tmp.1VAY1K1dYk ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 17 -ge 120 ]] + echo -n . .+ sleep 1 + retry=18 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.U99QmdgVtH +++ mktemp ++ local LAST_ERR=/tmp/tmp.NzJz27I8tT ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.U99QmdgVtH ++ cat /tmp/tmp.NzJz27I8tT ++ rm /tmp/tmp.U99QmdgVtH /tmp/tmp.NzJz27I8tT ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 18 -ge 120 ]] + echo -n . .+ sleep 1 + retry=19 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9Sr1TLN4Sf +++ mktemp ++ local LAST_ERR=/tmp/tmp.jrxI2eosio ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9Sr1TLN4Sf ++ cat /tmp/tmp.jrxI2eosio ++ rm /tmp/tmp.9Sr1TLN4Sf /tmp/tmp.jrxI2eosio ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 19 -ge 120 ]] + echo -n . .+ sleep 1 + retry=20 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.CJElKLdYQo +++ mktemp ++ local LAST_ERR=/tmp/tmp.JMiicXnraV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.CJElKLdYQo ++ cat /tmp/tmp.JMiicXnraV ++ rm /tmp/tmp.CJElKLdYQo /tmp/tmp.JMiicXnraV ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 20 -ge 120 ]] + echo -n . .+ sleep 1 + retry=21 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tSL0oQky9c +++ mktemp ++ local LAST_ERR=/tmp/tmp.pSlnhwROiy ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.tSL0oQky9c ++ cat /tmp/tmp.pSlnhwROiy ++ rm /tmp/tmp.tSL0oQky9c /tmp/tmp.pSlnhwROiy ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 21 -ge 120 ]] + echo -n . .+ sleep 1 + retry=22 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hQJ5ujHdMD +++ mktemp ++ local LAST_ERR=/tmp/tmp.TyGs4esrsf ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.hQJ5ujHdMD ++ cat /tmp/tmp.TyGs4esrsf ++ rm /tmp/tmp.hQJ5ujHdMD /tmp/tmp.TyGs4esrsf ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 22 -ge 120 ]] + echo -n . .+ sleep 1 + retry=23 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.j4iTL50XeP +++ mktemp ++ local LAST_ERR=/tmp/tmp.VjaP9xIYTp ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.j4iTL50XeP ++ cat /tmp/tmp.VjaP9xIYTp ++ rm /tmp/tmp.j4iTL50XeP /tmp/tmp.VjaP9xIYTp ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 23 -ge 120 ]] + echo -n . .+ sleep 1 + retry=24 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jBdckWkZAD +++ mktemp ++ local LAST_ERR=/tmp/tmp.XtWPlw1uEn ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jBdckWkZAD ++ cat /tmp/tmp.XtWPlw1uEn ++ rm /tmp/tmp.jBdckWkZAD /tmp/tmp.XtWPlw1uEn ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 24 -ge 120 ]] + echo -n . .+ sleep 1 + retry=25 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ot0Zo0ymmY +++ mktemp ++ local LAST_ERR=/tmp/tmp.auSBrU64j0 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ot0Zo0ymmY ++ cat /tmp/tmp.auSBrU64j0 ++ rm /tmp/tmp.ot0Zo0ymmY /tmp/tmp.auSBrU64j0 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 25 -ge 120 ]] + echo -n . .+ sleep 1 + retry=26 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GDz2034o8G +++ mktemp ++ local LAST_ERR=/tmp/tmp.WDOnOpkEsY ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.GDz2034o8G ++ cat /tmp/tmp.WDOnOpkEsY ++ rm /tmp/tmp.GDz2034o8G /tmp/tmp.WDOnOpkEsY ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 26 -ge 120 ]] + echo -n . .+ sleep 1 + retry=27 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.paFWWjQVhH +++ mktemp ++ local LAST_ERR=/tmp/tmp.1SeGPrUnJf ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.paFWWjQVhH ++ cat /tmp/tmp.1SeGPrUnJf ++ rm /tmp/tmp.paFWWjQVhH /tmp/tmp.1SeGPrUnJf ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 27 -ge 120 ]] + echo -n . .+ sleep 1 + retry=28 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FtycyJYsim +++ mktemp ++ local LAST_ERR=/tmp/tmp.C3W23u6W0T ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.FtycyJYsim ++ cat /tmp/tmp.C3W23u6W0T ++ rm /tmp/tmp.FtycyJYsim /tmp/tmp.C3W23u6W0T ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 28 -ge 120 ]] + echo -n . .+ sleep 1 + retry=29 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ecbPfH9Yci +++ mktemp ++ local LAST_ERR=/tmp/tmp.IyHCpgYJoN ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ecbPfH9Yci ++ cat /tmp/tmp.IyHCpgYJoN ++ rm /tmp/tmp.ecbPfH9Yci /tmp/tmp.IyHCpgYJoN ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 29 -ge 120 ]] + echo -n . .+ sleep 1 + retry=30 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Fa1GmUBg5C +++ mktemp ++ local LAST_ERR=/tmp/tmp.41XXmTllno ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Fa1GmUBg5C ++ cat /tmp/tmp.41XXmTllno ++ rm /tmp/tmp.Fa1GmUBg5C /tmp/tmp.41XXmTllno ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 30 -ge 120 ]] + echo -n . .+ sleep 1 + retry=31 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YAMNAHhg6e +++ mktemp ++ local LAST_ERR=/tmp/tmp.186FPyAGNu ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.YAMNAHhg6e ++ cat /tmp/tmp.186FPyAGNu ++ rm /tmp/tmp.YAMNAHhg6e /tmp/tmp.186FPyAGNu ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 31 -ge 120 ]] + echo -n . .+ sleep 1 + retry=32 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WO4SLS4ErD +++ mktemp ++ local LAST_ERR=/tmp/tmp.UbUDp2N4Zd ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.WO4SLS4ErD ++ cat /tmp/tmp.UbUDp2N4Zd ++ rm /tmp/tmp.WO4SLS4ErD /tmp/tmp.UbUDp2N4Zd ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 32 -ge 120 ]] + echo -n . .+ sleep 1 + retry=33 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kGgiXhhJep +++ mktemp ++ local LAST_ERR=/tmp/tmp.InJrxgk5mJ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.kGgiXhhJep ++ cat /tmp/tmp.InJrxgk5mJ ++ rm /tmp/tmp.kGgiXhhJep /tmp/tmp.InJrxgk5mJ ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 33 -ge 120 ]] + echo -n . .+ sleep 1 + retry=34 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.73O9vBenUl +++ mktemp ++ local LAST_ERR=/tmp/tmp.rGWoGI2cxD ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.73O9vBenUl ++ cat /tmp/tmp.rGWoGI2cxD ++ rm /tmp/tmp.73O9vBenUl /tmp/tmp.rGWoGI2cxD ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 34 -ge 120 ]] + echo -n . .+ sleep 1 + retry=35 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DW27SSKLIT +++ mktemp ++ local LAST_ERR=/tmp/tmp.OsKdV5iWcD ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.DW27SSKLIT ++ cat /tmp/tmp.OsKdV5iWcD ++ rm /tmp/tmp.DW27SSKLIT /tmp/tmp.OsKdV5iWcD ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 35 -ge 120 ]] + echo -n . .+ sleep 1 + retry=36 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7v9gsOZ73J +++ mktemp ++ local LAST_ERR=/tmp/tmp.ecndAFSLhL ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.7v9gsOZ73J ++ cat /tmp/tmp.ecndAFSLhL ++ rm /tmp/tmp.7v9gsOZ73J /tmp/tmp.ecndAFSLhL ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 36 -ge 120 ]] + echo -n . .+ sleep 1 + retry=37 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ysTpxCcaPq +++ mktemp ++ local LAST_ERR=/tmp/tmp.WOiXKKvzgt ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ysTpxCcaPq ++ cat /tmp/tmp.WOiXKKvzgt ++ rm /tmp/tmp.ysTpxCcaPq /tmp/tmp.WOiXKKvzgt ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 37 -ge 120 ]] + echo -n . .+ sleep 1 + retry=38 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VDXwP46VdJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.IXFWQdPDPB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.VDXwP46VdJ ++ cat /tmp/tmp.IXFWQdPDPB ++ rm /tmp/tmp.VDXwP46VdJ /tmp/tmp.IXFWQdPDPB ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 38 -ge 120 ]] + echo -n . .+ sleep 1 + retry=39 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yXq7C5nrZq +++ mktemp ++ local LAST_ERR=/tmp/tmp.vwdxRLOwi3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.yXq7C5nrZq ++ cat /tmp/tmp.vwdxRLOwi3 ++ rm /tmp/tmp.yXq7C5nrZq /tmp/tmp.vwdxRLOwi3 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 39 -ge 120 ]] + echo -n . .+ sleep 1 + retry=40 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.COqMycAg0M +++ mktemp ++ local LAST_ERR=/tmp/tmp.HdjbwUvfAF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.COqMycAg0M ++ cat /tmp/tmp.HdjbwUvfAF ++ rm /tmp/tmp.COqMycAg0M /tmp/tmp.HdjbwUvfAF ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 40 -ge 120 ]] + echo -n . .+ sleep 1 + retry=41 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OjDUAOUApJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.72KCMy0VeN ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.OjDUAOUApJ ++ cat /tmp/tmp.72KCMy0VeN ++ rm /tmp/tmp.OjDUAOUApJ /tmp/tmp.72KCMy0VeN ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 41 -ge 120 ]] + echo -n . .+ sleep 1 + retry=42 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.h7nWx5thKp +++ mktemp ++ local LAST_ERR=/tmp/tmp.bzv09Ozh6B ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.h7nWx5thKp ++ cat /tmp/tmp.bzv09Ozh6B ++ rm /tmp/tmp.h7nWx5thKp /tmp/tmp.bzv09Ozh6B ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 42 -ge 120 ]] + echo -n . .+ sleep 1 + retry=43 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OMjKIV74cs +++ mktemp ++ local LAST_ERR=/tmp/tmp.01yTpWXJcF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.OMjKIV74cs ++ cat /tmp/tmp.01yTpWXJcF ++ rm /tmp/tmp.OMjKIV74cs /tmp/tmp.01yTpWXJcF ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 43 -ge 120 ]] + echo -n . .+ sleep 1 + retry=44 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZNKNJ6XpKE +++ mktemp ++ local LAST_ERR=/tmp/tmp.6jk305KCtv ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ZNKNJ6XpKE ++ cat /tmp/tmp.6jk305KCtv ++ rm /tmp/tmp.ZNKNJ6XpKE /tmp/tmp.6jk305KCtv ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 44 -ge 120 ]] + echo -n . .+ sleep 1 + retry=45 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yBEuJlzOcG +++ mktemp ++ local LAST_ERR=/tmp/tmp.9HCCBezjM8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.yBEuJlzOcG ++ cat /tmp/tmp.9HCCBezjM8 ++ rm /tmp/tmp.yBEuJlzOcG /tmp/tmp.9HCCBezjM8 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 45 -ge 120 ]] + echo -n . .+ sleep 1 + retry=46 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6coRVebhGy +++ mktemp ++ local LAST_ERR=/tmp/tmp.geH3IO00Fo ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.6coRVebhGy ++ cat /tmp/tmp.geH3IO00Fo ++ rm /tmp/tmp.6coRVebhGy /tmp/tmp.geH3IO00Fo ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 46 -ge 120 ]] + echo -n . .+ sleep 1 + retry=47 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.22Xuj934t4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZJRDDjU6a2 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.22Xuj934t4 ++ cat /tmp/tmp.ZJRDDjU6a2 ++ rm /tmp/tmp.22Xuj934t4 /tmp/tmp.ZJRDDjU6a2 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 47 -ge 120 ]] + echo -n . .+ sleep 1 + retry=48 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uYUbAMVOEO +++ mktemp ++ local LAST_ERR=/tmp/tmp.Y25boFjkdR ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.uYUbAMVOEO ++ cat /tmp/tmp.Y25boFjkdR ++ rm /tmp/tmp.uYUbAMVOEO /tmp/tmp.Y25boFjkdR ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 48 -ge 120 ]] + echo -n . .+ sleep 1 + retry=49 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.d6ifutYBam +++ mktemp ++ local LAST_ERR=/tmp/tmp.d7vwUtoktD ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.d6ifutYBam ++ cat /tmp/tmp.d7vwUtoktD ++ rm /tmp/tmp.d6ifutYBam /tmp/tmp.d7vwUtoktD ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 49 -ge 120 ]] + echo -n . .+ sleep 1 + retry=50 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7fuowVnWDd +++ mktemp ++ local LAST_ERR=/tmp/tmp.Wbnoz9yT1t ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.7fuowVnWDd ++ cat /tmp/tmp.Wbnoz9yT1t ++ rm /tmp/tmp.7fuowVnWDd /tmp/tmp.Wbnoz9yT1t ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 50 -ge 120 ]] + echo -n . .+ sleep 1 + retry=51 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EtpadoygJZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.5nM41LEUpy ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.EtpadoygJZ ++ cat /tmp/tmp.5nM41LEUpy ++ rm /tmp/tmp.EtpadoygJZ /tmp/tmp.5nM41LEUpy ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 51 -ge 120 ]] + echo -n . .+ sleep 1 + retry=52 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RcejmBGiMb +++ mktemp ++ local LAST_ERR=/tmp/tmp.HTb5FE3f1u ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.RcejmBGiMb ++ cat /tmp/tmp.HTb5FE3f1u ++ rm /tmp/tmp.RcejmBGiMb /tmp/tmp.HTb5FE3f1u ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 52 -ge 120 ]] + echo -n . .+ sleep 1 + retry=53 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.90nYp0tscT +++ mktemp ++ local LAST_ERR=/tmp/tmp.WY1rhEbDp1 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.90nYp0tscT ++ cat /tmp/tmp.WY1rhEbDp1 ++ rm /tmp/tmp.90nYp0tscT /tmp/tmp.WY1rhEbDp1 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 53 -ge 120 ]] + echo -n . .+ sleep 1 + retry=54 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PIpDobJMUR +++ mktemp ++ local LAST_ERR=/tmp/tmp.Luktn0KksT ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.PIpDobJMUR ++ cat /tmp/tmp.Luktn0KksT ++ rm /tmp/tmp.PIpDobJMUR /tmp/tmp.Luktn0KksT ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 54 -ge 120 ]] + echo -n . .+ sleep 1 + retry=55 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TrAAMs2RLy +++ mktemp ++ local LAST_ERR=/tmp/tmp.GDS67hhxIl ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.TrAAMs2RLy ++ cat /tmp/tmp.GDS67hhxIl ++ rm /tmp/tmp.TrAAMs2RLy /tmp/tmp.GDS67hhxIl ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 55 -ge 120 ]] + echo -n . .+ sleep 1 + retry=56 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AP7EHEeiFY +++ mktemp ++ local LAST_ERR=/tmp/tmp.McjEWol1FJ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.AP7EHEeiFY ++ cat /tmp/tmp.McjEWol1FJ ++ rm /tmp/tmp.AP7EHEeiFY /tmp/tmp.McjEWol1FJ ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 56 -ge 120 ]] + echo -n . .+ sleep 1 + retry=57 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BLE6qWf22X +++ mktemp ++ local LAST_ERR=/tmp/tmp.RfNp6T4iXd ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.BLE6qWf22X ++ cat /tmp/tmp.RfNp6T4iXd ++ rm /tmp/tmp.BLE6qWf22X /tmp/tmp.RfNp6T4iXd ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 57 -ge 120 ]] + echo -n . .+ sleep 1 + retry=58 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VQ3O26EzHG +++ mktemp ++ local LAST_ERR=/tmp/tmp.4DbYC2ovJW ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.VQ3O26EzHG ++ cat /tmp/tmp.4DbYC2ovJW ++ rm /tmp/tmp.VQ3O26EzHG /tmp/tmp.4DbYC2ovJW ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 58 -ge 120 ]] + echo -n . .+ sleep 1 + retry=59 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vApHFy4B9f +++ mktemp ++ local LAST_ERR=/tmp/tmp.6I6M1cfgSE ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.vApHFy4B9f ++ cat /tmp/tmp.6I6M1cfgSE ++ rm /tmp/tmp.vApHFy4B9f /tmp/tmp.6I6M1cfgSE ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 59 -ge 120 ]] + echo -n . .+ sleep 1 + retry=60 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UmEQWtByfu +++ mktemp ++ local LAST_ERR=/tmp/tmp.DEP9E98ifr ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.UmEQWtByfu ++ cat /tmp/tmp.DEP9E98ifr ++ rm /tmp/tmp.UmEQWtByfu /tmp/tmp.DEP9E98ifr ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 60 -ge 120 ]] + echo -n . .+ sleep 1 + retry=61 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9QkgqbnTE1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZSslnk1YDb ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9QkgqbnTE1 ++ cat /tmp/tmp.ZSslnk1YDb ++ rm /tmp/tmp.9QkgqbnTE1 /tmp/tmp.ZSslnk1YDb ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 61 -ge 120 ]] + echo -n . .+ sleep 1 + retry=62 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.z4XcFpjVft +++ mktemp ++ local LAST_ERR=/tmp/tmp.yIseXwebNQ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.z4XcFpjVft ++ cat /tmp/tmp.yIseXwebNQ ++ rm /tmp/tmp.z4XcFpjVft /tmp/tmp.yIseXwebNQ ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 62 -ge 120 ]] + echo -n . .+ sleep 1 + retry=63 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Xl8nBUgbRX +++ mktemp ++ local LAST_ERR=/tmp/tmp.nMyj08Zt0r ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Xl8nBUgbRX ++ cat /tmp/tmp.nMyj08Zt0r ++ rm /tmp/tmp.Xl8nBUgbRX /tmp/tmp.nMyj08Zt0r ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 63 -ge 120 ]] + echo -n . .+ sleep 1 + retry=64 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qgZTt6rlAU +++ mktemp ++ local LAST_ERR=/tmp/tmp.1tzrhwNlbF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.qgZTt6rlAU ++ cat /tmp/tmp.1tzrhwNlbF ++ rm /tmp/tmp.qgZTt6rlAU /tmp/tmp.1tzrhwNlbF ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 64 -ge 120 ]] + echo -n . .+ sleep 1 + retry=65 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.88COC6od5Y +++ mktemp ++ local LAST_ERR=/tmp/tmp.2kffeG36EP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.88COC6od5Y ++ cat /tmp/tmp.2kffeG36EP ++ rm /tmp/tmp.88COC6od5Y /tmp/tmp.2kffeG36EP ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 65 -ge 120 ]] + echo -n . .+ sleep 1 + retry=66 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TIPuEaUdoi +++ mktemp ++ local LAST_ERR=/tmp/tmp.sJIIFPWdM5 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.TIPuEaUdoi ++ cat /tmp/tmp.sJIIFPWdM5 ++ rm /tmp/tmp.TIPuEaUdoi /tmp/tmp.sJIIFPWdM5 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 66 -ge 120 ]] + echo -n . .+ sleep 1 + retry=67 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5H4vye5rf0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.LUzfgggoKf ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5H4vye5rf0 ++ cat /tmp/tmp.LUzfgggoKf ++ rm /tmp/tmp.5H4vye5rf0 /tmp/tmp.LUzfgggoKf ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 67 -ge 120 ]] + echo -n . .+ sleep 1 + retry=68 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YAq7xA8slH +++ mktemp ++ local LAST_ERR=/tmp/tmp.tmQxjE8BG8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.YAq7xA8slH ++ cat /tmp/tmp.tmQxjE8BG8 ++ rm /tmp/tmp.YAq7xA8slH /tmp/tmp.tmQxjE8BG8 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 68 -ge 120 ]] + echo -n . .+ sleep 1 + retry=69 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nedR409rcY +++ mktemp ++ local LAST_ERR=/tmp/tmp.jyzqOumsu8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.nedR409rcY ++ cat /tmp/tmp.jyzqOumsu8 ++ rm /tmp/tmp.nedR409rcY /tmp/tmp.jyzqOumsu8 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 69 -ge 120 ]] + echo -n . .+ sleep 1 + retry=70 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TogYkWEqt3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.PDheSbtp5V ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.TogYkWEqt3 ++ cat /tmp/tmp.PDheSbtp5V ++ rm /tmp/tmp.TogYkWEqt3 /tmp/tmp.PDheSbtp5V ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 70 -ge 120 ]] + echo -n . .+ sleep 1 + retry=71 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.92xjia2NLf +++ mktemp ++ local LAST_ERR=/tmp/tmp.knUQXcVCGq ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.92xjia2NLf ++ cat /tmp/tmp.knUQXcVCGq ++ rm /tmp/tmp.92xjia2NLf /tmp/tmp.knUQXcVCGq ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 71 -ge 120 ]] + echo -n . .+ sleep 1 + retry=72 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OYb6c5CEDk +++ mktemp ++ local LAST_ERR=/tmp/tmp.OvrTHC7ObA ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.OYb6c5CEDk ++ cat /tmp/tmp.OvrTHC7ObA ++ rm /tmp/tmp.OYb6c5CEDk /tmp/tmp.OvrTHC7ObA ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 72 -ge 120 ]] + echo -n . .+ sleep 1 + retry=73 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Qz5QD2jt6X +++ mktemp ++ local LAST_ERR=/tmp/tmp.qBpyuFCDeC ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Qz5QD2jt6X ++ cat /tmp/tmp.qBpyuFCDeC ++ rm /tmp/tmp.Qz5QD2jt6X /tmp/tmp.qBpyuFCDeC ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 73 -ge 120 ]] + echo -n . .+ sleep 1 + retry=74 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.x0guXikwu8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.0rMWelXZtB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.x0guXikwu8 ++ cat /tmp/tmp.0rMWelXZtB ++ rm /tmp/tmp.x0guXikwu8 /tmp/tmp.0rMWelXZtB ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 74 -ge 120 ]] + echo -n . .+ sleep 1 + retry=75 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sZ2VEeQhJH +++ mktemp ++ local LAST_ERR=/tmp/tmp.UiLe481ZhY ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.sZ2VEeQhJH ++ cat /tmp/tmp.UiLe481ZhY ++ rm /tmp/tmp.sZ2VEeQhJH /tmp/tmp.UiLe481ZhY ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 75 -ge 120 ]] + echo -n . .+ sleep 1 + retry=76 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3GAQqgSILH +++ mktemp ++ local LAST_ERR=/tmp/tmp.kTThdJBWOY ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.3GAQqgSILH ++ cat /tmp/tmp.kTThdJBWOY ++ rm /tmp/tmp.3GAQqgSILH /tmp/tmp.kTThdJBWOY ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 76 -ge 120 ]] + echo -n . .+ sleep 1 + retry=77 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pRuPRIsnEC +++ mktemp ++ local LAST_ERR=/tmp/tmp.srCjXjMyWW ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.pRuPRIsnEC ++ cat /tmp/tmp.srCjXjMyWW ++ rm /tmp/tmp.pRuPRIsnEC /tmp/tmp.srCjXjMyWW ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 77 -ge 120 ]] + echo -n . .+ sleep 1 + retry=78 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RhpfhHTWKi +++ mktemp ++ local LAST_ERR=/tmp/tmp.KX16AacFZd ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.RhpfhHTWKi ++ cat /tmp/tmp.KX16AacFZd ++ rm /tmp/tmp.RhpfhHTWKi /tmp/tmp.KX16AacFZd ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 78 -ge 120 ]] + echo -n . .+ sleep 1 + retry=79 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Vp9pizbqOu +++ mktemp ++ local LAST_ERR=/tmp/tmp.IaK2NX7dEx ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Vp9pizbqOu ++ cat /tmp/tmp.IaK2NX7dEx ++ rm /tmp/tmp.Vp9pizbqOu /tmp/tmp.IaK2NX7dEx ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 79 -ge 120 ]] + echo -n . .+ sleep 1 + retry=80 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gpe1s1uIfk +++ mktemp ++ local LAST_ERR=/tmp/tmp.K88i8tdM3I ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.gpe1s1uIfk ++ cat /tmp/tmp.K88i8tdM3I ++ rm /tmp/tmp.gpe1s1uIfk /tmp/tmp.K88i8tdM3I ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 80 -ge 120 ]] + echo -n . .+ sleep 1 + retry=81 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.itY2dNFeCK +++ mktemp ++ local LAST_ERR=/tmp/tmp.rXV1AdU4LP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.itY2dNFeCK ++ cat /tmp/tmp.rXV1AdU4LP ++ rm /tmp/tmp.itY2dNFeCK /tmp/tmp.rXV1AdU4LP ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 81 -ge 120 ]] + echo -n . .+ sleep 1 + retry=82 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Z8dtMITuTy +++ mktemp ++ local LAST_ERR=/tmp/tmp.sf7ChxqF0q ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Z8dtMITuTy ++ cat /tmp/tmp.sf7ChxqF0q ++ rm /tmp/tmp.Z8dtMITuTy /tmp/tmp.sf7ChxqF0q ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 82 -ge 120 ]] + echo -n . .+ sleep 1 + retry=83 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FRhxCj6cxI +++ mktemp ++ local LAST_ERR=/tmp/tmp.NAqd6NfRJ5 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.FRhxCj6cxI ++ cat /tmp/tmp.NAqd6NfRJ5 ++ rm /tmp/tmp.FRhxCj6cxI /tmp/tmp.NAqd6NfRJ5 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 83 -ge 120 ]] + echo -n . .+ sleep 1 + retry=84 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.r088ni18k6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.4qGHIT2cnL ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.r088ni18k6 ++ cat /tmp/tmp.4qGHIT2cnL ++ rm /tmp/tmp.r088ni18k6 /tmp/tmp.4qGHIT2cnL ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 84 -ge 120 ]] + echo -n . .+ sleep 1 + retry=85 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zc5jpedvnO +++ mktemp ++ local LAST_ERR=/tmp/tmp.oXDKRBaDnW ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.zc5jpedvnO ++ cat /tmp/tmp.oXDKRBaDnW ++ rm /tmp/tmp.zc5jpedvnO /tmp/tmp.oXDKRBaDnW ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 85 -ge 120 ]] + echo -n . .+ sleep 1 + retry=86 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.exbeZDT2Di +++ mktemp ++ local LAST_ERR=/tmp/tmp.ArK5s4fTEB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.exbeZDT2Di ++ cat /tmp/tmp.ArK5s4fTEB ++ rm /tmp/tmp.exbeZDT2Di /tmp/tmp.ArK5s4fTEB ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 86 -ge 120 ]] + echo -n . .+ sleep 1 + retry=87 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eEuVii0d8O +++ mktemp ++ local LAST_ERR=/tmp/tmp.OXJkPXLdZB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.eEuVii0d8O ++ cat /tmp/tmp.OXJkPXLdZB ++ rm /tmp/tmp.eEuVii0d8O /tmp/tmp.OXJkPXLdZB ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 87 -ge 120 ]] + echo -n . .+ sleep 1 + retry=88 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.i6CiIjKkAv +++ mktemp ++ local LAST_ERR=/tmp/tmp.T5wckY06zk ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.i6CiIjKkAv ++ cat /tmp/tmp.T5wckY06zk ++ rm /tmp/tmp.i6CiIjKkAv /tmp/tmp.T5wckY06zk ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 88 -ge 120 ]] + echo -n . .+ sleep 1 + retry=89 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Rq50FXEqVx +++ mktemp ++ local LAST_ERR=/tmp/tmp.9WHL71BiON ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Rq50FXEqVx ++ cat /tmp/tmp.9WHL71BiON ++ rm /tmp/tmp.Rq50FXEqVx /tmp/tmp.9WHL71BiON ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 89 -ge 120 ]] + echo -n . .+ sleep 1 + retry=90 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.z2PBUrKiwi +++ mktemp ++ local LAST_ERR=/tmp/tmp.98tkGKK38j ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.z2PBUrKiwi ++ cat /tmp/tmp.98tkGKK38j ++ rm /tmp/tmp.z2PBUrKiwi /tmp/tmp.98tkGKK38j ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 90 -ge 120 ]] + echo -n . .+ sleep 1 + retry=91 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rljiBw2TWA +++ mktemp ++ local LAST_ERR=/tmp/tmp.cLYGGGRlQm ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.rljiBw2TWA ++ cat /tmp/tmp.cLYGGGRlQm ++ rm /tmp/tmp.rljiBw2TWA /tmp/tmp.cLYGGGRlQm ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 91 -ge 120 ]] + echo -n . .+ sleep 1 + retry=92 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TzhrTPaudB +++ mktemp ++ local LAST_ERR=/tmp/tmp.Wk5lu9g3jN ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.TzhrTPaudB ++ cat /tmp/tmp.Wk5lu9g3jN ++ rm /tmp/tmp.TzhrTPaudB /tmp/tmp.Wk5lu9g3jN ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 92 -ge 120 ]] + echo -n . .+ sleep 1 + retry=93 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IXDMMLgmjD +++ mktemp ++ local LAST_ERR=/tmp/tmp.ho7CLjU93h ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.IXDMMLgmjD ++ cat /tmp/tmp.ho7CLjU93h ++ rm /tmp/tmp.IXDMMLgmjD /tmp/tmp.ho7CLjU93h ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 93 -ge 120 ]] + echo -n . .+ sleep 1 + retry=94 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eI6Ot0uRUz +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ud3BYefJcr ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.eI6Ot0uRUz ++ cat /tmp/tmp.Ud3BYefJcr ++ rm /tmp/tmp.eI6Ot0uRUz /tmp/tmp.Ud3BYefJcr ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 94 -ge 120 ]] + echo -n . .+ sleep 1 + retry=95 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qBQZRcxOuw +++ mktemp ++ local LAST_ERR=/tmp/tmp.EDphBkO8j1 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.qBQZRcxOuw ++ cat /tmp/tmp.EDphBkO8j1 ++ rm /tmp/tmp.qBQZRcxOuw /tmp/tmp.EDphBkO8j1 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 95 -ge 120 ]] + echo -n . .+ sleep 1 + retry=96 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fZftOEAWvp +++ mktemp ++ local LAST_ERR=/tmp/tmp.unxPMGZhVQ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.fZftOEAWvp ++ cat /tmp/tmp.unxPMGZhVQ ++ rm /tmp/tmp.fZftOEAWvp /tmp/tmp.unxPMGZhVQ ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 96 -ge 120 ]] + echo -n . .+ sleep 1 + retry=97 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ght92sdDYQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.qFg81Fg8W1 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Ght92sdDYQ ++ cat /tmp/tmp.qFg81Fg8W1 ++ rm /tmp/tmp.Ght92sdDYQ /tmp/tmp.qFg81Fg8W1 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 97 -ge 120 ]] + echo -n . .+ sleep 1 + retry=98 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xwxne4vcrH +++ mktemp ++ local LAST_ERR=/tmp/tmp.rQiIDuwbQe ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xwxne4vcrH ++ cat /tmp/tmp.rQiIDuwbQe ++ rm /tmp/tmp.xwxne4vcrH /tmp/tmp.rQiIDuwbQe ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 98 -ge 120 ]] + echo -n . .+ sleep 1 + retry=99 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FRtCiO4HDH +++ mktemp ++ local LAST_ERR=/tmp/tmp.6lJL3QD3ZC ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.FRtCiO4HDH ++ cat /tmp/tmp.6lJL3QD3ZC ++ rm /tmp/tmp.FRtCiO4HDH /tmp/tmp.6lJL3QD3ZC ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 99 -ge 120 ]] + echo -n . .+ sleep 1 + retry=100 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FvKNvXfawr +++ mktemp ++ local LAST_ERR=/tmp/tmp.K5ZPGGmAPk ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.FvKNvXfawr ++ cat /tmp/tmp.K5ZPGGmAPk ++ rm /tmp/tmp.FvKNvXfawr /tmp/tmp.K5ZPGGmAPk ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 100 -ge 120 ]] + echo -n . .+ sleep 1 + retry=101 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Hl9p8BGS6K +++ mktemp ++ local LAST_ERR=/tmp/tmp.gKDnIA6Jdv ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Hl9p8BGS6K ++ cat /tmp/tmp.gKDnIA6Jdv ++ rm /tmp/tmp.Hl9p8BGS6K /tmp/tmp.gKDnIA6Jdv ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 101 -ge 120 ]] + echo -n . .+ sleep 1 + retry=102 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pOLhVpor3A +++ mktemp ++ local LAST_ERR=/tmp/tmp.yZ6b2GOLAL ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.pOLhVpor3A ++ cat /tmp/tmp.yZ6b2GOLAL ++ rm /tmp/tmp.pOLhVpor3A /tmp/tmp.yZ6b2GOLAL ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 102 -ge 120 ]] + echo -n . .+ sleep 1 + retry=103 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IMn30dNeCJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.SIh3VqzDr5 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.IMn30dNeCJ ++ cat /tmp/tmp.SIh3VqzDr5 ++ rm /tmp/tmp.IMn30dNeCJ /tmp/tmp.SIh3VqzDr5 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 103 -ge 120 ]] + echo -n . .+ sleep 1 + retry=104 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RqOrMkHCGB +++ mktemp ++ local LAST_ERR=/tmp/tmp.GOHMRWnnF0 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.RqOrMkHCGB ++ cat /tmp/tmp.GOHMRWnnF0 ++ rm /tmp/tmp.RqOrMkHCGB /tmp/tmp.GOHMRWnnF0 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 104 -ge 120 ]] + echo -n . .+ sleep 1 + retry=105 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fBifyK9siW +++ mktemp ++ local LAST_ERR=/tmp/tmp.tQARq0XWrS ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.fBifyK9siW ++ cat /tmp/tmp.tQARq0XWrS ++ rm /tmp/tmp.fBifyK9siW /tmp/tmp.tQARq0XWrS ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 105 -ge 120 ]] + echo -n . .+ sleep 1 + retry=106 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LUe6QQK4dB +++ mktemp ++ local LAST_ERR=/tmp/tmp.P94f7VSsO6 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.LUe6QQK4dB ++ cat /tmp/tmp.P94f7VSsO6 ++ rm /tmp/tmp.LUe6QQK4dB /tmp/tmp.P94f7VSsO6 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 106 -ge 120 ]] + echo -n . .+ sleep 1 + retry=107 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4om6JM3W2I +++ mktemp ++ local LAST_ERR=/tmp/tmp.5AqbmCtGXy ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4om6JM3W2I ++ cat /tmp/tmp.5AqbmCtGXy ++ rm /tmp/tmp.4om6JM3W2I /tmp/tmp.5AqbmCtGXy ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 107 -ge 120 ]] + echo -n . .+ sleep 1 + retry=108 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AhupcqGcC8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.6eCHQZMgHG ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.AhupcqGcC8 ++ cat /tmp/tmp.6eCHQZMgHG ++ rm /tmp/tmp.AhupcqGcC8 /tmp/tmp.6eCHQZMgHG ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 108 -ge 120 ]] + echo -n . .+ sleep 1 + retry=109 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.21cnZGYhj2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.QCk0t93a0C ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.21cnZGYhj2 ++ cat /tmp/tmp.QCk0t93a0C ++ rm /tmp/tmp.21cnZGYhj2 /tmp/tmp.QCk0t93a0C ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 109 -ge 120 ]] + echo -n . .+ sleep 1 + retry=110 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.haW3F0N1aV +++ mktemp ++ local LAST_ERR=/tmp/tmp.7D0Hx6Z2zG ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.haW3F0N1aV ++ cat /tmp/tmp.7D0Hx6Z2zG ++ rm /tmp/tmp.haW3F0N1aV /tmp/tmp.7D0Hx6Z2zG ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 110 -ge 120 ]] + echo -n . .+ sleep 1 + retry=111 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7YJ38mWJAm +++ mktemp ++ local LAST_ERR=/tmp/tmp.y8UYuZL5IV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.7YJ38mWJAm ++ cat /tmp/tmp.y8UYuZL5IV ++ rm /tmp/tmp.7YJ38mWJAm /tmp/tmp.y8UYuZL5IV ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 111 -ge 120 ]] + echo -n . .+ sleep 1 + retry=112 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Wo3SEqt1vO +++ mktemp ++ local LAST_ERR=/tmp/tmp.zhiFaCVNj3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Wo3SEqt1vO ++ cat /tmp/tmp.zhiFaCVNj3 ++ rm /tmp/tmp.Wo3SEqt1vO /tmp/tmp.zhiFaCVNj3 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 112 -ge 120 ]] + echo -n . .+ sleep 1 + retry=113 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.94KpZUIk4T +++ mktemp ++ local LAST_ERR=/tmp/tmp.SG0pqiKeRi ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.94KpZUIk4T ++ cat /tmp/tmp.SG0pqiKeRi ++ rm /tmp/tmp.94KpZUIk4T /tmp/tmp.SG0pqiKeRi ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 113 -ge 120 ]] + echo -n . .+ sleep 1 + retry=114 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GztzdcgkGA +++ mktemp ++ local LAST_ERR=/tmp/tmp.gckyHnJFxk ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.GztzdcgkGA ++ cat /tmp/tmp.gckyHnJFxk ++ rm /tmp/tmp.GztzdcgkGA /tmp/tmp.gckyHnJFxk ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 114 -ge 120 ]] + echo -n . .+ sleep 1 + retry=115 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dUJIlGw6Y4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.9juOMg6M96 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.dUJIlGw6Y4 ++ cat /tmp/tmp.9juOMg6M96 ++ rm /tmp/tmp.dUJIlGw6Y4 /tmp/tmp.9juOMg6M96 ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 115 -ge 120 ]] + echo -n . .+ sleep 1 + retry=116 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0sKuKdEqM5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.DepXnLyTdX ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.0sKuKdEqM5 ++ cat /tmp/tmp.DepXnLyTdX ++ rm /tmp/tmp.0sKuKdEqM5 /tmp/tmp.DepXnLyTdX ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 116 -ge 120 ]] + echo -n . .+ sleep 1 + retry=117 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WOHMBYXIlx +++ mktemp ++ local LAST_ERR=/tmp/tmp.VCF5g87wlf ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.WOHMBYXIlx ++ cat /tmp/tmp.VCF5g87wlf ++ rm /tmp/tmp.WOHMBYXIlx /tmp/tmp.VCF5g87wlf ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 117 -ge 120 ]] + echo -n . .+ sleep 1 + retry=118 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hat2tf2lXW +++ mktemp ++ local LAST_ERR=/tmp/tmp.2Am9auOlRi ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.hat2tf2lXW ++ cat /tmp/tmp.2Am9auOlRi ++ rm /tmp/tmp.hat2tf2lXW /tmp/tmp.2Am9auOlRi ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 118 -ge 120 ]] + echo -n . .+ sleep 1 + retry=119 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ipq4V8eF96 +++ mktemp ++ local LAST_ERR=/tmp/tmp.7afVGjfLjy ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ipq4V8eF96 ++ cat /tmp/tmp.7afVGjfLjy ++ rm /tmp/tmp.ipq4V8eF96 /tmp/tmp.7afVGjfLjy ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 119 -ge 120 ]] + echo -n . .+ sleep 1 + retry=120 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BjAc5wcJUG +++ mktemp ++ local LAST_ERR=/tmp/tmp.LpkupiOAso ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.BjAc5wcJUG ++ cat /tmp/tmp.LpkupiOAso ++ rm /tmp/tmp.BjAc5wcJUG /tmp/tmp.LpkupiOAso ++ return 0 + [[ 1Gi == \2\G\i ]] + [[ 120 -ge 120 ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-0 was not resized, max retries exceeded' persistentvolumeclaim/mongod-data-some-name-rs0-0 was not resized, max retries exceeded + return 1 + return 1 + echo 'Enabling PVC resize' Enabling PVC resize + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "add", "path": "/spec/enableVolumeExpansion", "value":true }]' ++ mktemp + local LAST_OUT=/tmp/tmp.ibAB1AWudu ++ mktemp + local LAST_ERR=/tmp/tmp.UWobQCeebo + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "add", "path": "/spec/enableVolumeExpansion", "value":true }]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ibAB1AWudu perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.UWobQCeebo + rm /tmp/tmp.ibAB1AWudu /tmp/tmp.UWobQCeebo + return 0 + sleep 10 + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wfHJFcfgLD +++ mktemp ++ local LAST_ERR=/tmp/tmp.cKqhzEMnbV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.wfHJFcfgLD ++ cat /tmp/tmp.cKqhzEMnbV ++ rm /tmp/tmp.wfHJFcfgLD /tmp/tmp.cKqhzEMnbV ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 1 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.t9vunmOrJL +++ mktemp ++ local LAST_ERR=/tmp/tmp.M2lnDcBU5w ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.t9vunmOrJL ++ cat /tmp/tmp.M2lnDcBU5w ++ rm /tmp/tmp.t9vunmOrJL /tmp/tmp.M2lnDcBU5w ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 2 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.t0KOCpJC5V +++ mktemp ++ local LAST_ERR=/tmp/tmp.z01FF7TWTc ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.t0KOCpJC5V ++ cat /tmp/tmp.z01FF7TWTc ++ rm /tmp/tmp.t0KOCpJC5V /tmp/tmp.z01FF7TWTc ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 3 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9pw5NZtToT +++ mktemp ++ local LAST_ERR=/tmp/tmp.T4F02ouSCt ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9pw5NZtToT ++ cat /tmp/tmp.T4F02ouSCt ++ rm /tmp/tmp.9pw5NZtToT /tmp/tmp.T4F02ouSCt ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 4 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.l08gPsp45R +++ mktemp ++ local LAST_ERR=/tmp/tmp.1oHmJxN2mg ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.l08gPsp45R ++ cat /tmp/tmp.1oHmJxN2mg ++ rm /tmp/tmp.l08gPsp45R /tmp/tmp.1oHmJxN2mg ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 5 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IRRs2DFYOI +++ mktemp ++ local LAST_ERR=/tmp/tmp.BMFjeTqBtJ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.IRRs2DFYOI ++ cat /tmp/tmp.BMFjeTqBtJ ++ rm /tmp/tmp.IRRs2DFYOI /tmp/tmp.BMFjeTqBtJ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 6 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2vbqjj1tKq +++ mktemp ++ local LAST_ERR=/tmp/tmp.dJ6axnkIdj ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.2vbqjj1tKq ++ cat /tmp/tmp.dJ6axnkIdj ++ rm /tmp/tmp.2vbqjj1tKq /tmp/tmp.dJ6axnkIdj ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 7 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vQWa4sYres +++ mktemp ++ local LAST_ERR=/tmp/tmp.aCeH6yjctI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.vQWa4sYres ++ cat /tmp/tmp.aCeH6yjctI ++ rm /tmp/tmp.vQWa4sYres /tmp/tmp.aCeH6yjctI ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 8 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6aFN3rhxhx +++ mktemp ++ local LAST_ERR=/tmp/tmp.vPXZcftKMl ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.6aFN3rhxhx ++ cat /tmp/tmp.vPXZcftKMl ++ rm /tmp/tmp.6aFN3rhxhx /tmp/tmp.vPXZcftKMl ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo + wait_all_pvc_resize 2Gi + local expected_size=2Gi + local max_retry=120 + local sleep_time=5 ++ kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name +++ mktemp ++ local LAST_OUT=/tmp/tmp.ygfyY3Nj2H +++ mktemp ++ local LAST_ERR=/tmp/tmp.MrOHxHXbL4 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pvc -l app.kubernetes.io/component=mongod -o name ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ygfyY3Nj2H ++ cat /tmp/tmp.MrOHxHXbL4 ++ rm /tmp/tmp.ygfyY3Nj2H /tmp/tmp.MrOHxHXbL4 ++ return 0 + for pvc in '$(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name)' + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-0 2Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-0 + local expected_size=2Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.aOtdk9QvNA +++ mktemp ++ local LAST_ERR=/tmp/tmp.FmT2lCP29F ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.aOtdk9QvNA ++ cat /tmp/tmp.FmT2lCP29F ++ rm /tmp/tmp.aOtdk9QvNA /tmp/tmp.FmT2lCP29F ++ return 0 + [[ 2Gi == \2\G\i ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized + return 0 + for pvc in '$(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name)' + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-1 2Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-1 + local expected_size=2Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-1 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-1 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EjdyOQuyyo +++ mktemp ++ local LAST_ERR=/tmp/tmp.39sIuIn312 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.EjdyOQuyyo ++ cat /tmp/tmp.39sIuIn312 ++ rm /tmp/tmp.EjdyOQuyyo /tmp/tmp.39sIuIn312 ++ return 0 + [[ 2Gi == \2\G\i ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-1 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-1 was resized + return 0 + for pvc in '$(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name)' + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-2 2Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-2 + local expected_size=2Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-2 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-2 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SVEzUv6rKC +++ mktemp ++ local LAST_ERR=/tmp/tmp.DBFDxPH18u ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.SVEzUv6rKC ++ cat /tmp/tmp.DBFDxPH18u ++ rm /tmp/tmp.SVEzUv6rKC /tmp/tmp.DBFDxPH18u ++ return 0 + [[ 2Gi == \2\G\i ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-2 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-2 was resized + return 0 + return 0 + echo + [[ 0 == 1 ]] + [[ -n '' ]] + desc 'create resourcequota' + set +o xtrace ----------------------------------------------------------------------------------- create resourcequota ----------------------------------------------------------------------------------- + apply_resourcequota 7Gi + local quota=7Gi + local default_sc ++ get_default_storageclass ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QwDuCfk8HI +++ mktemp ++ local LAST_ERR=/tmp/tmp.B1kdADCl29 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.QwDuCfk8HI ++ cat /tmp/tmp.B1kdADCl29 ++ rm /tmp/tmp.QwDuCfk8HI /tmp/tmp.B1kdADCl29 ++ return 0 + default_sc=standard-rwo + echo 'Applying resourcequota for default storageclass standard-rwo with quota 7Gi' Applying resourcequota for default storageclass standard-rwo with quota 7Gi + sed s/STORAGECLASS/standard-rwo/ + kubectl_bin apply -f - + sed s/QUOTA/7Gi/ ++ mktemp + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/e2e-tests/pvc-resize/conf/resourcequota.yml + local LAST_OUT=/tmp/tmp.vKBYh1slIu ++ mktemp + local LAST_ERR=/tmp/tmp.9g8o5peYA3 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.vKBYh1slIu resourcequota/default-storage-quota created + cat /tmp/tmp.9g8o5peYA3 + rm /tmp/tmp.vKBYh1slIu /tmp/tmp.9g8o5peYA3 + return 0 + patch_pvc_request some-name 3G + local cluster=some-name + local size=3G + echo 'Patching PVC request to 3G in some-name' Patching PVC request to 3G in some-name + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"3G"}]' ++ mktemp + local LAST_OUT=/tmp/tmp.6I3zx9psbu ++ mktemp + local LAST_ERR=/tmp/tmp.ta97PsXoH8 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"3G"}]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.6I3zx9psbu perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.ta97PsXoH8 + rm /tmp/tmp.6I3zx9psbu /tmp/tmp.ta97PsXoH8 + return 0 + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fzyfZFfvpK +++ mktemp ++ local LAST_ERR=/tmp/tmp.XGZDLLlX4t ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.fzyfZFfvpK ++ cat /tmp/tmp.XGZDLLlX4t ++ rm /tmp/tmp.fzyfZFfvpK /tmp/tmp.XGZDLLlX4t ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo + echo + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-0 3Gi + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-0 + local expected_size=3Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AeJ8zyJBRN +++ mktemp ++ local LAST_ERR=/tmp/tmp.XhIrpL7wXV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.AeJ8zyJBRN ++ cat /tmp/tmp.XhIrpL7wXV ++ rm /tmp/tmp.AeJ8zyJBRN /tmp/tmp.XhIrpL7wXV ++ return 0 + [[ 2Gi == \3\G\i ]] + [[ 0 -ge 120 ]] + echo -n . .+ sleep 5 + retry=1 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.471f32dz2M +++ mktemp ++ local LAST_ERR=/tmp/tmp.dQlZmI15lr ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.471f32dz2M ++ cat /tmp/tmp.dQlZmI15lr ++ rm /tmp/tmp.471f32dz2M /tmp/tmp.dQlZmI15lr ++ return 0 + [[ 2Gi == \3\G\i ]] + [[ 1 -ge 120 ]] + echo -n . .+ sleep 5 + retry=2 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VBNY4jSlJY +++ mktemp ++ local LAST_ERR=/tmp/tmp.zs27mp86kP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.VBNY4jSlJY ++ cat /tmp/tmp.zs27mp86kP ++ rm /tmp/tmp.VBNY4jSlJY /tmp/tmp.zs27mp86kP ++ return 0 + [[ 2Gi == \3\G\i ]] + [[ 2 -ge 120 ]] + echo -n . .+ sleep 5 + retry=3 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sTlTZsr3fb +++ mktemp ++ local LAST_ERR=/tmp/tmp.a10a2f8ygF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.sTlTZsr3fb ++ cat /tmp/tmp.a10a2f8ygF ++ rm /tmp/tmp.sTlTZsr3fb /tmp/tmp.a10a2f8ygF ++ return 0 + [[ 2Gi == \3\G\i ]] + [[ 3 -ge 120 ]] + echo -n . .+ sleep 5 + retry=4 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qzePE9t5Ve +++ mktemp ++ local LAST_ERR=/tmp/tmp.hkjLSifVRP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.qzePE9t5Ve ++ cat /tmp/tmp.hkjLSifVRP ++ rm /tmp/tmp.qzePE9t5Ve /tmp/tmp.hkjLSifVRP ++ return 0 + [[ 2Gi == \3\G\i ]] + [[ 4 -ge 120 ]] + echo -n . .+ sleep 5 + retry=5 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.J1zikCJZRL +++ mktemp ++ local LAST_ERR=/tmp/tmp.ca2lzjTtgN ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.J1zikCJZRL ++ cat /tmp/tmp.ca2lzjTtgN ++ rm /tmp/tmp.J1zikCJZRL /tmp/tmp.ca2lzjTtgN ++ return 0 + [[ 2Gi == \3\G\i ]] + [[ 5 -ge 120 ]] + echo -n . .+ sleep 5 + retry=6 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.H6EGoEt6bs +++ mktemp ++ local LAST_ERR=/tmp/tmp.cCMxBvNmFj ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.H6EGoEt6bs ++ cat /tmp/tmp.cCMxBvNmFj ++ rm /tmp/tmp.H6EGoEt6bs /tmp/tmp.cCMxBvNmFj ++ return 0 + [[ 2Gi == \3\G\i ]] + [[ 6 -ge 120 ]] + echo -n . .+ sleep 5 + retry=7 ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qo4SxzO98l +++ mktemp ++ local LAST_ERR=/tmp/tmp.By7BlM64qP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.qo4SxzO98l ++ cat /tmp/tmp.By7BlM64qP ++ rm /tmp/tmp.qo4SxzO98l /tmp/tmp.By7BlM64qP ++ return 0 + [[ 3Gi == \3\G\i ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized + return 0 + apply_resourcequota 9Gi + local quota=9Gi + local default_sc ++ get_default_storageclass ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yjK2hDX5Im +++ mktemp ++ local LAST_ERR=/tmp/tmp.E1zxmfKcV1 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.yjK2hDX5Im ++ cat /tmp/tmp.E1zxmfKcV1 ++ rm /tmp/tmp.yjK2hDX5Im /tmp/tmp.E1zxmfKcV1 ++ return 0 + default_sc=standard-rwo + echo 'Applying resourcequota for default storageclass standard-rwo with quota 9Gi' Applying resourcequota for default storageclass standard-rwo with quota 9Gi + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/e2e-tests/pvc-resize/conf/resourcequota.yml + sed s/STORAGECLASS/standard-rwo/ + sed s/QUOTA/9Gi/ + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.lWWS0MRun0 ++ mktemp + local LAST_ERR=/tmp/tmp.7PVarMjLPG + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.lWWS0MRun0 resourcequota/default-storage-quota configured + cat /tmp/tmp.7PVarMjLPG + rm /tmp/tmp.lWWS0MRun0 /tmp/tmp.7PVarMjLPG + return 0 + patch_pvc_request some-name 3G + local cluster=some-name + local size=3G + echo 'Patching PVC request to 3G in some-name' Patching PVC request to 3G in some-name + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"3G"}]' ++ mktemp + local LAST_OUT=/tmp/tmp.ffbDoTpbLt ++ mktemp + local LAST_ERR=/tmp/tmp.t08CyjODO8 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"3G"}]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ffbDoTpbLt perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.t08CyjODO8 + rm /tmp/tmp.ffbDoTpbLt /tmp/tmp.t08CyjODO8 + return 0 + wait_cluster_consistency some-name 42 + local cluster_name=some-name + local wait_time=42 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.onCjV2f2xz +++ mktemp ++ local LAST_ERR=/tmp/tmp.eeXb2MOspp ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.onCjV2f2xz ++ cat /tmp/tmp.eeXb2MOspp ++ rm /tmp/tmp.onCjV2f2xz /tmp/tmp.eeXb2MOspp ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 1 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SHQv131gwn +++ mktemp ++ local LAST_ERR=/tmp/tmp.mxbKo3fc7k ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.SHQv131gwn ++ cat /tmp/tmp.mxbKo3fc7k ++ rm /tmp/tmp.SHQv131gwn /tmp/tmp.mxbKo3fc7k ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 2 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FlcG44SlqF +++ mktemp ++ local LAST_ERR=/tmp/tmp.BD3QzwtYIY ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.FlcG44SlqF ++ cat /tmp/tmp.BD3QzwtYIY ++ rm /tmp/tmp.FlcG44SlqF /tmp/tmp.BD3QzwtYIY ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 3 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.H3MVbEQO6m +++ mktemp ++ local LAST_ERR=/tmp/tmp.jptWr2MSSi ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.H3MVbEQO6m ++ cat /tmp/tmp.jptWr2MSSi ++ rm /tmp/tmp.H3MVbEQO6m /tmp/tmp.jptWr2MSSi ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 4 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hhaWniOVAy +++ mktemp ++ local LAST_ERR=/tmp/tmp.jSaJBM4nT4 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.hhaWniOVAy ++ cat /tmp/tmp.jSaJBM4nT4 ++ rm /tmp/tmp.hhaWniOVAy /tmp/tmp.jSaJBM4nT4 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 5 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HkuWmobW8o +++ mktemp ++ local LAST_ERR=/tmp/tmp.wxaMNICIgT ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.HkuWmobW8o ++ cat /tmp/tmp.wxaMNICIgT ++ rm /tmp/tmp.HkuWmobW8o /tmp/tmp.wxaMNICIgT ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 6 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WWvu7LB97G +++ mktemp ++ local LAST_ERR=/tmp/tmp.oZNvuZOd4K ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.WWvu7LB97G ++ cat /tmp/tmp.oZNvuZOd4K ++ rm /tmp/tmp.WWvu7LB97G /tmp/tmp.oZNvuZOd4K ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 7 -ge 42 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wWp54eLR62 +++ mktemp ++ local LAST_ERR=/tmp/tmp.A76ObV98xb ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.wWp54eLR62 ++ cat /tmp/tmp.A76ObV98xb ++ rm /tmp/tmp.wWp54eLR62 /tmp/tmp.A76ObV98xb ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo + echo + wait_all_pvc_resize 3Gi + local expected_size=3Gi + local max_retry=120 + local sleep_time=5 ++ kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name +++ mktemp ++ local LAST_OUT=/tmp/tmp.30S6BMsPO6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.uFLY6uYuaa ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pvc -l app.kubernetes.io/component=mongod -o name ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.30S6BMsPO6 ++ cat /tmp/tmp.uFLY6uYuaa ++ rm /tmp/tmp.30S6BMsPO6 /tmp/tmp.uFLY6uYuaa ++ return 0 + for pvc in '$(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name)' + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-0 3Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-0 + local expected_size=3Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-0 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2tmGOYSyPt +++ mktemp ++ local LAST_ERR=/tmp/tmp.rjcB0AY86p ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.2tmGOYSyPt ++ cat /tmp/tmp.rjcB0AY86p ++ rm /tmp/tmp.2tmGOYSyPt /tmp/tmp.rjcB0AY86p ++ return 0 + [[ 3Gi == \3\G\i ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-0 was resized + return 0 + for pvc in '$(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name)' + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-1 3Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-1 + local expected_size=3Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-1 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-1 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Vo0cKK2mDn +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZeD6zAQuiZ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Vo0cKK2mDn ++ cat /tmp/tmp.ZeD6zAQuiZ ++ rm /tmp/tmp.Vo0cKK2mDn /tmp/tmp.ZeD6zAQuiZ ++ return 0 + [[ 3Gi == \3\G\i ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-1 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-1 was resized + return 0 + for pvc in '$(kubectl_bin get pvc -l app.kubernetes.io/component=mongod -o name)' + wait_pvc_resize persistentvolumeclaim/mongod-data-some-name-rs0-2 3Gi 120 5 + local pvc=persistentvolumeclaim/mongod-data-some-name-rs0-2 + local expected_size=3Gi + local max_retry=120 + local sleep_time=5 + local retry=0 + echo 'Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-2 to be resized' Waiting for persistentvolumeclaim/mongod-data-some-name-rs0-2 to be resized ++ kubectl_bin get persistentvolumeclaim/mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.iEvPa5EBz8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.w6zqEgc23r ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get persistentvolumeclaim/mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.iEvPa5EBz8 ++ cat /tmp/tmp.w6zqEgc23r ++ rm /tmp/tmp.iEvPa5EBz8 /tmp/tmp.w6zqEgc23r ++ return 0 + [[ 3Gi == \3\G\i ]] + echo + echo 'persistentvolumeclaim/mongod-data-some-name-rs0-2 was resized' persistentvolumeclaim/mongod-data-some-name-rs0-2 was resized + return 0 + return 0 + desc 'test downscale' + set +o xtrace ----------------------------------------------------------------------------------- test downscale ----------------------------------------------------------------------------------- + patch_pvc_request some-name 1G + local cluster=some-name + local size=1G + echo 'Patching PVC request to 1G in some-name' Patching PVC request to 1G in some-name + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"1G"}]' ++ mktemp + local LAST_OUT=/tmp/tmp.BPh6T6juBC ++ mktemp + local LAST_ERR=/tmp/tmp.uHIM9RGCtQ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"1G"}]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.BPh6T6juBC perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.uHIM9RGCtQ + rm /tmp/tmp.BPh6T6juBC /tmp/tmp.uHIM9RGCtQ + return 0 + wait_cluster_status some-name error + local cluster=some-name + local expected=error + echo -n 'Waiting for psmdb/some-name status to be error' Waiting for psmdb/some-name status to be error++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uRV8M1kTYy +++ mktemp ++ local LAST_ERR=/tmp/tmp.14PXk2h2gm ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.uRV8M1kTYy ++ cat /tmp/tmp.14PXk2h2gm ++ rm /tmp/tmp.uRV8M1kTYy /tmp/tmp.14PXk2h2gm ++ return 0 + [[ error == \e\r\r\o\r ]] + echo + echo 'psmdb/some-name status is error' psmdb/some-name status is error + patch_pvc_request some-name 3G + local cluster=some-name + local size=3G + echo 'Patching PVC request to 3G in some-name' Patching PVC request to 3G in some-name + kubectl_bin patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"3G"}]' ++ mktemp + local LAST_OUT=/tmp/tmp.gL77dph6Zq ++ mktemp + local LAST_ERR=/tmp/tmp.n1VpqVNs67 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch psmdb some-name --type=json '-p=[{"op": "replace", "path": "/spec/replsets/0/volumeSpec/persistentVolumeClaim/resources/requests/storage", "value":"3G"}]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.gL77dph6Zq perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.n1VpqVNs67 + rm /tmp/tmp.gL77dph6Zq /tmp/tmp.n1VpqVNs67 + return 0 + wait_cluster_status some-name ready + local cluster=some-name + local expected=ready + echo -n 'Waiting for psmdb/some-name status to be ready' Waiting for psmdb/some-name status to be ready++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.R1S0CtOOtv +++ mktemp ++ local LAST_ERR=/tmp/tmp.Zc839bFzV0 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.R1S0CtOOtv ++ cat /tmp/tmp.Zc839bFzV0 ++ rm /tmp/tmp.R1S0CtOOtv /tmp/tmp.Zc839bFzV0 ++ return 0 + [[ error == \r\e\a\d\y ]] + [[ 7 -ge 60 ]] + echo -n . .+ sleep 5 + retry=8 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2xrpsTodFX +++ mktemp ++ local LAST_ERR=/tmp/tmp.MYIJH2Vmja ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.2xrpsTodFX ++ cat /tmp/tmp.MYIJH2Vmja ++ rm /tmp/tmp.2xrpsTodFX /tmp/tmp.MYIJH2Vmja ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo + echo 'psmdb/some-name status is ready' psmdb/some-name status is ready + destroy pvc-resize-28274 + local namespace=pvc-resize-28274 + local ignore_logs=true + desc 'destroy cluster/operator and all other resources' + set +o xtrace ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + '[' true == false ']' + delete_backups + desc 'Delete psmdb-backup' + set +o xtrace ----------------------------------------------------------------------------------- Delete psmdb-backup ----------------------------------------------------------------------------------- ++ kubectl_bin get psmdb-backup --no-headers ++ wc -l +++ mktemp ++ local LAST_OUT=/tmp/tmp.BILeMA2YSt +++ mktemp ++ local LAST_ERR=/tmp/tmp.5ihidfj2u9 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb-backup --no-headers ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.BILeMA2YSt ++ cat /tmp/tmp.5ihidfj2u9 No resources found in pvc-resize-28274 namespace. ++ rm /tmp/tmp.BILeMA2YSt /tmp/tmp.5ihidfj2u9 ++ return 0 + '[' 0 '!=' 0 ']' + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.4RGG87mQSL ++ mktemp + local LAST_ERR=/tmp/tmp.ewXv0PSILj + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.4RGG87mQSL customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.ewXv0PSILj + rm /tmp/tmp.4RGG87mQSL /tmp/tmp.ewXv0PSILj + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/deploy/crd.yaml ++ grep -v '\-\-\-' + for crd_name in '$(yq eval '\''.metadata.name'\'' "${src_dir}/deploy/crd.yaml" | grep -v '\''\-\-\-'\'')' + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.sPUQ7xreCi ++ mktemp + local LAST_ERR=/tmp/tmp.v2TpZ8AC2K + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.sPUQ7xreCi + cat /tmp/tmp.v2TpZ8AC2K + rm /tmp/tmp.sPUQ7xreCi /tmp/tmp.v2TpZ8AC2K + return 0 + for crd_name in '$(yq eval '\''.metadata.name'\'' "${src_dir}/deploy/crd.yaml" | grep -v '\''\-\-\-'\'')' + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.yJhRcmJ1wF ++ mktemp + local LAST_ERR=/tmp/tmp.FX02SLfF6c + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.yJhRcmJ1wF + cat /tmp/tmp.FX02SLfF6c + rm /tmp/tmp.yJhRcmJ1wF /tmp/tmp.FX02SLfF6c + return 0 + for crd_name in '$(yq eval '\''.metadata.name'\'' "${src_dir}/deploy/crd.yaml" | grep -v '\''\-\-\-'\'')' + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + grep -v NAMESPACE + kubectl patch perconaservermongodbs.psmdb.percona.com -n pvc-resize-28274 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaservermongodb.psmdb.percona.com/some-name patched + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.sO8vRbCN2V ++ mktemp + local LAST_ERR=/tmp/tmp.tMcXw4WNUg + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.sO8vRbCN2V + cat /tmp/tmp.tMcXw4WNUg + rm /tmp/tmp.sO8vRbCN2V /tmp/tmp.tMcXw4WNUg + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.d7zIsXGEwn ++ mktemp + local LAST_ERR=/tmp/tmp.S9eJGG454I + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1904/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.d7zIsXGEwn clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.S9eJGG454I + rm /tmp/tmp.d7zIsXGEwn /tmp/tmp.S9eJGG454I + return 0 + destroy_cert_manager + kubectl_bin delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.jwe0F3vaix ++ mktemp + local LAST_ERR=/tmp/tmp.cQqywCkF6H + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.jwe0F3vaix + cat /tmp/tmp.cQqywCkF6H Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.jwe0F3vaix + cat /tmp/tmp.cQqywCkF6H Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 4 + for i in '$(seq 0 2)' + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.jwe0F3vaix + cat /tmp/tmp.cQqywCkF6H Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 8 + cat /tmp/tmp.jwe0F3vaix + cat /tmp/tmp.cQqywCkF6H Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.17.2/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + rm /tmp/tmp.jwe0F3vaix /tmp/tmp.cQqywCkF6H + return 1 + true + '[' -n '' ']' + '[' -n psmdb-operator ']' + kubectl_bin delete --grace-period=0 --force=true namespace pvc-resize-28274 + rm -rf /tmp/tmp.i92JTyYaWV ++ mktemp + kubectl_bin delete --grace-period=0 --force=true namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.2CqKGVjMww + desc 'test passed' ++ mktemp + set +o xtrace ----------------------------------------------------------------------------------- test passed ----------------------------------------------------------------------------------- + local LAST_OUT=/tmp/tmp.g9VdW2BQ7q ++ mktemp + local LAST_ERR=/tmp/tmp.jNgqq10E4f + local exit_status=0 + local timeout=4 ++ seq 0 2 + local LAST_ERR=/tmp/tmp.RvlDImgPSE + local exit_status=0 + local timeout=4 + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace pvc-resize-28274 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace psmdb-operator