Log: /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/e2e-tests/logs/pvc-auto-resize.log Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 + '[' 0 == 1 ']' + ensure_default_sc_allows_expansion + local default_sc ++ get_default_storageclass ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.limH0ceMMX +++ mktemp ++ local LAST_ERR=/tmp/tmp.iGYYPQgXm5 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.limH0ceMMX ++ cat /tmp/tmp.iGYYPQgXm5 ++ rm /tmp/tmp.limH0ceMMX /tmp/tmp.iGYYPQgXm5 ++ return 0 + default_sc=standard-rwo + echo 'Checking if default storageclass standard-rwo allows volume expansion' Checking if default storageclass standard-rwo allows volume expansion + local allowVolumeExpansion ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.name=="standard-rwo")].allowVolumeExpansion}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GiqZ2g99Ki +++ mktemp ++ local LAST_ERR=/tmp/tmp.3HAVBm5VEH ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.name=="standard-rwo")].allowVolumeExpansion}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.GiqZ2g99Ki ++ cat /tmp/tmp.3HAVBm5VEH ++ rm /tmp/tmp.GiqZ2g99Ki /tmp/tmp.3HAVBm5VEH ++ return 0 + allowVolumeExpansion=true + [[ true != \t\r\u\e ]] + create_infra pvc-auto-resize-27302 + local ns=pvc-auto-resize-27302 + [[ 1 == 1 ]] + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.eRAWA70Fta ++ mktemp + local LAST_ERR=/tmp/tmp.v175DXb1xE + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.eRAWA70Fta + cat /tmp/tmp.v175DXb1xE + rm /tmp/tmp.eRAWA70Fta /tmp/tmp.v175DXb1xE + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/deploy/crd.yaml ++ grep -v '\-\-\-' grep: warning: stray \ before - grep: warning: stray \ before - + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.ZY9AbXFweH ++ mktemp + local LAST_ERR=/tmp/tmp.damKQBhyqJ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ZY9AbXFweH + cat /tmp/tmp.damKQBhyqJ + rm /tmp/tmp.ZY9AbXFweH /tmp/tmp.damKQBhyqJ + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.mzS45BSAxI ++ mktemp + local LAST_ERR=/tmp/tmp.m31v0qry0V + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.mzS45BSAxI + cat /tmp/tmp.m31v0qry0V + rm /tmp/tmp.mzS45BSAxI /tmp/tmp.m31v0qry0V + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbs" + kubectl patch perconaservermongodbs.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbs" + : + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.p6rXg283U9 ++ mktemp + local LAST_ERR=/tmp/tmp.ZebnbW22xD + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.p6rXg283U9 + cat /tmp/tmp.ZebnbW22xD + rm /tmp/tmp.p6rXg283U9 /tmp/tmp.ZebnbW22xD + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.7NpIWMJeh8 ++ mktemp + local LAST_ERR=/tmp/tmp.C3KX9iptww + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.7NpIWMJeh8 + cat /tmp/tmp.C3KX9iptww + rm /tmp/tmp.7NpIWMJeh8 /tmp/tmp.C3KX9iptww + return 0 + check_crd_for_deletion PR-2247-091270b4 + local git_tag=PR-2247-091270b4 ++ curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/PR-2247-091270b4/deploy/crd.yaml ++ yq eval .metadata.name ++ /usr/sbin/sed s/---//g ++ /usr/sbin/sed ':a;N;$!ba;s/\n/ /g' + for crd_name in $(curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/${git_tag}/deploy/crd.yaml | yq eval '.metadata.name' | $sed 's/---//g' | $sed ':a;N;$!ba;s/\n/ /g') ++ kubectl_bin get crd/null -o 'jsonpath={.status.conditions[-1].type}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eIllaJvyGz +++ mktemp ++ local LAST_ERR=/tmp/tmp.gg3q75GcWt ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.eIllaJvyGz ++ cat /tmp/tmp.gg3q75GcWt Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 0 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.eIllaJvyGz ++ cat /tmp/tmp.gg3q75GcWt Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 4 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.eIllaJvyGz ++ cat /tmp/tmp.gg3q75GcWt Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 8 ++ cat /tmp/tmp.eIllaJvyGz ++ cat /tmp/tmp.gg3q75GcWt Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ rm /tmp/tmp.eIllaJvyGz /tmp/tmp.gg3q75GcWt ++ return 1 + [[ '' == \T\e\r\m\i\n\a\t\i\n\g ]] + '[' -n psmdb-operator ']' + create_namespace psmdb-operator + local namespace=psmdb-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ grep chaos-mesh.org ++ kubectl get crd ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + '[' -n '' ']' + awk '{print$1}' + desc 'cleaned up old namespaces psmdb-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace psmdb-operator --ignore-not-found + xargs kubectl delete ns ++ mktemp + kubectl_bin get ns + grep -E -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' + local LAST_OUT=/tmp/tmp.tQEUItlcDK ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.LEXMpV0Bno + local exit_status=0 + local timeout=4 + local LAST_OUT=/tmp/tmp.W37A8okT9P ++ seq 0 2 ++ mktemp + local LAST_ERR=/tmp/tmp.itFPDjYBjA + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete namespace psmdb-operator --ignore-not-found + for i in $(seq 0 2) + set +e + kubectl get ns + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.W37A8okT9P + cat /tmp/tmp.itFPDjYBjA + rm /tmp/tmp.W37A8okT9P /tmp/tmp.itFPDjYBjA + return 0 error: resource(s) were provided, but no name was specified + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.tQEUItlcDK + cat /tmp/tmp.LEXMpV0Bno + rm /tmp/tmp.tQEUItlcDK /tmp/tmp.LEXMpV0Bno + return 0 + kubectl_bin wait --for=delete namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.lTJbAKsRwp ++ mktemp + local LAST_ERR=/tmp/tmp.uzQULoe9vw + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.lTJbAKsRwp + cat /tmp/tmp.uzQULoe9vw + rm /tmp/tmp.lTJbAKsRwp /tmp/tmp.uzQULoe9vw + return 0 + desc 'create namespace psmdb-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.xUjMEHbYWA ++ mktemp + local LAST_ERR=/tmp/tmp.YUJS0ln9lP + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.xUjMEHbYWA namespace/psmdb-operator created + cat /tmp/tmp.YUJS0ln9lP + rm /tmp/tmp.xUjMEHbYWA /tmp/tmp.YUJS0ln9lP + return 0 + set_kube_ctx psmdb-operator + local namespace=psmdb-operator ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.KBObXOPNF8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.NEYPPQYpbv ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.KBObXOPNF8 ++ cat /tmp/tmp.NEYPPQYpbv ++ rm /tmp/tmp.KBObXOPNF8 /tmp/tmp.NEYPPQYpbv ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2247-091270b4-5-cluster5 --namespace=psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.a4waEpVIwK ++ mktemp + local LAST_ERR=/tmp/tmp.odVyokLwQ0 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2247-091270b4-5-cluster5 --namespace=psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.a4waEpVIwK Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2247-091270b4-5-cluster5" modified. + cat /tmp/tmp.odVyokLwQ0 + rm /tmp/tmp.a4waEpVIwK /tmp/tmp.odVyokLwQ0 + return 0 + deploy_operator + desc 'start PSMDB operator: perconalab/percona-server-mongodb-operator:PR-2247-091270b4' + set +o xtrace ----------------------------------------------------------------------------------- start PSMDB operator: perconalab/percona-server-mongodb-operator:PR-2247-091270b4 ----------------------------------------------------------------------------------- + local cr_file + '[' -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/e2e-tests/pvc-auto-resize/conf/crd.yaml ']' + cr_file=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/deploy/crd.yaml + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.akBcGexuNL ++ mktemp + local LAST_ERR=/tmp/tmp.t7zDSMuq7I + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.akBcGexuNL customresourcedefinition.apiextensions.k8s.io/perconaservermongodbbackups.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbrestores.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com serverside-applied + cat /tmp/tmp.t7zDSMuq7I + rm /tmp/tmp.akBcGexuNL /tmp/tmp.t7zDSMuq7I + return 0 + '[' -n psmdb-operator ']' + apply_rbac cw-rbac + local operator_namespace=psmdb-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: psmdb-operator^' + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.1TuOOGHgEQ ++ mktemp + local LAST_ERR=/tmp/tmp.jxp8gadJdu + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.1TuOOGHgEQ clusterrole.rbac.authorization.k8s.io/percona-server-mongodb-operator created serviceaccount/percona-server-mongodb-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-server-mongodb-operator created + cat /tmp/tmp.jxp8gadJdu + rm /tmp/tmp.1TuOOGHgEQ /tmp/tmp.jxp8gadJdu + return 0 + kubectl_bin apply -n psmdb-operator -f - + yq eval ' (.spec.template.spec.containers[].image = "perconalab/percona-server-mongodb-operator:PR-2247-091270b4") | ((.. | select(.[] == "DISABLE_TELEMETRY")) |= .value="true") | ((.. | select(.[] == "LOG_LEVEL")) |= .value="DEBUG")' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/deploy/cw-operator.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.aAvDUHjq47 ++ mktemp + local LAST_ERR=/tmp/tmp.CdP0DwZaFX + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.aAvDUHjq47 deployment.apps/percona-server-mongodb-operator created + cat /tmp/tmp.CdP0DwZaFX + rm /tmp/tmp.aAvDUHjq47 /tmp/tmp.CdP0DwZaFX + return 0 + sleep 20 ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.MPEGn18abu +++ mktemp ++ local LAST_ERR=/tmp/tmp.HQksItASEA ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.MPEGn18abu ++ cat /tmp/tmp.HQksItASEA ++ rm /tmp/tmp.MPEGn18abu /tmp/tmp.HQksItASEA ++ return 0 + wait_operator_pod percona-server-mongodb-operator-64bc9b5fc9-nzkx5 + local pod=percona-server-mongodb-operator-64bc9b5fc9-nzkx5 + set +o xtrace waiting for pod/percona-server-mongodb-operator-64bc9b5fc9-nzkx5 to be ready.OK + echo 'Print operator info from log' Print operator info from log + grep 'Manager starting up' ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.tRJt0UzaxK +++ mktemp ++ local LAST_ERR=/tmp/tmp.3RG5MZl8sw ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.tRJt0UzaxK ++ cat /tmp/tmp.3RG5MZl8sw ++ rm /tmp/tmp.tRJt0UzaxK /tmp/tmp.3RG5MZl8sw ++ return 0 + kubectl_bin logs -n psmdb-operator percona-server-mongodb-operator-64bc9b5fc9-nzkx5 ++ mktemp + local LAST_OUT=/tmp/tmp.6q4aL6F8Z3 ++ mktemp + local LAST_ERR=/tmp/tmp.qM5VKwzT6i + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl logs -n psmdb-operator percona-server-mongodb-operator-64bc9b5fc9-nzkx5 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.6q4aL6F8Z3 + cat /tmp/tmp.qM5VKwzT6i + rm /tmp/tmp.6q4aL6F8Z3 /tmp/tmp.qM5VKwzT6i + return 0 2026-02-19T14:48:44.321Z INFO setup Manager starting up {"gitCommit": "091270b48773244359d74f74a10340fc1c0cff51", "gitBranch": "PR-2247-091270b4", "buildTime": "", "goVersion": "go1.25.7", "os": "linux", "arch": "amd64"} + create_namespace pvc-auto-resize-27302 + local namespace=pvc-auto-resize-27302 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + grep -E -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' + awk '{print$1}' ++ mktemp + '[' -n '' ']' + desc 'cleaned up old namespaces pvc-auto-resize-27302' + set +o xtrace + xargs kubectl delete ns ----------------------------------------------------------------------------------- cleaned up old namespaces pvc-auto-resize-27302 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pvc-auto-resize-27302 --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.y4XWu4fgHq ++ mktemp + local LAST_OUT=/tmp/tmp.Jx8WKaHUXt + local LAST_ERR=/tmp/tmp.0aeRzKXgjV ++ mktemp + local exit_status=0 + local timeout=4 ++ seq 0 2 + local LAST_ERR=/tmp/tmp.mrG26vZ50C + local exit_status=0 + local timeout=4 + for i in $(seq 0 2) + set +e + kubectl get ns ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete namespace pvc-auto-resize-27302 --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.y4XWu4fgHq + cat /tmp/tmp.0aeRzKXgjV + rm /tmp/tmp.y4XWu4fgHq /tmp/tmp.0aeRzKXgjV + return 0 error: resource(s) were provided, but no name was specified + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Jx8WKaHUXt + cat /tmp/tmp.mrG26vZ50C + rm /tmp/tmp.Jx8WKaHUXt /tmp/tmp.mrG26vZ50C + return 0 + kubectl_bin wait --for=delete namespace pvc-auto-resize-27302 ++ mktemp + local LAST_OUT=/tmp/tmp.MdbMt15ksS ++ mktemp + local LAST_ERR=/tmp/tmp.YTBg1IGPRm + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete namespace pvc-auto-resize-27302 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.MdbMt15ksS + cat /tmp/tmp.YTBg1IGPRm + rm /tmp/tmp.MdbMt15ksS /tmp/tmp.YTBg1IGPRm + return 0 + desc 'create namespace pvc-auto-resize-27302' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pvc-auto-resize-27302 ----------------------------------------------------------------------------------- + kubectl_bin create namespace pvc-auto-resize-27302 ++ mktemp + local LAST_OUT=/tmp/tmp.mkYafG5rwO ++ mktemp + local LAST_ERR=/tmp/tmp.yyXiLAlDu5 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace pvc-auto-resize-27302 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.mkYafG5rwO namespace/pvc-auto-resize-27302 created + cat /tmp/tmp.yyXiLAlDu5 + rm /tmp/tmp.mkYafG5rwO /tmp/tmp.yyXiLAlDu5 + return 0 + set_kube_ctx pvc-auto-resize-27302 + local namespace=pvc-auto-resize-27302 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.Id4nvtp78K +++ mktemp ++ local LAST_ERR=/tmp/tmp.NIOa2qLxFX ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Id4nvtp78K ++ cat /tmp/tmp.NIOa2qLxFX ++ rm /tmp/tmp.Id4nvtp78K /tmp/tmp.NIOa2qLxFX ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2247-091270b4-5-cluster5 --namespace=pvc-auto-resize-27302 ++ mktemp + local LAST_OUT=/tmp/tmp.6FxCPIX2RJ ++ mktemp + local LAST_ERR=/tmp/tmp.HELcjB6KBK + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2247-091270b4-5-cluster5 --namespace=pvc-auto-resize-27302 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.6FxCPIX2RJ Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2247-091270b4-5-cluster5" modified. + cat /tmp/tmp.HELcjB6KBK + rm /tmp/tmp.6FxCPIX2RJ /tmp/tmp.HELcjB6KBK + return 0 + desc 'create secrets and psmdb client' + set +o xtrace ----------------------------------------------------------------------------------- create secrets and psmdb client ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/e2e-tests/conf/secrets.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.fNblzFZov8 ++ mktemp + local LAST_ERR=/tmp/tmp.7sJDzUT91A + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/e2e-tests/conf/secrets.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/e2e-tests/conf/client.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.fNblzFZov8 secret/some-users created deployment.apps/psmdb-client created + cat /tmp/tmp.7sJDzUT91A + rm /tmp/tmp.fNblzFZov8 /tmp/tmp.7sJDzUT91A + return 0 + desc 'create PSMDB cluster with storageAutoscaling enabled' + set +o xtrace ----------------------------------------------------------------------------------- create PSMDB cluster with storageAutoscaling enabled ----------------------------------------------------------------------------------- + cluster=some-name + spinup_psmdb some-name-rs0 /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/e2e-tests/pvc-auto-resize/conf/some-name.yml + local cluster=some-name-rs0 + local config=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/e2e-tests/pvc-auto-resize/conf/some-name.yml + local size=3 + desc 'create first PSMDB cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PSMDB cluster ----------------------------------------------------------------------------------- + apply_cluster /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/e2e-tests/pvc-auto-resize/conf/some-name.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/e2e-tests/pvc-auto-resize/conf/some-name.yml + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/e2e-tests/pvc-auto-resize/conf/some-name.yml ++ mktemp + yq eval '(.spec | select(.image == null)).image = "perconalab/percona-server-mongodb-operator:main-mongod8.0"' + yq eval '(.spec | select(has("pmm"))).pmm.image = "percona/pmm-client:2.44.1-1"' + yq eval '(.spec | select(has("initImage"))).initImage = "perconalab/percona-server-mongodb-operator:PR-2247-091270b4"' + yq eval '(.spec | select(has("backup"))).backup.image = "perconalab/percona-server-mongodb-operator:main-backup"' + /usr/sbin/sed -e s/NAME_SPACE/pvc-auto-resize-27302/g + yq eval '.spec.upgradeOptions.apply="Never"' + local LAST_OUT=/tmp/tmp.3tBFExfptY ++ mktemp + local LAST_ERR=/tmp/tmp.e1BOl8ASx6 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.3tBFExfptY perconaservermongodb.psmdb.percona.com/some-name created + cat /tmp/tmp.e1BOl8ASx6 + rm /tmp/tmp.3tBFExfptY /tmp/tmp.e1BOl8ASx6 + return 0 + desc 'check if Pod is started' + set +o xtrace ----------------------------------------------------------------------------------- check if Pod is started ----------------------------------------------------------------------------------- + wait_for_running some-name-rs0 3 + local name=some-name-rs0 + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=rs0 + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-rs0-0 + local pod=some-name-rs0-0 + set +o xtrace waiting for pod/some-name-rs0-0 to be ready...............OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-rs0-1 + local pod=some-name-rs0-1 + set +o xtrace waiting for pod/some-name-rs0-1 to be ready.............OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.do8WtVNAXv +++ mktemp ++ local LAST_ERR=/tmp/tmp.Dl67tKP7Ln ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.do8WtVNAXv ++ cat /tmp/tmp.Dl67tKP7Ln ++ rm /tmp/tmp.do8WtVNAXv /tmp/tmp.Dl67tKP7Ln ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-rs0-2 + local pod=some-name-rs0-2 + set +o xtrace waiting for pod/some-name-rs0-2 to be ready.............OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.u21VxVxpWd +++ mktemp ++ local LAST_ERR=/tmp/tmp.DyHEv0iACx ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.u21VxVxpWd ++ cat /tmp/tmp.DyHEv0iACx ++ rm /tmp/tmp.u21VxVxpWd /tmp/tmp.DyHEv0iACx ++ return 0 + [[ '' == \t\r\u\e ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KZtMtRCN4l +++ mktemp ++ local LAST_ERR=/tmp/tmp.lsagY2J0tj ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.KZtMtRCN4l ++ cat /tmp/tmp.lsagY2J0tj ++ rm /tmp/tmp.KZtMtRCN4l /tmp/tmp.lsagY2J0tj ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness + sleep 20 + compare_kubectl statefulset/some-name-rs0 + local resource=statefulset/some-name-rs0 + local postfix= + local skip_generation_check= + local expected_result=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/e2e-tests/pvc-auto-resize/compare/statefulset_some-name-rs0.yml + local new_result=/tmp/tmp.xS08x4x5XN/statefulset_some-name-rs0.yml + '[' -n '' -a -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/e2e-tests/pvc-auto-resize/compare/statefulset_some-name-rs0-oc.yml ']' + kubectl_bin get -o yaml statefulset/some-name-rs0 + yq eval ' del(.metadata.ownerReferences[].apiVersion) | del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "NAMESPACE")) | del(.metadata.selfLink) | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.status) | (.. | select(tag == "!!str")) |= sub("pvc-auto-resize-27302", "NAME_SPACE") | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "extensions/v1beta1")) = "apps/v1" | (.. | select(. == "batch/v1beta1")) = "batch/v1" ' - ++ mktemp + local LAST_OUT=/tmp/tmp.HhKGCjEld4 ++ mktemp + local LAST_ERR=/tmp/tmp.YRanYx5AuJ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get -o yaml statefulset/some-name-rs0 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.HhKGCjEld4 + cat /tmp/tmp.YRanYx5AuJ + rm /tmp/tmp.HhKGCjEld4 /tmp/tmp.YRanYx5AuJ + return 0 + yq -i eval 'del(.spec.persistentVolumeClaimRetentionPolicy)' /tmp/tmp.xS08x4x5XN/statefulset_some-name-rs0.yml + version_gt 1.22 ++ echo '1.32 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + yq -i eval 'del(.spec.internalTrafficPolicy)' /tmp/tmp.xS08x4x5XN/statefulset_some-name-rs0.yml + yq -i eval 'del(.spec.allocateLoadBalancerNodePorts)' /tmp/tmp.xS08x4x5XN/statefulset_some-name-rs0.yml + [[ /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/e2e-tests/pvc-auto-resize/compare/statefulset_some-name-rs0.yml == */cronjob* ]] + '[' -n '' ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/e2e-tests/pvc-auto-resize/compare/statefulset_some-name-rs0.yml /tmp/tmp.xS08x4x5XN/statefulset_some-name-rs0.yml + log 'compare_kubectl: statefulset/some-name-rs0 OK' + set +o xtrace [2026-02-19T14:51:22+0000] compare_kubectl: statefulset/some-name-rs0 OK + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + run_mongo 'db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})' userAdmin:userAdmin123456@some-name-rs0.pvc-auto-resize-27302 + local 'command=db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})' + local uri=userAdmin:userAdmin123456@some-name-rs0.pvc-auto-resize-27302 + local driver=mongodb+srv + local suffix=.svc.cluster.local + local mongo_flag= + local replica_set=rs0 + [[ userAdmin:userAdmin123456@some-name-rs0.pvc-auto-resize-27302 == *cfg* ]] ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Xqz0VQSuti +++ mktemp ++ local LAST_ERR=/tmp/tmp.ICRbRd6Ncp ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Xqz0VQSuti ++ cat /tmp/tmp.ICRbRd6Ncp ++ rm /tmp/tmp.Xqz0VQSuti /tmp/tmp.ICRbRd6Ncp ++ return 0 + local client_container=psmdb-client-86cb5d8484-p4cbr + kubectl_bin exec psmdb-client-86cb5d8484-p4cbr -- bash -c 'printf '\''db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})\n'\'' | mongo mongodb+srv://userAdmin:userAdmin123456@some-name-rs0.pvc-auto-resize-27302.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.oT4fGnwIyr ++ mktemp + local LAST_ERR=/tmp/tmp.xfsSRBXkaz + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-86cb5d8484-p4cbr -- bash -c 'printf '\''db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})\n'\'' | mongo mongodb+srv://userAdmin:userAdmin123456@some-name-rs0.pvc-auto-resize-27302.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.oT4fGnwIyr Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-rs0-2.some-name-rs0.pvc-auto-resize-27302.svc.cluster.local:27017,some-name-rs0-0.some-name-rs0.pvc-auto-resize-27302.svc.cluster.local:27017,some-name-rs0-1.some-name-rs0.pvc-auto-resize-27302.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb&replicaSet=rs0&ssl=false Implicit session: session { "id" : UUID("87ce1315-00a2-40e0-a2a9-a25724368bdf") } Percona Server for MongoDB server version: v8.0.19-7 WARNING: shell and server versions do not match Successfully added user: { "user" : "myApp", "roles" : [ { "db" : "myApp", "role" : "readWrite" } ] } bye + cat /tmp/tmp.xfsSRBXkaz + rm /tmp/tmp.oT4fGnwIyr /tmp/tmp.xfsSRBXkaz + return 0 + run_mongo 'use myApp\n db.test.insert({ x: 100500 })' myApp:myPass@some-name-rs0.pvc-auto-resize-27302 + local 'command=use myApp\n db.test.insert({ x: 100500 })' + local uri=myApp:myPass@some-name-rs0.pvc-auto-resize-27302 + local driver=mongodb+srv + local suffix=.svc.cluster.local + local mongo_flag= + local replica_set=rs0 + [[ myApp:myPass@some-name-rs0.pvc-auto-resize-27302 == *cfg* ]] ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XroTRPogN9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.FJv1s5t111 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.XroTRPogN9 ++ cat /tmp/tmp.FJv1s5t111 ++ rm /tmp/tmp.XroTRPogN9 /tmp/tmp.FJv1s5t111 ++ return 0 + local client_container=psmdb-client-86cb5d8484-p4cbr + kubectl_bin exec psmdb-client-86cb5d8484-p4cbr -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb+srv://myApp:myPass@some-name-rs0.pvc-auto-resize-27302.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.Wa2G0hLOCe ++ mktemp + local LAST_ERR=/tmp/tmp.mC1uO7mtbn + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-86cb5d8484-p4cbr -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb+srv://myApp:myPass@some-name-rs0.pvc-auto-resize-27302.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Wa2G0hLOCe Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-rs0-0.some-name-rs0.pvc-auto-resize-27302.svc.cluster.local:27017,some-name-rs0-1.some-name-rs0.pvc-auto-resize-27302.svc.cluster.local:27017,some-name-rs0-2.some-name-rs0.pvc-auto-resize-27302.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb&replicaSet=rs0&ssl=false Implicit session: session { "id" : UUID("23b7afa0-8301-4969-a009-dd3c74474f31") } Percona Server for MongoDB server version: v8.0.19-7 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.mC1uO7mtbn + rm /tmp/tmp.Wa2G0hLOCe /tmp/tmp.mC1uO7mtbn + return 0 + desc 'verify initial PVC size is 1Gi' + set +o xtrace ----------------------------------------------------------------------------------- verify initial PVC size is 1Gi ----------------------------------------------------------------------------------- ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XViif5g7NB +++ mktemp ++ local LAST_ERR=/tmp/tmp.rm959RWurr ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.XViif5g7NB ++ cat /tmp/tmp.rm959RWurr ++ rm /tmp/tmp.XViif5g7NB /tmp/tmp.rm959RWurr ++ return 0 + initial_size=1Gi + echo 'Initial PVC size: 1Gi' Initial PVC size: 1Gi + [[ 1Gi != \1\G\i ]] + desc 'fill disk to trigger autoscaling threshold 50 percent' + set +o xtrace ----------------------------------------------------------------------------------- fill disk to trigger autoscaling threshold 50 percent ----------------------------------------------------------------------------------- + fill_disk_to_threshold some-name-rs0-0 50 600 + local pod=some-name-rs0-0 + local threshold=50 + local size_mb=600 + echo 'Filling disk on some-name-rs0-0 to approximately 50% (creating 600MB file)' Filling disk on some-name-rs0-0 to approximately 50% (creating 600MB file) + kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=600 2>/dev/null || true' ++ mktemp + local LAST_OUT=/tmp/tmp.qvvAIcPRxl ++ mktemp + local LAST_ERR=/tmp/tmp.kf2ELcLKpZ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=600 2>/dev/null || true' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.qvvAIcPRxl + cat /tmp/tmp.kf2ELcLKpZ + rm /tmp/tmp.qvvAIcPRxl /tmp/tmp.kf2ELcLKpZ + return 0 + local usage ++ kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KcF2ieuyFa +++ mktemp ++ local LAST_ERR=/tmp/tmp.51i29WFRqL ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.KcF2ieuyFa ++ cat /tmp/tmp.51i29WFRqL ++ rm /tmp/tmp.KcF2ieuyFa /tmp/tmp.51i29WFRqL ++ return 0 + usage=85 + echo 'Current disk usage on some-name-rs0-0: 85%' Current disk usage on some-name-rs0-0: 85% + desc 'wait for auto-resize to trigger' + set +o xtrace ----------------------------------------------------------------------------------- wait for auto-resize to trigger ----------------------------------------------------------------------------------- + wait_for_auto_resize mongod-data-some-name-rs0-0 3Gi 60 10 + local pvc=mongod-data-some-name-rs0-0 + local expected_size=3Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 3Gi' Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 3Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HZGdAsZOvD +++ mktemp ++ local LAST_ERR=/tmp/tmp.MpMlXg3mqv ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.HZGdAsZOvD ++ cat /tmp/tmp.MpMlXg3mqv ++ rm /tmp/tmp.HZGdAsZOvD /tmp/tmp.MpMlXg3mqv ++ return 0 + [[ 1Gi == \3\G\i ]] + [[ 0 -ge 60 ]] + echo -n . .+ sleep 10 + retry=1 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.r5GqL27ukk +++ mktemp ++ local LAST_ERR=/tmp/tmp.EHSaZZEsX8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.r5GqL27ukk ++ cat /tmp/tmp.EHSaZZEsX8 ++ rm /tmp/tmp.r5GqL27ukk /tmp/tmp.EHSaZZEsX8 ++ return 0 + [[ 1Gi == \3\G\i ]] + [[ 1 -ge 60 ]] + echo -n . .+ sleep 10 + retry=2 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.um0w976x61 +++ mktemp ++ local LAST_ERR=/tmp/tmp.jiRzEMbixl ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.um0w976x61 ++ cat /tmp/tmp.jiRzEMbixl ++ rm /tmp/tmp.um0w976x61 /tmp/tmp.jiRzEMbixl ++ return 0 + [[ 1Gi == \3\G\i ]] + [[ 2 -ge 60 ]] + echo -n . .+ sleep 10 + retry=3 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tYIHWTcYhv +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ow5Qnjmf1a ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.tYIHWTcYhv ++ cat /tmp/tmp.Ow5Qnjmf1a ++ rm /tmp/tmp.tYIHWTcYhv /tmp/tmp.Ow5Qnjmf1a ++ return 0 + [[ 3Gi == \3\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-0 successfully resized to 3Gi' PVC mongod-data-some-name-rs0-0 successfully resized to 3Gi + wait_for_auto_resize mongod-data-some-name-rs0-1 3Gi 60 10 + local pvc=mongod-data-some-name-rs0-1 + local expected_size=3Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-1 to auto-resize to 3Gi' Waiting for PVC mongod-data-some-name-rs0-1 to auto-resize to 3Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7qt7Ip7UFV +++ mktemp ++ local LAST_ERR=/tmp/tmp.7IRPFOKU15 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.7qt7Ip7UFV ++ cat /tmp/tmp.7IRPFOKU15 ++ rm /tmp/tmp.7qt7Ip7UFV /tmp/tmp.7IRPFOKU15 ++ return 0 + [[ 3Gi == \3\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-1 successfully resized to 3Gi' PVC mongod-data-some-name-rs0-1 successfully resized to 3Gi + wait_for_auto_resize mongod-data-some-name-rs0-2 3Gi 60 10 + local pvc=mongod-data-some-name-rs0-2 + local expected_size=3Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-2 to auto-resize to 3Gi' Waiting for PVC mongod-data-some-name-rs0-2 to auto-resize to 3Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-2 ++ local pvc=mongod-data-some-name-rs0-2 ++ kubectl_bin get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MIuFJALEOY +++ mktemp ++ local LAST_ERR=/tmp/tmp.OhC4SBUx0h ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.MIuFJALEOY ++ cat /tmp/tmp.OhC4SBUx0h ++ rm /tmp/tmp.MIuFJALEOY /tmp/tmp.OhC4SBUx0h ++ return 0 + [[ 3Gi == \3\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-2 successfully resized to 3Gi' PVC mongod-data-some-name-rs0-2 successfully resized to 3Gi + desc 'verify autoscaling status is updated' + set +o xtrace ----------------------------------------------------------------------------------- verify autoscaling status is updated ----------------------------------------------------------------------------------- + check_autoscaling_status some-name 1 mongod-data-some-name-rs0-0 mongod-data-some-name-rs0-1 mongod-data-some-name-rs0-2 + local cluster=some-name + local expected_resize_count=1 + shift 2 + pvc_names=('mongod-data-some-name-rs0-0' 'mongod-data-some-name-rs0-1' 'mongod-data-some-name-rs0-2') + local pvc_names + local max_retry=30 + local sleep_time=10 + echo 'Checking autoscaling status for PVCs: mongod-data-some-name-rs0-0 mongod-data-some-name-rs0-1 mongod-data-some-name-rs0-2' Checking autoscaling status for PVCs: mongod-data-some-name-rs0-0 mongod-data-some-name-rs0-1 mongod-data-some-name-rs0-2 + echo 'Expected resize count per PVC: 1' Expected resize count per PVC: 1 + for pvc_name in "${pvc_names[@]}" + local retry=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-0.resizeCount}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.naOV2a7sIH +++ mktemp ++ local LAST_ERR=/tmp/tmp.6ikpFoBT1C ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-0.resizeCount}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.naOV2a7sIH ++ cat /tmp/tmp.6ikpFoBT1C ++ rm /tmp/tmp.naOV2a7sIH /tmp/tmp.6ikpFoBT1C ++ return 0 + [[ '' -eq 1 ]] + [[ 0 -ge 30 ]] + echo -n . .+ sleep 10 + retry=1 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-0.resizeCount}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JpuPvLnfwm +++ mktemp ++ local LAST_ERR=/tmp/tmp.7iPJlm1J4b ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-0.resizeCount}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.JpuPvLnfwm ++ cat /tmp/tmp.7iPJlm1J4b ++ rm /tmp/tmp.JpuPvLnfwm /tmp/tmp.7iPJlm1J4b ++ return 0 + [[ 1 -eq 1 ]] + for pvc_name in "${pvc_names[@]}" + local retry=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-1.resizeCount}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5WdDnZnnFS +++ mktemp ++ local LAST_ERR=/tmp/tmp.ITAetlb9sj ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-1.resizeCount}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5WdDnZnnFS ++ cat /tmp/tmp.ITAetlb9sj ++ rm /tmp/tmp.5WdDnZnnFS /tmp/tmp.ITAetlb9sj ++ return 0 + [[ 1 -eq 1 ]] + for pvc_name in "${pvc_names[@]}" + local retry=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-2.resizeCount}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pdrj3zF5lf +++ mktemp ++ local LAST_ERR=/tmp/tmp.gSPW8u3Wb5 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-2.resizeCount}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.pdrj3zF5lf ++ cat /tmp/tmp.gSPW8u3Wb5 ++ rm /tmp/tmp.pdrj3zF5lf /tmp/tmp.gSPW8u3Wb5 ++ return 0 + [[ 1 -eq 1 ]] + desc 'wait for cluster consistency after resize' + set +o xtrace ----------------------------------------------------------------------------------- wait for cluster consistency after resize ----------------------------------------------------------------------------------- + wait_cluster_consistency some-name 64 + local cluster_name=some-name + local wait_time=64 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dofkxrmcWp +++ mktemp ++ local LAST_ERR=/tmp/tmp.KmBrzOzOtB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.dofkxrmcWp ++ cat /tmp/tmp.KmBrzOzOtB ++ rm /tmp/tmp.dofkxrmcWp /tmp/tmp.KmBrzOzOtB ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo .OK .OK + desc 'verify cluster is ready' + set +o xtrace ----------------------------------------------------------------------------------- verify cluster is ready ----------------------------------------------------------------------------------- + kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' + grep -q ready ++ mktemp + local LAST_OUT=/tmp/tmp.McHg7UQOxx ++ mktemp + local LAST_ERR=/tmp/tmp.hV3Flj4MiA + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get psmdb some-name -o 'jsonpath={.status.state}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.McHg7UQOxx + cat /tmp/tmp.hV3Flj4MiA + rm /tmp/tmp.McHg7UQOxx /tmp/tmp.hV3Flj4MiA + return 0 + desc 'test namespace resource quota enforcement' + set +o xtrace ----------------------------------------------------------------------------------- test namespace resource quota enforcement ----------------------------------------------------------------------------------- + apply_resourcequota 9Gi + local quota=9Gi + local default_sc ++ get_default_storageclass ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gUcOpDd90Q +++ mktemp ++ local LAST_ERR=/tmp/tmp.DSH6lRlAbN ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.gUcOpDd90Q ++ cat /tmp/tmp.DSH6lRlAbN ++ rm /tmp/tmp.gUcOpDd90Q /tmp/tmp.DSH6lRlAbN ++ return 0 + default_sc=standard-rwo + echo 'Applying resourcequota for default storageclass standard-rwo with quota 9Gi' Applying resourcequota for default storageclass standard-rwo with quota 9Gi + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/e2e-tests/pvc-auto-resize/conf/resourcequota.yml + sed s/QUOTA/9Gi/ + sed s/STORAGECLASS/standard-rwo/ + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.CGs0hkJZOX ++ mktemp + local LAST_ERR=/tmp/tmp.PMZXbyZW4G + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.CGs0hkJZOX resourcequota/default-storage-quota created + cat /tmp/tmp.PMZXbyZW4G + rm /tmp/tmp.CGs0hkJZOX /tmp/tmp.PMZXbyZW4G + return 0 + echo 'Filling disk again to trigger autoscaling with quota limit' Filling disk again to trigger autoscaling with quota limit + fill_disk_to_threshold some-name-rs0-0 50 1700 + local pod=some-name-rs0-0 + local threshold=50 + local size_mb=1700 + echo 'Filling disk on some-name-rs0-0 to approximately 50% (creating 1700MB file)' Filling disk on some-name-rs0-0 to approximately 50% (creating 1700MB file) + kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=1700 2>/dev/null || true' ++ mktemp + local LAST_OUT=/tmp/tmp.0QmYjjYbbU ++ mktemp + local LAST_ERR=/tmp/tmp.BTHcYatvKC + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=1700 2>/dev/null || true' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.0QmYjjYbbU + cat /tmp/tmp.BTHcYatvKC + rm /tmp/tmp.0QmYjjYbbU /tmp/tmp.BTHcYatvKC + return 0 + local usage ++ kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1it1r3Yl7K +++ mktemp ++ local LAST_ERR=/tmp/tmp.QmuYY0uZxP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.1it1r3Yl7K ++ cat /tmp/tmp.QmuYY0uZxP ++ rm /tmp/tmp.1it1r3Yl7K /tmp/tmp.QmuYY0uZxP ++ return 0 + usage=65 + echo 'Current disk usage on some-name-rs0-0: 65%' Current disk usage on some-name-rs0-0: 65% + echo 'Waiting to verify PVC does not resize beyond quota (should stay at 3Gi)' Waiting to verify PVC does not resize beyond quota (should stay at 3Gi) + sleep 30 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3ZklTsMOP9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.04KyPO65TH ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.3ZklTsMOP9 ++ cat /tmp/tmp.04KyPO65TH ++ rm /tmp/tmp.3ZklTsMOP9 /tmp/tmp.04KyPO65TH ++ return 0 + current_size=3Gi + [[ 3Gi != \3\G\i ]] + desc 'increase quota to allow further autoscaling' + set +o xtrace ----------------------------------------------------------------------------------- increase quota to allow further autoscaling ----------------------------------------------------------------------------------- + apply_resourcequota 15Gi + local quota=15Gi + local default_sc ++ get_default_storageclass ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.miED1EzkE1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.kUj3WdorM4 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.miED1EzkE1 ++ cat /tmp/tmp.kUj3WdorM4 ++ rm /tmp/tmp.miED1EzkE1 /tmp/tmp.kUj3WdorM4 ++ return 0 + default_sc=standard-rwo + echo 'Applying resourcequota for default storageclass standard-rwo with quota 15Gi' Applying resourcequota for default storageclass standard-rwo with quota 15Gi + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/e2e-tests/pvc-auto-resize/conf/resourcequota.yml + sed s/STORAGECLASS/standard-rwo/ + sed s/QUOTA/15Gi/ + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.LhHeX2rLu2 ++ mktemp + local LAST_ERR=/tmp/tmp.XCgeiYnVrw + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.LhHeX2rLu2 resourcequota/default-storage-quota configured + cat /tmp/tmp.XCgeiYnVrw + rm /tmp/tmp.LhHeX2rLu2 /tmp/tmp.XCgeiYnVrw + return 0 + wait_for_auto_resize mongod-data-some-name-rs0-0 5Gi 60 10 + local pvc=mongod-data-some-name-rs0-0 + local expected_size=5Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 5Gi' Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 5Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.l9nMic11pd +++ mktemp ++ local LAST_ERR=/tmp/tmp.t6QmnUxejq ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.l9nMic11pd ++ cat /tmp/tmp.t6QmnUxejq ++ rm /tmp/tmp.l9nMic11pd /tmp/tmp.t6QmnUxejq ++ return 0 + [[ 3Gi == \5\G\i ]] + [[ 0 -ge 60 ]] + echo -n . .+ sleep 10 + retry=1 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8SlXlKeuai +++ mktemp ++ local LAST_ERR=/tmp/tmp.Df9gH79OT3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.8SlXlKeuai ++ cat /tmp/tmp.Df9gH79OT3 ++ rm /tmp/tmp.8SlXlKeuai /tmp/tmp.Df9gH79OT3 ++ return 0 + [[ 3Gi == \5\G\i ]] + [[ 1 -ge 60 ]] + echo -n . .+ sleep 10 + retry=2 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Zdu7x6rSRn +++ mktemp ++ local LAST_ERR=/tmp/tmp.DBqusG8nWQ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Zdu7x6rSRn ++ cat /tmp/tmp.DBqusG8nWQ ++ rm /tmp/tmp.Zdu7x6rSRn /tmp/tmp.DBqusG8nWQ ++ return 0 + [[ 3Gi == \5\G\i ]] + [[ 2 -ge 60 ]] + echo -n . .+ sleep 10 + retry=3 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KnQswwFdcq +++ mktemp ++ local LAST_ERR=/tmp/tmp.tiKpwn2U2r ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.KnQswwFdcq ++ cat /tmp/tmp.tiKpwn2U2r ++ rm /tmp/tmp.KnQswwFdcq /tmp/tmp.tiKpwn2U2r ++ return 0 + [[ 3Gi == \5\G\i ]] + [[ 3 -ge 60 ]] + echo -n . .+ sleep 10 + retry=4 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7k06E3c62h +++ mktemp ++ local LAST_ERR=/tmp/tmp.fxXhMt4pn6 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.7k06E3c62h ++ cat /tmp/tmp.fxXhMt4pn6 ++ rm /tmp/tmp.7k06E3c62h /tmp/tmp.fxXhMt4pn6 ++ return 0 + [[ 3Gi == \5\G\i ]] + [[ 4 -ge 60 ]] + echo -n . .+ sleep 10 + retry=5 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.F6LucSwPwp +++ mktemp ++ local LAST_ERR=/tmp/tmp.nD9JCFKFDd ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.F6LucSwPwp ++ cat /tmp/tmp.nD9JCFKFDd ++ rm /tmp/tmp.F6LucSwPwp /tmp/tmp.nD9JCFKFDd ++ return 0 + [[ 5Gi == \5\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-0 successfully resized to 5Gi' PVC mongod-data-some-name-rs0-0 successfully resized to 5Gi + wait_for_auto_resize mongod-data-some-name-rs0-1 5Gi 60 10 + local pvc=mongod-data-some-name-rs0-1 + local expected_size=5Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-1 to auto-resize to 5Gi' Waiting for PVC mongod-data-some-name-rs0-1 to auto-resize to 5Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XUExx5qnPC +++ mktemp ++ local LAST_ERR=/tmp/tmp.3AYzCvQgvI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.XUExx5qnPC ++ cat /tmp/tmp.3AYzCvQgvI ++ rm /tmp/tmp.XUExx5qnPC /tmp/tmp.3AYzCvQgvI ++ return 0 + [[ 5Gi == \5\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-1 successfully resized to 5Gi' PVC mongod-data-some-name-rs0-1 successfully resized to 5Gi + wait_for_auto_resize mongod-data-some-name-rs0-2 5Gi 60 10 + local pvc=mongod-data-some-name-rs0-2 + local expected_size=5Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-2 to auto-resize to 5Gi' Waiting for PVC mongod-data-some-name-rs0-2 to auto-resize to 5Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-2 ++ local pvc=mongod-data-some-name-rs0-2 ++ kubectl_bin get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.adaGf0UhWp +++ mktemp ++ local LAST_ERR=/tmp/tmp.kFBLbiTbZw ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.adaGf0UhWp ++ cat /tmp/tmp.kFBLbiTbZw ++ rm /tmp/tmp.adaGf0UhWp /tmp/tmp.kFBLbiTbZw ++ return 0 + [[ 3Gi == \5\G\i ]] + [[ 0 -ge 60 ]] + echo -n . .+ sleep 10 + retry=1 ++ get_pvc_size mongod-data-some-name-rs0-2 ++ local pvc=mongod-data-some-name-rs0-2 ++ kubectl_bin get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.i1ea7SueZk +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ut4T2KO8LV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.i1ea7SueZk ++ cat /tmp/tmp.Ut4T2KO8LV ++ rm /tmp/tmp.i1ea7SueZk /tmp/tmp.Ut4T2KO8LV ++ return 0 + [[ 5Gi == \5\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-2 successfully resized to 5Gi' PVC mongod-data-some-name-rs0-2 successfully resized to 5Gi + desc 'verify cluster consistency after quota test' + set +o xtrace ----------------------------------------------------------------------------------- verify cluster consistency after quota test ----------------------------------------------------------------------------------- + wait_cluster_consistency some-name 64 + local cluster_name=some-name + local wait_time=64 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JgCCbysNgn +++ mktemp ++ local LAST_ERR=/tmp/tmp.clXVy8qoD0 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.JgCCbysNgn ++ cat /tmp/tmp.clXVy8qoD0 ++ rm /tmp/tmp.JgCCbysNgn /tmp/tmp.clXVy8qoD0 ++ return 0 + [[ error == \r\e\a\d\y ]] + let retry+=1 + '[' 1 -ge 64 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1WpqSvKtXn +++ mktemp ++ local LAST_ERR=/tmp/tmp.Y0353T9jtS ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.1WpqSvKtXn ++ cat /tmp/tmp.Y0353T9jtS ++ rm /tmp/tmp.1WpqSvKtXn /tmp/tmp.Y0353T9jtS ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo .OK .OK + desc 'test maxSize limit' + set +o xtrace ----------------------------------------------------------------------------------- test maxSize limit ----------------------------------------------------------------------------------- + apply_resourcequota 30Gi + local quota=30Gi + local default_sc ++ get_default_storageclass ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MjrSXwgbDB +++ mktemp ++ local LAST_ERR=/tmp/tmp.hkzrudE2Fe ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.MjrSXwgbDB ++ cat /tmp/tmp.hkzrudE2Fe ++ rm /tmp/tmp.MjrSXwgbDB /tmp/tmp.hkzrudE2Fe ++ return 0 + default_sc=standard-rwo + echo 'Applying resourcequota for default storageclass standard-rwo with quota 30Gi' Applying resourcequota for default storageclass standard-rwo with quota 30Gi + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/e2e-tests/pvc-auto-resize/conf/resourcequota.yml + sed s/STORAGECLASS/standard-rwo/ + sed s/QUOTA/30Gi/ + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.MBoBFLXxdm ++ mktemp + local LAST_ERR=/tmp/tmp.RByA63MQdf + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.MBoBFLXxdm resourcequota/default-storage-quota configured + cat /tmp/tmp.RByA63MQdf + rm /tmp/tmp.MBoBFLXxdm /tmp/tmp.RByA63MQdf + return 0 + echo 'Continuing to fill disk to test maxSize enforcement' Continuing to fill disk to test maxSize enforcement + fill_disk_to_threshold some-name-rs0-0 50 2900 + local pod=some-name-rs0-0 + local threshold=50 + local size_mb=2900 + echo 'Filling disk on some-name-rs0-0 to approximately 50% (creating 2900MB file)' Filling disk on some-name-rs0-0 to approximately 50% (creating 2900MB file) + kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=2900 2>/dev/null || true' ++ mktemp + local LAST_OUT=/tmp/tmp.x4jctAuIQa ++ mktemp + local LAST_ERR=/tmp/tmp.KHjAGu8Rdp + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=2900 2>/dev/null || true' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.x4jctAuIQa + cat /tmp/tmp.KHjAGu8Rdp + rm /tmp/tmp.x4jctAuIQa /tmp/tmp.KHjAGu8Rdp + return 0 + local usage ++ kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zKoh3UHuVw +++ mktemp ++ local LAST_ERR=/tmp/tmp.NCuIAj2Jvn ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.zKoh3UHuVw ++ cat /tmp/tmp.NCuIAj2Jvn ++ rm /tmp/tmp.zKoh3UHuVw /tmp/tmp.NCuIAj2Jvn ++ return 0 + usage=63 + echo 'Current disk usage on some-name-rs0-0: 63%' Current disk usage on some-name-rs0-0: 63% + wait_for_auto_resize mongod-data-some-name-rs0-0 7Gi 60 10 + local pvc=mongod-data-some-name-rs0-0 + local expected_size=7Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 7Gi' Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 7Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fddrW1EWUu +++ mktemp ++ local LAST_ERR=/tmp/tmp.WsgKNP1QXF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.fddrW1EWUu ++ cat /tmp/tmp.WsgKNP1QXF ++ rm /tmp/tmp.fddrW1EWUu /tmp/tmp.WsgKNP1QXF ++ return 0 + [[ 5Gi == \7\G\i ]] + [[ 0 -ge 60 ]] + echo -n . .+ sleep 10 + retry=1 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.irlkU9DJyj +++ mktemp ++ local LAST_ERR=/tmp/tmp.J0aKpJGinV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.irlkU9DJyj ++ cat /tmp/tmp.J0aKpJGinV ++ rm /tmp/tmp.irlkU9DJyj /tmp/tmp.J0aKpJGinV ++ return 0 + [[ 7Gi == \7\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-0 successfully resized to 7Gi' PVC mongod-data-some-name-rs0-0 successfully resized to 7Gi + wait_for_auto_resize mongod-data-some-name-rs0-1 7Gi 60 10 + local pvc=mongod-data-some-name-rs0-1 + local expected_size=7Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-1 to auto-resize to 7Gi' Waiting for PVC mongod-data-some-name-rs0-1 to auto-resize to 7Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qibgtcIwzt +++ mktemp ++ local LAST_ERR=/tmp/tmp.q0DCowv2Pg ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.qibgtcIwzt ++ cat /tmp/tmp.q0DCowv2Pg ++ rm /tmp/tmp.qibgtcIwzt /tmp/tmp.q0DCowv2Pg ++ return 0 + [[ 5Gi == \7\G\i ]] + [[ 0 -ge 60 ]] + echo -n . .+ sleep 10 + retry=1 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Z4EMUsiRNV +++ mktemp ++ local LAST_ERR=/tmp/tmp.d6KbY66VPw ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Z4EMUsiRNV ++ cat /tmp/tmp.d6KbY66VPw ++ rm /tmp/tmp.Z4EMUsiRNV /tmp/tmp.d6KbY66VPw ++ return 0 + [[ 5Gi == \7\G\i ]] + [[ 1 -ge 60 ]] + echo -n . .+ sleep 10 + retry=2 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LVks5jJ09E +++ mktemp ++ local LAST_ERR=/tmp/tmp.Qk7ngCc2wc ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.LVks5jJ09E ++ cat /tmp/tmp.Qk7ngCc2wc ++ rm /tmp/tmp.LVks5jJ09E /tmp/tmp.Qk7ngCc2wc ++ return 0 + [[ 5Gi == \7\G\i ]] + [[ 2 -ge 60 ]] + echo -n . .+ sleep 10 + retry=3 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ty2r2MmMkt +++ mktemp ++ local LAST_ERR=/tmp/tmp.vH3UVLWwTF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ty2r2MmMkt ++ cat /tmp/tmp.vH3UVLWwTF ++ rm /tmp/tmp.ty2r2MmMkt /tmp/tmp.vH3UVLWwTF ++ return 0 + [[ 5Gi == \7\G\i ]] + [[ 3 -ge 60 ]] + echo -n . .+ sleep 10 + retry=4 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2iV7oACBTo +++ mktemp ++ local LAST_ERR=/tmp/tmp.q79pn58uQU ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.2iV7oACBTo ++ cat /tmp/tmp.q79pn58uQU ++ rm /tmp/tmp.2iV7oACBTo /tmp/tmp.q79pn58uQU ++ return 0 + [[ 7Gi == \7\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-1 successfully resized to 7Gi' PVC mongod-data-some-name-rs0-1 successfully resized to 7Gi + wait_for_auto_resize mongod-data-some-name-rs0-2 7Gi 60 10 + local pvc=mongod-data-some-name-rs0-2 + local expected_size=7Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-2 to auto-resize to 7Gi' Waiting for PVC mongod-data-some-name-rs0-2 to auto-resize to 7Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-2 ++ local pvc=mongod-data-some-name-rs0-2 ++ kubectl_bin get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jg6n67tfCT +++ mktemp ++ local LAST_ERR=/tmp/tmp.wpbWD5aGEU ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jg6n67tfCT ++ cat /tmp/tmp.wpbWD5aGEU ++ rm /tmp/tmp.jg6n67tfCT /tmp/tmp.wpbWD5aGEU ++ return 0 + [[ 7Gi == \7\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-2 successfully resized to 7Gi' PVC mongod-data-some-name-rs0-2 successfully resized to 7Gi + fill_disk_to_threshold some-name-rs0-0 50 4000 + local pod=some-name-rs0-0 + local threshold=50 + local size_mb=4000 + echo 'Filling disk on some-name-rs0-0 to approximately 50% (creating 4000MB file)' Filling disk on some-name-rs0-0 to approximately 50% (creating 4000MB file) + kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=4000 2>/dev/null || true' ++ mktemp + local LAST_OUT=/tmp/tmp.inhomR3pwR ++ mktemp + local LAST_ERR=/tmp/tmp.vSNlXUJvgN + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=4000 2>/dev/null || true' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.inhomR3pwR + cat /tmp/tmp.vSNlXUJvgN + rm /tmp/tmp.inhomR3pwR /tmp/tmp.vSNlXUJvgN + return 0 + local usage ++ kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4hP2HK9Ruq +++ mktemp ++ local LAST_ERR=/tmp/tmp.J56hUsQATz ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4hP2HK9Ruq ++ cat /tmp/tmp.J56hUsQATz ++ rm /tmp/tmp.4hP2HK9Ruq /tmp/tmp.J56hUsQATz ++ return 0 + usage=61 + echo 'Current disk usage on some-name-rs0-0: 61%' Current disk usage on some-name-rs0-0: 61% + wait_for_auto_resize mongod-data-some-name-rs0-0 9Gi 60 10 + local pvc=mongod-data-some-name-rs0-0 + local expected_size=9Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 9Gi' Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 9Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2PETTnAgMy +++ mktemp ++ local LAST_ERR=/tmp/tmp.IH1uCUCWxd ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.2PETTnAgMy ++ cat /tmp/tmp.IH1uCUCWxd ++ rm /tmp/tmp.2PETTnAgMy /tmp/tmp.IH1uCUCWxd ++ return 0 + [[ 7Gi == \9\G\i ]] + [[ 0 -ge 60 ]] + echo -n . .+ sleep 10 + retry=1 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NFekRK8488 +++ mktemp ++ local LAST_ERR=/tmp/tmp.6VmwbLpP53 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.NFekRK8488 ++ cat /tmp/tmp.6VmwbLpP53 ++ rm /tmp/tmp.NFekRK8488 /tmp/tmp.6VmwbLpP53 ++ return 0 + [[ 7Gi == \9\G\i ]] + [[ 1 -ge 60 ]] + echo -n . .+ sleep 10 + retry=2 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vpFOeVr0TV +++ mktemp ++ local LAST_ERR=/tmp/tmp.il1Zw1i5GO ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.vpFOeVr0TV ++ cat /tmp/tmp.il1Zw1i5GO ++ rm /tmp/tmp.vpFOeVr0TV /tmp/tmp.il1Zw1i5GO ++ return 0 + [[ 7Gi == \9\G\i ]] + [[ 2 -ge 60 ]] + echo -n . .+ sleep 10 + retry=3 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nkGzgvxh9M +++ mktemp ++ local LAST_ERR=/tmp/tmp.3qT9foN53l ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.nkGzgvxh9M ++ cat /tmp/tmp.3qT9foN53l ++ rm /tmp/tmp.nkGzgvxh9M /tmp/tmp.3qT9foN53l ++ return 0 + [[ 7Gi == \9\G\i ]] + [[ 3 -ge 60 ]] + echo -n . .+ sleep 10 + retry=4 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZsjHDWRDyL +++ mktemp ++ local LAST_ERR=/tmp/tmp.ARX91w2i9o ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ZsjHDWRDyL ++ cat /tmp/tmp.ARX91w2i9o ++ rm /tmp/tmp.ZsjHDWRDyL /tmp/tmp.ARX91w2i9o ++ return 0 + [[ 9Gi == \9\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-0 successfully resized to 9Gi' PVC mongod-data-some-name-rs0-0 successfully resized to 9Gi + wait_for_auto_resize mongod-data-some-name-rs0-1 9Gi 60 10 + local pvc=mongod-data-some-name-rs0-1 + local expected_size=9Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-1 to auto-resize to 9Gi' Waiting for PVC mongod-data-some-name-rs0-1 to auto-resize to 9Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YkIEnRnOwc +++ mktemp ++ local LAST_ERR=/tmp/tmp.L0SAA8GpXB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.YkIEnRnOwc ++ cat /tmp/tmp.L0SAA8GpXB ++ rm /tmp/tmp.YkIEnRnOwc /tmp/tmp.L0SAA8GpXB ++ return 0 + [[ 9Gi == \9\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-1 successfully resized to 9Gi' PVC mongod-data-some-name-rs0-1 successfully resized to 9Gi + wait_for_auto_resize mongod-data-some-name-rs0-2 9Gi 60 10 + local pvc=mongod-data-some-name-rs0-2 + local expected_size=9Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-2 to auto-resize to 9Gi' Waiting for PVC mongod-data-some-name-rs0-2 to auto-resize to 9Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-2 ++ local pvc=mongod-data-some-name-rs0-2 ++ kubectl_bin get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.x0FJlFc1pO +++ mktemp ++ local LAST_ERR=/tmp/tmp.OAOpwvynnu ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.x0FJlFc1pO ++ cat /tmp/tmp.OAOpwvynnu ++ rm /tmp/tmp.x0FJlFc1pO /tmp/tmp.OAOpwvynnu ++ return 0 + [[ 7Gi == \9\G\i ]] + [[ 0 -ge 60 ]] + echo -n . .+ sleep 10 + retry=1 ++ get_pvc_size mongod-data-some-name-rs0-2 ++ local pvc=mongod-data-some-name-rs0-2 ++ kubectl_bin get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uuX49sI1m9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.ffPRvmQml6 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.uuX49sI1m9 ++ cat /tmp/tmp.ffPRvmQml6 ++ rm /tmp/tmp.uuX49sI1m9 /tmp/tmp.ffPRvmQml6 ++ return 0 + [[ 9Gi == \9\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-2 successfully resized to 9Gi' PVC mongod-data-some-name-rs0-2 successfully resized to 9Gi + fill_disk_to_threshold some-name-rs0-0 50 5200 + local pod=some-name-rs0-0 + local threshold=50 + local size_mb=5200 + echo 'Filling disk on some-name-rs0-0 to approximately 50% (creating 5200MB file)' Filling disk on some-name-rs0-0 to approximately 50% (creating 5200MB file) + kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=5200 2>/dev/null || true' ++ mktemp + local LAST_OUT=/tmp/tmp.jtCnMJcNzL ++ mktemp + local LAST_ERR=/tmp/tmp.nKDp55Iwta + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=5200 2>/dev/null || true' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.jtCnMJcNzL + cat /tmp/tmp.nKDp55Iwta + rm /tmp/tmp.jtCnMJcNzL /tmp/tmp.nKDp55Iwta + return 0 + local usage ++ kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RMWwtPlzRH +++ mktemp ++ local LAST_ERR=/tmp/tmp.pHMMZWqJLB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.RMWwtPlzRH ++ cat /tmp/tmp.pHMMZWqJLB ++ rm /tmp/tmp.RMWwtPlzRH /tmp/tmp.pHMMZWqJLB ++ return 0 + usage=61 + echo 'Current disk usage on some-name-rs0-0: 61%' Current disk usage on some-name-rs0-0: 61% ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.154VrBKoKR +++ mktemp ++ local LAST_ERR=/tmp/tmp.4jKpIVejt4 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.154VrBKoKR ++ cat /tmp/tmp.4jKpIVejt4 ++ rm /tmp/tmp.154VrBKoKR /tmp/tmp.4jKpIVejt4 ++ return 0 + final_size=9Gi + [[ 9Gi == \1\0\G\i ]] + [[ 9Gi == \9\G\i ]] + echo 'PVC is at 9Gi, will resize to 10Gi (maxSize)' PVC is at 9Gi, will resize to 10Gi (maxSize) + wait_for_auto_resize mongod-data-some-name-rs0-0 10Gi 60 10 + local pvc=mongod-data-some-name-rs0-0 + local expected_size=10Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 10Gi' Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 10Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ts7O319GbG +++ mktemp ++ local LAST_ERR=/tmp/tmp.bsmgTACSRG ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Ts7O319GbG ++ cat /tmp/tmp.bsmgTACSRG ++ rm /tmp/tmp.Ts7O319GbG /tmp/tmp.bsmgTACSRG ++ return 0 + [[ 9Gi == \1\0\G\i ]] + [[ 0 -ge 60 ]] + echo -n . .+ sleep 10 + retry=1 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.l4A7Hs8lBN +++ mktemp ++ local LAST_ERR=/tmp/tmp.3uFmaf12ag ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.l4A7Hs8lBN ++ cat /tmp/tmp.3uFmaf12ag ++ rm /tmp/tmp.l4A7Hs8lBN /tmp/tmp.3uFmaf12ag ++ return 0 + [[ 9Gi == \1\0\G\i ]] + [[ 1 -ge 60 ]] + echo -n . .+ sleep 10 + retry=2 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uWEA5z2WZy +++ mktemp ++ local LAST_ERR=/tmp/tmp.soVPLc7qj6 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.uWEA5z2WZy ++ cat /tmp/tmp.soVPLc7qj6 ++ rm /tmp/tmp.uWEA5z2WZy /tmp/tmp.soVPLc7qj6 ++ return 0 + [[ 9Gi == \1\0\G\i ]] + [[ 2 -ge 60 ]] + echo -n . .+ sleep 10 + retry=3 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yG2WdhhoGj +++ mktemp ++ local LAST_ERR=/tmp/tmp.BowcKaeMhH ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.yG2WdhhoGj ++ cat /tmp/tmp.BowcKaeMhH ++ rm /tmp/tmp.yG2WdhhoGj /tmp/tmp.BowcKaeMhH ++ return 0 + [[ 10Gi == \1\0\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-0 successfully resized to 10Gi' PVC mongod-data-some-name-rs0-0 successfully resized to 10Gi + wait_for_auto_resize mongod-data-some-name-rs0-1 10Gi 60 10 + local pvc=mongod-data-some-name-rs0-1 + local expected_size=10Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-1 to auto-resize to 10Gi' Waiting for PVC mongod-data-some-name-rs0-1 to auto-resize to 10Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.f7MWk60lo1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.c2Ry8BVmAp ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.f7MWk60lo1 ++ cat /tmp/tmp.c2Ry8BVmAp ++ rm /tmp/tmp.f7MWk60lo1 /tmp/tmp.c2Ry8BVmAp ++ return 0 + [[ 9Gi == \1\0\G\i ]] + [[ 0 -ge 60 ]] + echo -n . .+ sleep 10 + retry=1 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QtB9T78nt6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.xcrU9kHtU8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.QtB9T78nt6 ++ cat /tmp/tmp.xcrU9kHtU8 ++ rm /tmp/tmp.QtB9T78nt6 /tmp/tmp.xcrU9kHtU8 ++ return 0 + [[ 9Gi == \1\0\G\i ]] + [[ 1 -ge 60 ]] + echo -n . .+ sleep 10 + retry=2 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.y9wzEDK87K +++ mktemp ++ local LAST_ERR=/tmp/tmp.DLWt4RgZGl ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.y9wzEDK87K ++ cat /tmp/tmp.DLWt4RgZGl ++ rm /tmp/tmp.y9wzEDK87K /tmp/tmp.DLWt4RgZGl ++ return 0 + [[ 9Gi == \1\0\G\i ]] + [[ 2 -ge 60 ]] + echo -n . .+ sleep 10 + retry=3 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hi0DCKpByq +++ mktemp ++ local LAST_ERR=/tmp/tmp.HMnqkdH6AR ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.hi0DCKpByq ++ cat /tmp/tmp.HMnqkdH6AR ++ rm /tmp/tmp.hi0DCKpByq /tmp/tmp.HMnqkdH6AR ++ return 0 + [[ 9Gi == \1\0\G\i ]] + [[ 3 -ge 60 ]] + echo -n . .+ sleep 10 + retry=4 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JwyGTWarPf +++ mktemp ++ local LAST_ERR=/tmp/tmp.0qU6nPlEBn ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.JwyGTWarPf ++ cat /tmp/tmp.0qU6nPlEBn ++ rm /tmp/tmp.JwyGTWarPf /tmp/tmp.0qU6nPlEBn ++ return 0 + [[ 9Gi == \1\0\G\i ]] + [[ 4 -ge 60 ]] + echo -n . .+ sleep 10 + retry=5 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AlkCzYiVos +++ mktemp ++ local LAST_ERR=/tmp/tmp.btKF9MeNFo ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.AlkCzYiVos ++ cat /tmp/tmp.btKF9MeNFo ++ rm /tmp/tmp.AlkCzYiVos /tmp/tmp.btKF9MeNFo ++ return 0 + [[ 10Gi == \1\0\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-1 successfully resized to 10Gi' PVC mongod-data-some-name-rs0-1 successfully resized to 10Gi + wait_for_auto_resize mongod-data-some-name-rs0-2 10Gi 60 10 + local pvc=mongod-data-some-name-rs0-2 + local expected_size=10Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-2 to auto-resize to 10Gi' Waiting for PVC mongod-data-some-name-rs0-2 to auto-resize to 10Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-2 ++ local pvc=mongod-data-some-name-rs0-2 ++ kubectl_bin get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3b4RJonCc5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.okW5oG5NF5 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.3b4RJonCc5 ++ cat /tmp/tmp.okW5oG5NF5 ++ rm /tmp/tmp.3b4RJonCc5 /tmp/tmp.okW5oG5NF5 ++ return 0 + [[ 10Gi == \1\0\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-2 successfully resized to 10Gi' PVC mongod-data-some-name-rs0-2 successfully resized to 10Gi + desc 'verify final autoscaling status' + set +o xtrace ----------------------------------------------------------------------------------- verify final autoscaling status ----------------------------------------------------------------------------------- + check_autoscaling_status some-name 4 mongod-data-some-name-rs0-0 mongod-data-some-name-rs0-1 mongod-data-some-name-rs0-2 + local cluster=some-name + local expected_resize_count=4 + shift 2 + pvc_names=('mongod-data-some-name-rs0-0' 'mongod-data-some-name-rs0-1' 'mongod-data-some-name-rs0-2') + local pvc_names + local max_retry=30 + local sleep_time=10 + echo 'Checking autoscaling status for PVCs: mongod-data-some-name-rs0-0 mongod-data-some-name-rs0-1 mongod-data-some-name-rs0-2' Checking autoscaling status for PVCs: mongod-data-some-name-rs0-0 mongod-data-some-name-rs0-1 mongod-data-some-name-rs0-2 + echo 'Expected resize count per PVC: 4' Expected resize count per PVC: 4 + for pvc_name in "${pvc_names[@]}" + local retry=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-0.resizeCount}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AXH4CnCqPe +++ mktemp ++ local LAST_ERR=/tmp/tmp.Uppjg0QfqD ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-0.resizeCount}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.AXH4CnCqPe ++ cat /tmp/tmp.Uppjg0QfqD ++ rm /tmp/tmp.AXH4CnCqPe /tmp/tmp.Uppjg0QfqD ++ return 0 + [[ 4 -eq 4 ]] + for pvc_name in "${pvc_names[@]}" + local retry=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-1.resizeCount}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3GD7iKsYfm +++ mktemp ++ local LAST_ERR=/tmp/tmp.bH8KUEJouh ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-1.resizeCount}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.3GD7iKsYfm ++ cat /tmp/tmp.bH8KUEJouh ++ rm /tmp/tmp.3GD7iKsYfm /tmp/tmp.bH8KUEJouh ++ return 0 + [[ 4 -eq 4 ]] + for pvc_name in "${pvc_names[@]}" + local retry=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-2.resizeCount}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SaEp2l8JbT +++ mktemp ++ local LAST_ERR=/tmp/tmp.dPQz0VEfG2 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-2.resizeCount}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.SaEp2l8JbT ++ cat /tmp/tmp.dPQz0VEfG2 ++ rm /tmp/tmp.SaEp2l8JbT /tmp/tmp.dPQz0VEfG2 ++ return 0 + [[ 4 -eq 4 ]] + wait_cluster_consistency some-name 64 + local cluster_name=some-name + local wait_time=64 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BYSYo53Khg +++ mktemp ++ local LAST_ERR=/tmp/tmp.SllB1dcXD8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.BYSYo53Khg ++ cat /tmp/tmp.SllB1dcXD8 ++ rm /tmp/tmp.BYSYo53Khg /tmp/tmp.SllB1dcXD8 ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo .OK .OK + destroy pvc-auto-resize-27302 + local namespace=pvc-auto-resize-27302 + local ignore_logs=true + [[ 0 == 1 ]] + desc 'destroy cluster/operator and all other resources' + set +o xtrace ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + '[' true == false ']' + delete_backups + desc 'Delete psmdb-backup' + set +o xtrace ----------------------------------------------------------------------------------- Delete psmdb-backup ----------------------------------------------------------------------------------- ++ wc -l ++ kubectl_bin get psmdb-backup --no-headers +++ mktemp ++ local LAST_OUT=/tmp/tmp.ohA28zbJ0l +++ mktemp ++ local LAST_ERR=/tmp/tmp.fevOP1KaUF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb-backup --no-headers ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ohA28zbJ0l ++ cat /tmp/tmp.fevOP1KaUF No resources found in pvc-auto-resize-27302 namespace. ++ rm /tmp/tmp.ohA28zbJ0l /tmp/tmp.fevOP1KaUF ++ return 0 + '[' 0 '!=' 0 ']' + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.tWqfjUuYO0 ++ mktemp + local LAST_ERR=/tmp/tmp.TlJUaVSxPs + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.tWqfjUuYO0 customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.TlJUaVSxPs + rm /tmp/tmp.tWqfjUuYO0 /tmp/tmp.TlJUaVSxPs + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/deploy/crd.yaml ++ grep -v '\-\-\-' grep: warning: stray \ before - grep: warning: stray \ before - + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.s6rRZklPZB ++ mktemp + local LAST_ERR=/tmp/tmp.kw4tJHflQz + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.s6rRZklPZB + cat /tmp/tmp.kw4tJHflQz + rm /tmp/tmp.s6rRZklPZB /tmp/tmp.kw4tJHflQz + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.0oVc2qYFCh ++ mktemp + local LAST_ERR=/tmp/tmp.Uvx20pI8vm + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.0oVc2qYFCh + cat /tmp/tmp.Uvx20pI8vm + rm /tmp/tmp.0oVc2qYFCh /tmp/tmp.Uvx20pI8vm + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch perconaservermongodbs.psmdb.percona.com -n pvc-auto-resize-27302 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaservermongodb.psmdb.percona.com/some-name patched + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.TBvheiBu8O ++ mktemp + local LAST_ERR=/tmp/tmp.PQs3xmnlqP + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.TBvheiBu8O customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com condition met + cat /tmp/tmp.PQs3xmnlqP + rm /tmp/tmp.TBvheiBu8O /tmp/tmp.PQs3xmnlqP + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.E2brmMhuuC ++ mktemp + local LAST_ERR=/tmp/tmp.3txFuhfqCH + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2247/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.E2brmMhuuC clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.3txFuhfqCH + rm /tmp/tmp.E2brmMhuuC /tmp/tmp.3txFuhfqCH + return 0 + destroy_cert_manager + kubectl_bin delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.GumvDEuaMV ++ mktemp + local LAST_ERR=/tmp/tmp.NjYx7jqXGf + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.GumvDEuaMV + cat /tmp/tmp.NjYx7jqXGf Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 0 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.GumvDEuaMV + cat /tmp/tmp.NjYx7jqXGf Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 4 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.GumvDEuaMV + cat /tmp/tmp.NjYx7jqXGf Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 8 + cat /tmp/tmp.GumvDEuaMV + cat /tmp/tmp.NjYx7jqXGf Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + rm /tmp/tmp.GumvDEuaMV /tmp/tmp.NjYx7jqXGf + return 1 + true + '[' -n '' ']' + '[' -n psmdb-operator ']' + kubectl_bin delete --grace-period=0 --force=true namespace pvc-auto-resize-27302 + rm -rf /tmp/tmp.xS08x4x5XN + kubectl_bin delete --grace-period=0 --force=true namespace psmdb-operator ++ mktemp ++ mktemp + desc 'test passed' + set +o xtrace ----------------------------------------------------------------------------------- test passed+ local LAST_OUT=/tmp/tmp.pRQru3hWyv ----------------------------------------------------------------------------------- + local LAST_OUT=/tmp/tmp.W7Wrz0vZMf ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.2HosALVFzA + local exit_status=0 + local timeout=4 + local LAST_ERR=/tmp/tmp.W0nAInEVnr + local exit_status=0 + local timeout=4 ++ seq 0 2 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete --grace-period=0 --force=true namespace psmdb-operator + for i in $(seq 0 2) + set +e + kubectl delete --grace-period=0 --force=true namespace pvc-auto-resize-27302