Log: /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/logs/pvc-auto-resize.log Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 + '[' 0 == 1 ']' + ensure_default_sc_allows_expansion + local default_sc ++ get_default_storageclass ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FdVdIwa9SJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.m1p7FzNwZG ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.FdVdIwa9SJ ++ cat /tmp/tmp.m1p7FzNwZG ++ rm /tmp/tmp.FdVdIwa9SJ /tmp/tmp.m1p7FzNwZG ++ return 0 + default_sc=standard-rwo + echo 'Checking if default storageclass standard-rwo allows volume expansion' Checking if default storageclass standard-rwo allows volume expansion + local allowVolumeExpansion ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.name=="standard-rwo")].allowVolumeExpansion}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FmqVeAC4uW +++ mktemp ++ local LAST_ERR=/tmp/tmp.ABeKNFyHWE ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.name=="standard-rwo")].allowVolumeExpansion}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.FmqVeAC4uW ++ cat /tmp/tmp.ABeKNFyHWE ++ rm /tmp/tmp.FmqVeAC4uW /tmp/tmp.ABeKNFyHWE ++ return 0 + allowVolumeExpansion=true + [[ true != true ]] + create_infra pvc-auto-resize-16581 + local ns=pvc-auto-resize-16581 + [[ 1 == 1 ]] + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.AU7dXM0MaX ++ mktemp + local LAST_ERR=/tmp/tmp.FuEKfGkctu + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.AU7dXM0MaX customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.FuEKfGkctu + rm /tmp/tmp.AU7dXM0MaX /tmp/tmp.FuEKfGkctu + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/crd.yaml ++ grep -v '\-\-\-' grep: warning: stray \ before - grep: warning: stray \ before - + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + grep -v NAMESPACE + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.aZTj9ew0mn ++ mktemp + local LAST_ERR=/tmp/tmp.8TDWVIpj5N + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.aZTj9ew0mn + cat /tmp/tmp.8TDWVIpj5N + rm /tmp/tmp.aZTj9ew0mn /tmp/tmp.8TDWVIpj5N + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.XNK9RBTK8I ++ mktemp + local LAST_ERR=/tmp/tmp.eK2yLmHDfR + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.XNK9RBTK8I + cat /tmp/tmp.eK2yLmHDfR + rm /tmp/tmp.XNK9RBTK8I /tmp/tmp.eK2yLmHDfR + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch perconaservermongodbs.psmdb.percona.com -n pvc-auto-resize-11351 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaservermongodb.psmdb.percona.com/some-name patched + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.TA45MeLtrx ++ mktemp + local LAST_ERR=/tmp/tmp.1oQZfhlW3m + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.TA45MeLtrx customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com condition met + cat /tmp/tmp.1oQZfhlW3m + rm /tmp/tmp.TA45MeLtrx /tmp/tmp.1oQZfhlW3m + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.Ie00E8pwzv ++ mktemp + local LAST_ERR=/tmp/tmp.wI3V77dp1S + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Ie00E8pwzv clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.wI3V77dp1S + rm /tmp/tmp.Ie00E8pwzv /tmp/tmp.wI3V77dp1S + return 0 + check_crd_for_deletion PR-2219-fe9d8c05 + local git_tag=PR-2219-fe9d8c05 ++ curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/PR-2219-fe9d8c05/deploy/crd.yaml ++ yq eval .metadata.name ++ /usr/sbin/sed s/---//g ++ /usr/sbin/sed ':a;N;$!ba;s/\n/ /g' + for crd_name in $(curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/${git_tag}/deploy/crd.yaml | yq eval '.metadata.name' | $sed 's/---//g' | $sed ':a;N;$!ba;s/\n/ /g') ++ kubectl_bin get crd/null -o 'jsonpath={.status.conditions[-1].type}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.baGSUdCLJK +++ mktemp ++ local LAST_ERR=/tmp/tmp.o7yrBOiE49 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.baGSUdCLJK ++ cat /tmp/tmp.o7yrBOiE49 Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 0 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.baGSUdCLJK ++ cat /tmp/tmp.o7yrBOiE49 Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 4 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.baGSUdCLJK ++ cat /tmp/tmp.o7yrBOiE49 Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 8 ++ cat /tmp/tmp.baGSUdCLJK ++ cat /tmp/tmp.o7yrBOiE49 Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ rm /tmp/tmp.baGSUdCLJK /tmp/tmp.o7yrBOiE49 ++ return 1 + [[ '' == Terminating ]] + '[' -n psmdb-operator ']' + create_namespace psmdb-operator + local namespace=psmdb-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + grep -E -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' + awk '{print$1}' + '[' -n '' ']' + desc 'cleaned up old namespaces psmdb-operator' + set +o xtrace ++ mktemp ----------------------------------------------------------------------------------- cleaned up old namespaces psmdb-operator ----------------------------------------------------------------------------------- + xargs kubectl delete ns + kubectl_bin delete namespace psmdb-operator --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.LSvUrcHO6b + local LAST_OUT=/tmp/tmp.V5bCBFEmOq ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.M1x4XvqPjj + local exit_status=0 + local timeout=4 ++ seq 0 2 + local LAST_ERR=/tmp/tmp.w01zORbc0u + local exit_status=0 + local timeout=4 + for i in $(seq 0 2) + set +e + kubectl get ns ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete namespace psmdb-operator --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.V5bCBFEmOq + cat /tmp/tmp.M1x4XvqPjj + rm /tmp/tmp.V5bCBFEmOq /tmp/tmp.M1x4XvqPjj + return 0 namespace "pvc-auto-resize-11351" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.LSvUrcHO6b namespace "psmdb-operator" deleted + cat /tmp/tmp.w01zORbc0u + rm /tmp/tmp.LSvUrcHO6b /tmp/tmp.w01zORbc0u + return 0 + kubectl_bin wait --for=delete namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.jt2dREMdkV ++ mktemp + local LAST_ERR=/tmp/tmp.83kjcNyDD8 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.jt2dREMdkV + cat /tmp/tmp.83kjcNyDD8 + rm /tmp/tmp.jt2dREMdkV /tmp/tmp.83kjcNyDD8 + return 0 + desc 'create namespace psmdb-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.8w7zUU55QS ++ mktemp + local LAST_ERR=/tmp/tmp.znPLvFqw8d + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.8w7zUU55QS namespace/psmdb-operator created + cat /tmp/tmp.znPLvFqw8d + rm /tmp/tmp.8w7zUU55QS /tmp/tmp.znPLvFqw8d + return 0 + set_kube_ctx psmdb-operator + local namespace=psmdb-operator ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.lPV3hpViIf +++ mktemp ++ local LAST_ERR=/tmp/tmp.FsE6flZKwm ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.lPV3hpViIf ++ cat /tmp/tmp.FsE6flZKwm ++ rm /tmp/tmp.lPV3hpViIf /tmp/tmp.FsE6flZKwm ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2219-fe9d8c05-8-cluster15 --namespace=psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.Qd9eREQ6t4 ++ mktemp + local LAST_ERR=/tmp/tmp.Bnaoi8RCPk + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2219-fe9d8c05-8-cluster15 --namespace=psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Qd9eREQ6t4 Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2219-fe9d8c05-8-cluster15" modified. + cat /tmp/tmp.Bnaoi8RCPk + rm /tmp/tmp.Qd9eREQ6t4 /tmp/tmp.Bnaoi8RCPk + return 0 + deploy_operator + desc 'start PSMDB operator: docker.io/perconalab/percona-server-mongodb-operator:PR-2219-fe9d8c05' + set +o xtrace ----------------------------------------------------------------------------------- start PSMDB operator: docker.io/perconalab/percona-server-mongodb-operator:PR-2219-fe9d8c05 ----------------------------------------------------------------------------------- + local cr_file + '[' -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/pvc-auto-resize/conf/crd.yaml ']' + cr_file=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/crd.yaml + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.KzidLNGqTP ++ mktemp + local LAST_ERR=/tmp/tmp.9gTd6kDNbA + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.KzidLNGqTP customresourcedefinition.apiextensions.k8s.io/perconaservermongodbbackups.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbrestores.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com serverside-applied + cat /tmp/tmp.9gTd6kDNbA + rm /tmp/tmp.KzidLNGqTP /tmp/tmp.9gTd6kDNbA + return 0 + '[' -n psmdb-operator ']' + apply_rbac cw-rbac + local operator_namespace=psmdb-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: psmdb-operator^' + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.RcW67oFT0J ++ mktemp + local LAST_ERR=/tmp/tmp.ycqrGzTzEb + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.RcW67oFT0J clusterrole.rbac.authorization.k8s.io/percona-server-mongodb-operator created serviceaccount/percona-server-mongodb-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-server-mongodb-operator created + cat /tmp/tmp.ycqrGzTzEb + rm /tmp/tmp.RcW67oFT0J /tmp/tmp.ycqrGzTzEb + return 0 + yq eval $'\n\t\t\t(.spec.template.spec.containers[].image = "docker.io/perconalab/percona-server-mongodb-operator:PR-2219-fe9d8c05") |\n\t\t\t((.. | select(.[] == "DISABLE_TELEMETRY")) |= .value="true") |\n\t\t\t((.. | select(.[] == "LOG_LEVEL")) |= .value="DEBUG")' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/cw-operator.yaml + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.eihMUNLnaH ++ mktemp + local LAST_ERR=/tmp/tmp.1UgLAVh3a8 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.eihMUNLnaH deployment.apps/percona-server-mongodb-operator created + cat /tmp/tmp.1UgLAVh3a8 + rm /tmp/tmp.eihMUNLnaH /tmp/tmp.1UgLAVh3a8 + return 0 + sleep 20 ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.jZhTwdTleE +++ mktemp ++ local LAST_ERR=/tmp/tmp.FVb1E2XZmh ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jZhTwdTleE ++ cat /tmp/tmp.FVb1E2XZmh ++ rm /tmp/tmp.jZhTwdTleE /tmp/tmp.FVb1E2XZmh ++ return 0 + wait_operator_pod percona-server-mongodb-operator-5796f4f5b5-qcqwz + local pod=percona-server-mongodb-operator-5796f4f5b5-qcqwz + set +o xtrace waiting for pod/percona-server-mongodb-operator-5796f4f5b5-qcqwz to be ready.OK + echo 'Print operator info from log' Print operator info from log + grep 'Manager starting up' ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.fNmhCk017L +++ mktemp ++ local LAST_ERR=/tmp/tmp.wQEHJpD1QL ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.fNmhCk017L ++ cat /tmp/tmp.wQEHJpD1QL ++ rm /tmp/tmp.fNmhCk017L /tmp/tmp.wQEHJpD1QL ++ return 0 + kubectl_bin logs -n psmdb-operator percona-server-mongodb-operator-5796f4f5b5-qcqwz ++ mktemp + local LAST_OUT=/tmp/tmp.UGravL6HOK ++ mktemp + local LAST_ERR=/tmp/tmp.encgFRVVya + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl logs -n psmdb-operator percona-server-mongodb-operator-5796f4f5b5-qcqwz + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.UGravL6HOK + cat /tmp/tmp.encgFRVVya + rm /tmp/tmp.UGravL6HOK /tmp/tmp.encgFRVVya + return 0 2026-04-10T11:21:53.626Z INFO setup Manager starting up {"gitCommit": "fe9d8c05ac7c569882d27d07bd9604e14009ab07", "gitBranch": "PR-2219-fe9d8c05", "buildTime": "", "goVersion": "go1.25.9", "os": "linux", "arch": "amd64"} + create_namespace pvc-auto-resize-16581 + local namespace=pvc-auto-resize-16581 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ awk '-F ' '{print $2}' ++ tail -n1 ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep validate-auth + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ awk '{print $1}' ++ grep chaos-mesh.org + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ kubectl get clusterrolebinding ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + '[' -n '' ']' + desc 'cleaned up old namespaces pvc-auto-resize-16581' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pvc-auto-resize-16581 ----------------------------------------------------------------------------------- + xargs kubectl delete ns + kubectl_bin delete namespace pvc-auto-resize-16581 --ignore-not-found ++ mktemp + awk '{print$1}' + kubectl_bin get ns + local LAST_OUT=/tmp/tmp.BNt4aztQoC + grep -E -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.n9hnk8aoeN ++ mktemp + local LAST_ERR=/tmp/tmp.ZRPLzCpDQ6 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get ns + local LAST_ERR=/tmp/tmp.Ct1yRsJrfc + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete namespace pvc-auto-resize-16581 --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.n9hnk8aoeN + cat /tmp/tmp.ZRPLzCpDQ6 + rm /tmp/tmp.n9hnk8aoeN /tmp/tmp.ZRPLzCpDQ6 + return 0 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.BNt4aztQoC + cat /tmp/tmp.Ct1yRsJrfc + rm /tmp/tmp.BNt4aztQoC /tmp/tmp.Ct1yRsJrfc + return 0 + kubectl_bin wait --for=delete namespace pvc-auto-resize-16581 ++ mktemp + local LAST_OUT=/tmp/tmp.m8lxkQg6UT ++ mktemp + local LAST_ERR=/tmp/tmp.fy5EFQ20kq + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete namespace pvc-auto-resize-16581 error: resource(s) were provided, but no name was specified + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.m8lxkQg6UT + cat /tmp/tmp.fy5EFQ20kq + rm /tmp/tmp.m8lxkQg6UT /tmp/tmp.fy5EFQ20kq + return 0 + desc 'create namespace pvc-auto-resize-16581' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pvc-auto-resize-16581 ----------------------------------------------------------------------------------- + kubectl_bin create namespace pvc-auto-resize-16581 ++ mktemp + local LAST_OUT=/tmp/tmp.PsIrvY0lVW ++ mktemp + local LAST_ERR=/tmp/tmp.HM4X6now2H + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace pvc-auto-resize-16581 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.PsIrvY0lVW namespace/pvc-auto-resize-16581 created + cat /tmp/tmp.HM4X6now2H + rm /tmp/tmp.PsIrvY0lVW /tmp/tmp.HM4X6now2H + return 0 + set_kube_ctx pvc-auto-resize-16581 + local namespace=pvc-auto-resize-16581 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.8uivqw8YkG +++ mktemp ++ local LAST_ERR=/tmp/tmp.3DL0QoOxLn ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.8uivqw8YkG ++ cat /tmp/tmp.3DL0QoOxLn ++ rm /tmp/tmp.8uivqw8YkG /tmp/tmp.3DL0QoOxLn ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2219-fe9d8c05-8-cluster15 --namespace=pvc-auto-resize-16581 ++ mktemp + local LAST_OUT=/tmp/tmp.nJficqyUGk ++ mktemp + local LAST_ERR=/tmp/tmp.QB115CiWEy + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2219-fe9d8c05-8-cluster15 --namespace=pvc-auto-resize-16581 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.nJficqyUGk Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2219-fe9d8c05-8-cluster15" modified. + cat /tmp/tmp.QB115CiWEy + rm /tmp/tmp.nJficqyUGk /tmp/tmp.QB115CiWEy + return 0 + desc 'create secrets and psmdb client' + set +o xtrace ----------------------------------------------------------------------------------- create secrets and psmdb client ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/conf/secrets.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.rX8YulFwIL ++ mktemp + local LAST_ERR=/tmp/tmp.SaPj5xfmy4 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/conf/secrets.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/conf/client.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.rX8YulFwIL secret/some-users created deployment.apps/psmdb-client created + cat /tmp/tmp.SaPj5xfmy4 + rm /tmp/tmp.rX8YulFwIL /tmp/tmp.SaPj5xfmy4 + return 0 + desc 'create PSMDB cluster with storageAutoscaling enabled' + set +o xtrace ----------------------------------------------------------------------------------- create PSMDB cluster with storageAutoscaling enabled ----------------------------------------------------------------------------------- + cluster=some-name + '[' 0 == 1 ']' + spinup_psmdb some-name-rs0 /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/pvc-auto-resize/conf/some-name.yml + local cluster=some-name-rs0 + local config=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/pvc-auto-resize/conf/some-name.yml + local size=3 + desc 'create first PSMDB cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PSMDB cluster ----------------------------------------------------------------------------------- + apply_cluster /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/pvc-auto-resize/conf/some-name.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/pvc-auto-resize/conf/some-name.yml + kubectl_bin apply -f - + yq eval '(.spec | select(.image == null)).image = "docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0"' + yq eval '(.spec | select(has("pmm"))).pmm.image = "docker.io/percona/pmm-client:2.44.1-1"' + yq eval '(.spec | select(has("backup"))).backup.image = "docker.io/perconalab/percona-server-mongodb-operator:main-backup"' + /usr/sbin/sed -e s/NAME_SPACE/pvc-auto-resize-16581/g ++ mktemp + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/pvc-auto-resize/conf/some-name.yml + yq eval '(.spec | select(has("initImage"))).initImage = "docker.io/perconalab/percona-server-mongodb-operator:PR-2219-fe9d8c05"' + yq eval '.spec.upgradeOptions.apply="Never"' + local LAST_OUT=/tmp/tmp.9Y7pMKz9jq ++ mktemp + local LAST_ERR=/tmp/tmp.8DuD18fBn7 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.9Y7pMKz9jq perconaservermongodb.psmdb.percona.com/some-name created + cat /tmp/tmp.8DuD18fBn7 + rm /tmp/tmp.9Y7pMKz9jq /tmp/tmp.8DuD18fBn7 + return 0 + desc 'check if Pod is started' + set +o xtrace ----------------------------------------------------------------------------------- check if Pod is started ----------------------------------------------------------------------------------- + wait_for_running some-name-rs0 3 + local name=some-name-rs0 + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=rs0 + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-rs0-0 + local pod=some-name-rs0-0 + set +o xtrace waiting for pod/some-name-rs0-0 to be ready..............OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-rs0-1 + local pod=some-name-rs0-1 + set +o xtrace waiting for pod/some-name-rs0-1 to be ready.............OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tEQjctxZ0u +++ mktemp ++ local LAST_ERR=/tmp/tmp.NJLNsz1yJP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.tEQjctxZ0u ++ cat /tmp/tmp.NJLNsz1yJP ++ rm /tmp/tmp.tEQjctxZ0u /tmp/tmp.NJLNsz1yJP ++ return 0 + [[ '' == true ]] + wait_pod some-name-rs0-2 + local pod=some-name-rs0-2 + set +o xtrace waiting for pod/some-name-rs0-2 to be ready.............OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dYJAP0nX32 +++ mktemp ++ local LAST_ERR=/tmp/tmp.AxovawWuWY ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.dYJAP0nX32 ++ cat /tmp/tmp.AxovawWuWY ++ rm /tmp/tmp.dYJAP0nX32 /tmp/tmp.AxovawWuWY ++ return 0 + [[ '' == true ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.I2xkcHlvcj +++ mktemp ++ local LAST_ERR=/tmp/tmp.LN3MMktXqt ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.I2xkcHlvcj ++ cat /tmp/tmp.LN3MMktXqt ++ rm /tmp/tmp.I2xkcHlvcj /tmp/tmp.LN3MMktXqt ++ return 0 + [[ '' == true ]] + sleep 10 + [[ true == true ]] + set +x Waiting for cluster readyness + sleep 20 + compare_kubectl statefulset/some-name-rs0 + local resource=statefulset/some-name-rs0 + local postfix= + local skip_generation_check= + local expected_result=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/pvc-auto-resize/compare/statefulset_some-name-rs0.yml + local new_result=/tmp/tmp.IJmbDctCRT/statefulset_some-name-rs0.yml + '[' -n '' -a -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/pvc-auto-resize/compare/statefulset_some-name-rs0-oc.yml ']' + kubectl_bin get -o yaml statefulset/some-name-rs0 + yq eval $'\n\t\t\tdel(.metadata.ownerReferences[].apiVersion) |\n\t\t\tdel(.metadata.managedFields) |\n\t\t\tdel(.. | select(has("creationTimestamp")).creationTimestamp) |\n\t\t\tdel(.. | select(has("namespace")).namespace) |\n\t\t\tdel(.. | select(has("uid")).uid) |\n\t\t\tdel(.metadata.resourceVersion) |\n\t\t\tdel(.spec.template.spec.containers[].env[] | select(.name == "NAMESPACE")) |\n\t\t\tdel(.metadata.selfLink) |\n\t\t\tdel(.metadata.annotations."cloud.google.com/neg") |\n\t\t\tdel(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") |\n\t\t\tdel(.. | select(has("image")).image) |\n\t\t\tdel(.. | select(has("clusterIP")).clusterIP) |\n\t\t\tdel(.. | select(has("clusterIPs")).clusterIPs) |\n\t\t\tdel(.. | select(has("dataSource")).dataSource) |\n\t\t\tdel(.. | select(has("procMount")).procMount) |\n\t\t\tdel(.. | select(has("storageClassName")).storageClassName) |\n\t\t\tdel(.. | select(has("finalizers")).finalizers) |\n\t\t\tdel(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") |\n\t\t\tdel(.. | select(has("volumeName")).volumeName) |\n\t\t\tdel(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") |\n\t\t\tdel(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") |\n\t\t\tdel(.spec.volumeMode) |\n\t\t\tdel(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") |\n\t\t\tdel(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") |\n\t\t\tdel(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") |\n\t\t\tdel(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") |\n\t\t\tdel(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") |\n\t\t\tdel(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) |\n\t\t\tdel(.. | select(has("healthCheckNodePort")).healthCheckNodePort) |\n\t\t\tdel(.. | select(has("nodePort")).nodePort) |\n\t\t\tdel(.status) |\n\t\t\t(.. | select(tag == "!!str")) |= sub("pvc-auto-resize-16581", "NAME_SPACE") |\n\t\t\tdel(.spec.volumeClaimTemplates[].apiVersion) |\n\t\t\tdel(.spec.volumeClaimTemplates[].kind) |\n\t\t\tdel(.spec.ipFamilies) |\n\t\t\tdel(.spec.ipFamilyPolicy) |\n\t\t\t(.. | select(. == "extensions/v1beta1")) = "apps/v1" |\n\t\t\t(.. | select(. == "batch/v1beta1")) = "batch/v1" ' - ++ mktemp + local LAST_OUT=/tmp/tmp.ND0dDFYXSb ++ mktemp + local LAST_ERR=/tmp/tmp.LH1lKHmthE + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get -o yaml statefulset/some-name-rs0 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ND0dDFYXSb + cat /tmp/tmp.LH1lKHmthE + rm /tmp/tmp.ND0dDFYXSb /tmp/tmp.LH1lKHmthE + return 0 + yq -i eval 'del(.spec.persistentVolumeClaimRetentionPolicy)' /tmp/tmp.IJmbDctCRT/statefulset_some-name-rs0.yml + version_gt 1.22 ++ echo '1.32 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + yq -i eval 'del(.spec.internalTrafficPolicy)' /tmp/tmp.IJmbDctCRT/statefulset_some-name-rs0.yml + yq -i eval 'del(.spec.allocateLoadBalancerNodePorts)' /tmp/tmp.IJmbDctCRT/statefulset_some-name-rs0.yml + [[ /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/pvc-auto-resize/compare/statefulset_some-name-rs0.yml == */cronjob* ]] + '[' -n '' ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/pvc-auto-resize/compare/statefulset_some-name-rs0.yml /tmp/tmp.IJmbDctCRT/statefulset_some-name-rs0.yml + log 'compare_kubectl: statefulset/some-name-rs0 OK' + set +o xtrace [2026-04-10T11:24:32+0000] compare_kubectl: statefulset/some-name-rs0 OK + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + run_mongo 'db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})' userAdmin:userAdmin123456@some-name-rs0.pvc-auto-resize-16581 + local 'command=db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})' + local uri=userAdmin:userAdmin123456@some-name-rs0.pvc-auto-resize-16581 + local driver=mongodb+srv + local suffix=.svc.cluster.local + local mongo_flag= + local replica_set=rs0 + [[ userAdmin:userAdmin123456@some-name-rs0.pvc-auto-resize-16581 == *cfg* ]] ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nZIH8RCKzf +++ mktemp ++ local LAST_ERR=/tmp/tmp.MO0ETtJr35 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.nZIH8RCKzf ++ cat /tmp/tmp.MO0ETtJr35 ++ rm /tmp/tmp.nZIH8RCKzf /tmp/tmp.MO0ETtJr35 ++ return 0 + local client_container=psmdb-client-bb8b97679-w9xjh + kubectl_bin exec psmdb-client-bb8b97679-w9xjh -- bash -c 'printf '\''db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})\n'\'' | mongo mongodb+srv://userAdmin:userAdmin123456@some-name-rs0.pvc-auto-resize-16581.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.ihCyWPZBbf ++ mktemp + local LAST_ERR=/tmp/tmp.p3cwgHUxqX + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-w9xjh -- bash -c 'printf '\''db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})\n'\'' | mongo mongodb+srv://userAdmin:userAdmin123456@some-name-rs0.pvc-auto-resize-16581.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ihCyWPZBbf Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-rs0-0.some-name-rs0.pvc-auto-resize-16581.svc.cluster.local:27017,some-name-rs0-2.some-name-rs0.pvc-auto-resize-16581.svc.cluster.local:27017,some-name-rs0-1.some-name-rs0.pvc-auto-resize-16581.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb&replicaSet=rs0&ssl=false Implicit session: session { "id" : UUID("c8a2d7ee-d885-4dbd-bd07-f804a74118f7") } Percona Server for MongoDB server version: v8.0.20-8 WARNING: shell and server versions do not match Successfully added user: { "user" : "myApp", "roles" : [ { "db" : "myApp", "role" : "readWrite" } ] } bye + cat /tmp/tmp.p3cwgHUxqX + rm /tmp/tmp.ihCyWPZBbf /tmp/tmp.p3cwgHUxqX + return 0 + run_mongo 'use myApp\n db.test.insert({ x: 100500 })' myApp:myPass@some-name-rs0.pvc-auto-resize-16581 + local 'command=use myApp\n db.test.insert({ x: 100500 })' + local uri=myApp:myPass@some-name-rs0.pvc-auto-resize-16581 + local driver=mongodb+srv + local suffix=.svc.cluster.local + local mongo_flag= + local replica_set=rs0 + [[ myApp:myPass@some-name-rs0.pvc-auto-resize-16581 == *cfg* ]] ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xeJ77y7UZP +++ mktemp ++ local LAST_ERR=/tmp/tmp.CJiqFnI4ZF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xeJ77y7UZP ++ cat /tmp/tmp.CJiqFnI4ZF ++ rm /tmp/tmp.xeJ77y7UZP /tmp/tmp.CJiqFnI4ZF ++ return 0 + local client_container=psmdb-client-bb8b97679-w9xjh + kubectl_bin exec psmdb-client-bb8b97679-w9xjh -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb+srv://myApp:myPass@some-name-rs0.pvc-auto-resize-16581.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.IJrpCf7n4S ++ mktemp + local LAST_ERR=/tmp/tmp.bPui8MLH5D + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-w9xjh -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb+srv://myApp:myPass@some-name-rs0.pvc-auto-resize-16581.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.IJrpCf7n4S Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-rs0-1.some-name-rs0.pvc-auto-resize-16581.svc.cluster.local:27017,some-name-rs0-0.some-name-rs0.pvc-auto-resize-16581.svc.cluster.local:27017,some-name-rs0-2.some-name-rs0.pvc-auto-resize-16581.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb&replicaSet=rs0&ssl=false Implicit session: session { "id" : UUID("08ca0a38-f490-428d-85fa-8e3247b038dd") } Percona Server for MongoDB server version: v8.0.20-8 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.bPui8MLH5D + rm /tmp/tmp.IJrpCf7n4S /tmp/tmp.bPui8MLH5D + return 0 + desc 'verify initial PVC size is 1Gi' + set +o xtrace ----------------------------------------------------------------------------------- verify initial PVC size is 1Gi ----------------------------------------------------------------------------------- ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oB726c8T8N +++ mktemp ++ local LAST_ERR=/tmp/tmp.dsIDrhD2R0 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.oB726c8T8N ++ cat /tmp/tmp.dsIDrhD2R0 ++ rm /tmp/tmp.oB726c8T8N /tmp/tmp.dsIDrhD2R0 ++ return 0 + initial_size=1Gi + echo 'Initial PVC size: 1Gi' Initial PVC size: 1Gi + [[ 1Gi != 1Gi ]] + desc 'fill disk to trigger autoscaling threshold 50 percent' + set +o xtrace ----------------------------------------------------------------------------------- fill disk to trigger autoscaling threshold 50 percent ----------------------------------------------------------------------------------- + fill_disk_to_threshold some-name-rs0-0 50 600 + local pod=some-name-rs0-0 + local threshold=50 + local size_mb=600 + echo 'Filling disk on some-name-rs0-0 to approximately 50% (creating 600MB file)' Filling disk on some-name-rs0-0 to approximately 50% (creating 600MB file) + kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=600 2>/dev/null || true' ++ mktemp + local LAST_OUT=/tmp/tmp.hBHrAnRmAt ++ mktemp + local LAST_ERR=/tmp/tmp.PwhunLlonE + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=600 2>/dev/null || true' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.hBHrAnRmAt + cat /tmp/tmp.PwhunLlonE + rm /tmp/tmp.hBHrAnRmAt /tmp/tmp.PwhunLlonE + return 0 + local usage ++ kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eFI70Sapil +++ mktemp ++ local LAST_ERR=/tmp/tmp.9IsouxSv2w ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.eFI70Sapil ++ cat /tmp/tmp.9IsouxSv2w ++ rm /tmp/tmp.eFI70Sapil /tmp/tmp.9IsouxSv2w ++ return 0 + usage=85 + echo 'Current disk usage on some-name-rs0-0: 85%' Current disk usage on some-name-rs0-0: 85% + desc 'wait for auto-resize to trigger' + set +o xtrace ----------------------------------------------------------------------------------- wait for auto-resize to trigger ----------------------------------------------------------------------------------- + wait_for_auto_resize mongod-data-some-name-rs0-0 3Gi 60 10 + local pvc=mongod-data-some-name-rs0-0 + local expected_size=3Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 3Gi' Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 3Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qIuXbPwbwe +++ mktemp ++ local LAST_ERR=/tmp/tmp.ifox7KoBsb ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.qIuXbPwbwe ++ cat /tmp/tmp.ifox7KoBsb ++ rm /tmp/tmp.qIuXbPwbwe /tmp/tmp.ifox7KoBsb ++ return 0 + [[ 1Gi == 3Gi ]] + [[ 0 -ge 60 ]] + echo -n . .+ sleep 10 + retry=1 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kurk3Nnddi +++ mktemp ++ local LAST_ERR=/tmp/tmp.xzyV7IqkPT ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.kurk3Nnddi ++ cat /tmp/tmp.xzyV7IqkPT ++ rm /tmp/tmp.kurk3Nnddi /tmp/tmp.xzyV7IqkPT ++ return 0 + [[ 1Gi == 3Gi ]] + [[ 1 -ge 60 ]] + echo -n . .+ sleep 10 + retry=2 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TuWIvIvoU8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.YykhcmaXKM ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.TuWIvIvoU8 ++ cat /tmp/tmp.YykhcmaXKM ++ rm /tmp/tmp.TuWIvIvoU8 /tmp/tmp.YykhcmaXKM ++ return 0 + [[ 1Gi == 3Gi ]] + [[ 2 -ge 60 ]] + echo -n . .+ sleep 10 + retry=3 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AUs3SvSokD +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ec31y0Bldg ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.AUs3SvSokD ++ cat /tmp/tmp.Ec31y0Bldg ++ rm /tmp/tmp.AUs3SvSokD /tmp/tmp.Ec31y0Bldg ++ return 0 + [[ 1Gi == 3Gi ]] + [[ 3 -ge 60 ]] + echo -n . .+ sleep 10 + retry=4 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.g0UnrMgwmS +++ mktemp ++ local LAST_ERR=/tmp/tmp.NhekML2IYb ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.g0UnrMgwmS ++ cat /tmp/tmp.NhekML2IYb ++ rm /tmp/tmp.g0UnrMgwmS /tmp/tmp.NhekML2IYb ++ return 0 + [[ 1Gi == 3Gi ]] + [[ 4 -ge 60 ]] + echo -n . .+ sleep 10 + retry=5 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RxCTSwjVuf +++ mktemp ++ local LAST_ERR=/tmp/tmp.EgZBWxyNcs ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.RxCTSwjVuf ++ cat /tmp/tmp.EgZBWxyNcs ++ rm /tmp/tmp.RxCTSwjVuf /tmp/tmp.EgZBWxyNcs ++ return 0 + [[ 3Gi == 3Gi ]] + echo + echo 'PVC mongod-data-some-name-rs0-0 successfully resized to 3Gi' PVC mongod-data-some-name-rs0-0 successfully resized to 3Gi + wait_for_auto_resize mongod-data-some-name-rs0-1 3Gi 60 10 + local pvc=mongod-data-some-name-rs0-1 + local expected_size=3Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-1 to auto-resize to 3Gi' Waiting for PVC mongod-data-some-name-rs0-1 to auto-resize to 3Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.B5hZlBRwxS +++ mktemp ++ local LAST_ERR=/tmp/tmp.Rnonxkgpaf ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.B5hZlBRwxS ++ cat /tmp/tmp.Rnonxkgpaf ++ rm /tmp/tmp.B5hZlBRwxS /tmp/tmp.Rnonxkgpaf ++ return 0 + [[ 1Gi == 3Gi ]] + [[ 0 -ge 60 ]] + echo -n . .+ sleep 10 + retry=1 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.aZHsdBR6mO +++ mktemp ++ local LAST_ERR=/tmp/tmp.2Y6rUJhijn ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.aZHsdBR6mO ++ cat /tmp/tmp.2Y6rUJhijn ++ rm /tmp/tmp.aZHsdBR6mO /tmp/tmp.2Y6rUJhijn ++ return 0 + [[ 1Gi == 3Gi ]] + [[ 1 -ge 60 ]] + echo -n . .+ sleep 10 + retry=2 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hCDX0OkMJ2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.SPCdI1LMVA ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.hCDX0OkMJ2 ++ cat /tmp/tmp.SPCdI1LMVA ++ rm /tmp/tmp.hCDX0OkMJ2 /tmp/tmp.SPCdI1LMVA ++ return 0 + [[ 3Gi == 3Gi ]] + echo + echo 'PVC mongod-data-some-name-rs0-1 successfully resized to 3Gi' PVC mongod-data-some-name-rs0-1 successfully resized to 3Gi + wait_for_auto_resize mongod-data-some-name-rs0-2 3Gi 60 10 + local pvc=mongod-data-some-name-rs0-2 + local expected_size=3Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-2 to auto-resize to 3Gi' Waiting for PVC mongod-data-some-name-rs0-2 to auto-resize to 3Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-2 ++ local pvc=mongod-data-some-name-rs0-2 ++ kubectl_bin get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sedpIJKqEP +++ mktemp ++ local LAST_ERR=/tmp/tmp.FsFgJ36tYp ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.sedpIJKqEP ++ cat /tmp/tmp.FsFgJ36tYp ++ rm /tmp/tmp.sedpIJKqEP /tmp/tmp.FsFgJ36tYp ++ return 0 + [[ 3Gi == 3Gi ]] + echo + echo 'PVC mongod-data-some-name-rs0-2 successfully resized to 3Gi' PVC mongod-data-some-name-rs0-2 successfully resized to 3Gi + desc 'verify autoscaling status is updated' + set +o xtrace ----------------------------------------------------------------------------------- verify autoscaling status is updated ----------------------------------------------------------------------------------- + check_autoscaling_status some-name 1 mongod-data-some-name-rs0-0 mongod-data-some-name-rs0-1 mongod-data-some-name-rs0-2 + local cluster=some-name + local expected_resize_count=1 + shift 2 + pvc_names=('mongod-data-some-name-rs0-0' 'mongod-data-some-name-rs0-1' 'mongod-data-some-name-rs0-2') + local pvc_names + local max_retry=30 + local sleep_time=10 + echo 'Checking autoscaling status for PVCs: mongod-data-some-name-rs0-0 mongod-data-some-name-rs0-1 mongod-data-some-name-rs0-2' Checking autoscaling status for PVCs: mongod-data-some-name-rs0-0 mongod-data-some-name-rs0-1 mongod-data-some-name-rs0-2 + echo 'Expected resize count per PVC: 1' Expected resize count per PVC: 1 + for pvc_name in "${pvc_names[@]}" + local retry=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-0.resizeCount}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xvmHpZVfnY +++ mktemp ++ local LAST_ERR=/tmp/tmp.bGrF5XN6ne ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-0.resizeCount}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xvmHpZVfnY ++ cat /tmp/tmp.bGrF5XN6ne ++ rm /tmp/tmp.xvmHpZVfnY /tmp/tmp.bGrF5XN6ne ++ return 0 + [[ 1 -eq 1 ]] + for pvc_name in "${pvc_names[@]}" + local retry=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-1.resizeCount}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RwgyuEthyZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.yFKR28ESq2 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-1.resizeCount}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.RwgyuEthyZ ++ cat /tmp/tmp.yFKR28ESq2 ++ rm /tmp/tmp.RwgyuEthyZ /tmp/tmp.yFKR28ESq2 ++ return 0 + [[ 1 -eq 1 ]] + for pvc_name in "${pvc_names[@]}" + local retry=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-2.resizeCount}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.CYYLPSH9ak +++ mktemp ++ local LAST_ERR=/tmp/tmp.W25ByL3lZF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-2.resizeCount}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.CYYLPSH9ak ++ cat /tmp/tmp.W25ByL3lZF ++ rm /tmp/tmp.CYYLPSH9ak /tmp/tmp.W25ByL3lZF ++ return 0 + [[ 1 -eq 1 ]] + desc 'wait for cluster consistency after resize' + set +o xtrace ----------------------------------------------------------------------------------- wait for cluster consistency after resize ----------------------------------------------------------------------------------- + wait_cluster_consistency some-name 64 + local cluster_name=some-name + local wait_time=64 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kOcfNtNixC +++ mktemp ++ local LAST_ERR=/tmp/tmp.2a0Fa0eLcF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.kOcfNtNixC ++ cat /tmp/tmp.2a0Fa0eLcF ++ rm /tmp/tmp.kOcfNtNixC /tmp/tmp.2a0Fa0eLcF ++ return 0 + [[ ready == ready ]] + echo .OK .OK + desc 'verify cluster is ready' + set +o xtrace ----------------------------------------------------------------------------------- verify cluster is ready ----------------------------------------------------------------------------------- + kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' + grep -q ready ++ mktemp + local LAST_OUT=/tmp/tmp.ltNFJB5nks ++ mktemp + local LAST_ERR=/tmp/tmp.JgKy7K3ggh + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get psmdb some-name -o 'jsonpath={.status.state}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ltNFJB5nks + cat /tmp/tmp.JgKy7K3ggh + rm /tmp/tmp.ltNFJB5nks /tmp/tmp.JgKy7K3ggh + return 0 + [[ -n '' ]] + [[ 0 == 1 ]] + desc 'test namespace resource quota enforcement' + set +o xtrace ----------------------------------------------------------------------------------- test namespace resource quota enforcement ----------------------------------------------------------------------------------- + apply_resourcequota 9Gi + local quota=9Gi + local default_sc ++ get_default_storageclass ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Pg2sOJAZa4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.1k5EmBVKjE ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Pg2sOJAZa4 ++ cat /tmp/tmp.1k5EmBVKjE ++ rm /tmp/tmp.Pg2sOJAZa4 /tmp/tmp.1k5EmBVKjE ++ return 0 + default_sc=standard-rwo + echo 'Applying resourcequota for default storageclass standard-rwo with quota 9Gi' Applying resourcequota for default storageclass standard-rwo with quota 9Gi + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/pvc-auto-resize/conf/resourcequota.yml + sed s/STORAGECLASS/standard-rwo/ + sed s/QUOTA/9Gi/ + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.hEmLCnWy8Q ++ mktemp + local LAST_ERR=/tmp/tmp.zeeER7zr6n + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.hEmLCnWy8Q resourcequota/default-storage-quota created + cat /tmp/tmp.zeeER7zr6n + rm /tmp/tmp.hEmLCnWy8Q /tmp/tmp.zeeER7zr6n + return 0 + echo 'Filling disk again to trigger autoscaling with quota limit' Filling disk again to trigger autoscaling with quota limit + fill_disk_to_threshold some-name-rs0-0 50 1700 + local pod=some-name-rs0-0 + local threshold=50 + local size_mb=1700 + echo 'Filling disk on some-name-rs0-0 to approximately 50% (creating 1700MB file)' Filling disk on some-name-rs0-0 to approximately 50% (creating 1700MB file) + kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=1700 2>/dev/null || true' ++ mktemp + local LAST_OUT=/tmp/tmp.rgbn1kxNBw ++ mktemp + local LAST_ERR=/tmp/tmp.YedpA20nsr + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=1700 2>/dev/null || true' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.rgbn1kxNBw + cat /tmp/tmp.YedpA20nsr + rm /tmp/tmp.rgbn1kxNBw /tmp/tmp.YedpA20nsr + return 0 + local usage ++ kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pgLulzEMjU +++ mktemp ++ local LAST_ERR=/tmp/tmp.oe0Jtj0MJz ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.pgLulzEMjU ++ cat /tmp/tmp.oe0Jtj0MJz ++ rm /tmp/tmp.pgLulzEMjU /tmp/tmp.oe0Jtj0MJz ++ return 0 + usage=65 + echo 'Current disk usage on some-name-rs0-0: 65%' Current disk usage on some-name-rs0-0: 65% + echo 'Waiting to verify PVC does not resize beyond quota (should stay at 3Gi)' Waiting to verify PVC does not resize beyond quota (should stay at 3Gi) + sleep 30 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MmPNfynTvd +++ mktemp ++ local LAST_ERR=/tmp/tmp.EMzO3mfuQD ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.MmPNfynTvd ++ cat /tmp/tmp.EMzO3mfuQD ++ rm /tmp/tmp.MmPNfynTvd /tmp/tmp.EMzO3mfuQD ++ return 0 + current_size=3Gi + [[ 3Gi != 3Gi ]] + desc 'increase quota to allow further autoscaling' + set +o xtrace ----------------------------------------------------------------------------------- increase quota to allow further autoscaling ----------------------------------------------------------------------------------- + apply_resourcequota 15Gi + local quota=15Gi + local default_sc ++ get_default_storageclass ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.t7RT0XZ5Rf +++ mktemp ++ local LAST_ERR=/tmp/tmp.EyyHbauBuX ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.t7RT0XZ5Rf ++ cat /tmp/tmp.EyyHbauBuX ++ rm /tmp/tmp.t7RT0XZ5Rf /tmp/tmp.EyyHbauBuX ++ return 0 + default_sc=standard-rwo + echo 'Applying resourcequota for default storageclass standard-rwo with quota 15Gi' Applying resourcequota for default storageclass standard-rwo with quota 15Gi + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/pvc-auto-resize/conf/resourcequota.yml + sed s/STORAGECLASS/standard-rwo/ + sed s/QUOTA/15Gi/ + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.iazkyHgVkF ++ mktemp + local LAST_ERR=/tmp/tmp.3DdFSs1cNp + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.iazkyHgVkF resourcequota/default-storage-quota configured + cat /tmp/tmp.3DdFSs1cNp + rm /tmp/tmp.iazkyHgVkF /tmp/tmp.3DdFSs1cNp + return 0 + wait_for_auto_resize mongod-data-some-name-rs0-0 5Gi 60 10 + local pvc=mongod-data-some-name-rs0-0 + local expected_size=5Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 5Gi' Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 5Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YQhK6r4x13 +++ mktemp ++ local LAST_ERR=/tmp/tmp.h8AHvOQFRU ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.YQhK6r4x13 ++ cat /tmp/tmp.h8AHvOQFRU ++ rm /tmp/tmp.YQhK6r4x13 /tmp/tmp.h8AHvOQFRU ++ return 0 + [[ 3Gi == 5Gi ]] + [[ 0 -ge 60 ]] + echo -n . .+ sleep 10 + retry=1 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WN4BPoPq95 +++ mktemp ++ local LAST_ERR=/tmp/tmp.WH46vs9XV0 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.WN4BPoPq95 ++ cat /tmp/tmp.WH46vs9XV0 ++ rm /tmp/tmp.WN4BPoPq95 /tmp/tmp.WH46vs9XV0 ++ return 0 + [[ 3Gi == 5Gi ]] + [[ 1 -ge 60 ]] + echo -n . .+ sleep 10 + retry=2 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5eH2SG5Xh4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.y6VJEYDYrc ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5eH2SG5Xh4 ++ cat /tmp/tmp.y6VJEYDYrc ++ rm /tmp/tmp.5eH2SG5Xh4 /tmp/tmp.y6VJEYDYrc ++ return 0 + [[ 3Gi == 5Gi ]] + [[ 2 -ge 60 ]] + echo -n . .+ sleep 10 + retry=3 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AjKqmJ3cdV +++ mktemp ++ local LAST_ERR=/tmp/tmp.RFoaz6ukKD ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.AjKqmJ3cdV ++ cat /tmp/tmp.RFoaz6ukKD ++ rm /tmp/tmp.AjKqmJ3cdV /tmp/tmp.RFoaz6ukKD ++ return 0 + [[ 3Gi == 5Gi ]] + [[ 3 -ge 60 ]] + echo -n . .+ sleep 10 + retry=4 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3a8osMUbQp +++ mktemp ++ local LAST_ERR=/tmp/tmp.TIqqDmfgGa ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.3a8osMUbQp ++ cat /tmp/tmp.TIqqDmfgGa ++ rm /tmp/tmp.3a8osMUbQp /tmp/tmp.TIqqDmfgGa ++ return 0 + [[ 3Gi == 5Gi ]] + [[ 4 -ge 60 ]] + echo -n . .+ sleep 10 + retry=5 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sf6oPSqSy7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.CS9XpdNkCn ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.sf6oPSqSy7 ++ cat /tmp/tmp.CS9XpdNkCn ++ rm /tmp/tmp.sf6oPSqSy7 /tmp/tmp.CS9XpdNkCn ++ return 0 + [[ 5Gi == 5Gi ]] + echo + echo 'PVC mongod-data-some-name-rs0-0 successfully resized to 5Gi' PVC mongod-data-some-name-rs0-0 successfully resized to 5Gi + wait_for_auto_resize mongod-data-some-name-rs0-1 5Gi 60 10 + local pvc=mongod-data-some-name-rs0-1 + local expected_size=5Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-1 to auto-resize to 5Gi' Waiting for PVC mongod-data-some-name-rs0-1 to auto-resize to 5Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UbRWKc83br +++ mktemp ++ local LAST_ERR=/tmp/tmp.90CnOUQ8jB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.UbRWKc83br ++ cat /tmp/tmp.90CnOUQ8jB ++ rm /tmp/tmp.UbRWKc83br /tmp/tmp.90CnOUQ8jB ++ return 0 + [[ 3Gi == 5Gi ]] + [[ 0 -ge 60 ]] + echo -n . .+ sleep 10 + retry=1 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eYCW6O1MBa +++ mktemp ++ local LAST_ERR=/tmp/tmp.QWAgSmCt9v ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.eYCW6O1MBa ++ cat /tmp/tmp.QWAgSmCt9v ++ rm /tmp/tmp.eYCW6O1MBa /tmp/tmp.QWAgSmCt9v ++ return 0 + [[ 3Gi == 5Gi ]] + [[ 1 -ge 60 ]] + echo -n . .+ sleep 10 + retry=2 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.z3ozg6QSRB +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ejs9QGLGnO ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.z3ozg6QSRB ++ cat /tmp/tmp.Ejs9QGLGnO ++ rm /tmp/tmp.z3ozg6QSRB /tmp/tmp.Ejs9QGLGnO ++ return 0 + [[ 3Gi == 5Gi ]] + [[ 2 -ge 60 ]] + echo -n . .+ sleep 10 + retry=3 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jjWCK8FygB +++ mktemp ++ local LAST_ERR=/tmp/tmp.RgV6NofZX8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jjWCK8FygB ++ cat /tmp/tmp.RgV6NofZX8 ++ rm /tmp/tmp.jjWCK8FygB /tmp/tmp.RgV6NofZX8 ++ return 0 + [[ 5Gi == 5Gi ]] + echo + echo 'PVC mongod-data-some-name-rs0-1 successfully resized to 5Gi' PVC mongod-data-some-name-rs0-1 successfully resized to 5Gi + wait_for_auto_resize mongod-data-some-name-rs0-2 5Gi 60 10 + local pvc=mongod-data-some-name-rs0-2 + local expected_size=5Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-2 to auto-resize to 5Gi' Waiting for PVC mongod-data-some-name-rs0-2 to auto-resize to 5Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-2 ++ local pvc=mongod-data-some-name-rs0-2 ++ kubectl_bin get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pbjqVPwF4z +++ mktemp ++ local LAST_ERR=/tmp/tmp.O7ImDAEiHp ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.pbjqVPwF4z ++ cat /tmp/tmp.O7ImDAEiHp ++ rm /tmp/tmp.pbjqVPwF4z /tmp/tmp.O7ImDAEiHp ++ return 0 + [[ 5Gi == 5Gi ]] + echo + echo 'PVC mongod-data-some-name-rs0-2 successfully resized to 5Gi' PVC mongod-data-some-name-rs0-2 successfully resized to 5Gi + desc 'verify cluster consistency after quota test' + set +o xtrace ----------------------------------------------------------------------------------- verify cluster consistency after quota test ----------------------------------------------------------------------------------- + wait_cluster_consistency some-name 64 + local cluster_name=some-name + local wait_time=64 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uV2sifYdea +++ mktemp ++ local LAST_ERR=/tmp/tmp.uZC0Sl8AY8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.uV2sifYdea ++ cat /tmp/tmp.uZC0Sl8AY8 ++ rm /tmp/tmp.uV2sifYdea /tmp/tmp.uZC0Sl8AY8 ++ return 0 + [[ error == ready ]] + let retry+=1 + '[' 1 -ge 64 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XOuJ3zZ5GI +++ mktemp ++ local LAST_ERR=/tmp/tmp.7L6gZ3etWO ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.XOuJ3zZ5GI ++ cat /tmp/tmp.7L6gZ3etWO ++ rm /tmp/tmp.XOuJ3zZ5GI /tmp/tmp.7L6gZ3etWO ++ return 0 + [[ ready == ready ]] + echo .OK .OK + desc 'test maxSize limit' + set +o xtrace ----------------------------------------------------------------------------------- test maxSize limit ----------------------------------------------------------------------------------- + apply_resourcequota 30Gi + local quota=30Gi + local default_sc ++ get_default_storageclass ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KCpEYhvY9F +++ mktemp ++ local LAST_ERR=/tmp/tmp.VQM7FGTKbk ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.KCpEYhvY9F ++ cat /tmp/tmp.VQM7FGTKbk ++ rm /tmp/tmp.KCpEYhvY9F /tmp/tmp.VQM7FGTKbk ++ return 0 + default_sc=standard-rwo + echo 'Applying resourcequota for default storageclass standard-rwo with quota 30Gi' Applying resourcequota for default storageclass standard-rwo with quota 30Gi + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/pvc-auto-resize/conf/resourcequota.yml + sed s/STORAGECLASS/standard-rwo/ + sed s/QUOTA/30Gi/ + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.qGD0nKR43L ++ mktemp + local LAST_ERR=/tmp/tmp.EUUZiOPuHk + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.qGD0nKR43L resourcequota/default-storage-quota configured + cat /tmp/tmp.EUUZiOPuHk + rm /tmp/tmp.qGD0nKR43L /tmp/tmp.EUUZiOPuHk + return 0 + echo 'Continuing to fill disk to test maxSize enforcement' Continuing to fill disk to test maxSize enforcement + fill_disk_to_threshold some-name-rs0-0 50 2900 + local pod=some-name-rs0-0 + local threshold=50 + local size_mb=2900 + echo 'Filling disk on some-name-rs0-0 to approximately 50% (creating 2900MB file)' Filling disk on some-name-rs0-0 to approximately 50% (creating 2900MB file) + kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=2900 2>/dev/null || true' ++ mktemp + local LAST_OUT=/tmp/tmp.Nu4fNtKaIR ++ mktemp + local LAST_ERR=/tmp/tmp.GxFYt6Fl6X + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=2900 2>/dev/null || true' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Nu4fNtKaIR + cat /tmp/tmp.GxFYt6Fl6X + rm /tmp/tmp.Nu4fNtKaIR /tmp/tmp.GxFYt6Fl6X + return 0 + local usage ++ kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cJ3DSOUIcB +++ mktemp ++ local LAST_ERR=/tmp/tmp.Lsg3NkAFNf ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.cJ3DSOUIcB ++ cat /tmp/tmp.Lsg3NkAFNf ++ rm /tmp/tmp.cJ3DSOUIcB /tmp/tmp.Lsg3NkAFNf ++ return 0 + usage=63 + echo 'Current disk usage on some-name-rs0-0: 63%' Current disk usage on some-name-rs0-0: 63% + wait_for_auto_resize mongod-data-some-name-rs0-0 7Gi 60 10 + local pvc=mongod-data-some-name-rs0-0 + local expected_size=7Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 7Gi' Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 7Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.R1jgxYS3ZZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.F8wCdERgAO ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.R1jgxYS3ZZ ++ cat /tmp/tmp.F8wCdERgAO ++ rm /tmp/tmp.R1jgxYS3ZZ /tmp/tmp.F8wCdERgAO ++ return 0 + [[ 5Gi == 7Gi ]] + [[ 0 -ge 60 ]] + echo -n . .+ sleep 10 + retry=1 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.V6O56wjBMK +++ mktemp ++ local LAST_ERR=/tmp/tmp.PkR9Jco9lL ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.V6O56wjBMK ++ cat /tmp/tmp.PkR9Jco9lL ++ rm /tmp/tmp.V6O56wjBMK /tmp/tmp.PkR9Jco9lL ++ return 0 + [[ 5Gi == 7Gi ]] + [[ 1 -ge 60 ]] + echo -n . .+ sleep 10 + retry=2 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SkhMW5qhvQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.79I5AfdYNN ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.SkhMW5qhvQ ++ cat /tmp/tmp.79I5AfdYNN ++ rm /tmp/tmp.SkhMW5qhvQ /tmp/tmp.79I5AfdYNN ++ return 0 + [[ 5Gi == 7Gi ]] + [[ 2 -ge 60 ]] + echo -n . .+ sleep 10 + retry=3 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lmcA5roOF7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.SddtC5zVXp ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.lmcA5roOF7 ++ cat /tmp/tmp.SddtC5zVXp ++ rm /tmp/tmp.lmcA5roOF7 /tmp/tmp.SddtC5zVXp ++ return 0 + [[ 5Gi == 7Gi ]] + [[ 3 -ge 60 ]] + echo -n . .+ sleep 10 + retry=4 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Rgmbg2aaOr +++ mktemp ++ local LAST_ERR=/tmp/tmp.fluhduZZjT ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Rgmbg2aaOr ++ cat /tmp/tmp.fluhduZZjT ++ rm /tmp/tmp.Rgmbg2aaOr /tmp/tmp.fluhduZZjT ++ return 0 + [[ 5Gi == 7Gi ]] + [[ 4 -ge 60 ]] + echo -n . .+ sleep 10 + retry=5 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zU6A7nAzzZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.rkoDEUVd1o ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.zU6A7nAzzZ ++ cat /tmp/tmp.rkoDEUVd1o ++ rm /tmp/tmp.zU6A7nAzzZ /tmp/tmp.rkoDEUVd1o ++ return 0 + [[ 5Gi == 7Gi ]] + [[ 5 -ge 60 ]] + echo -n . .+ sleep 10 + retry=6 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ynd7bRszah +++ mktemp ++ local LAST_ERR=/tmp/tmp.fCbNE7IDg3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Ynd7bRszah ++ cat /tmp/tmp.fCbNE7IDg3 ++ rm /tmp/tmp.Ynd7bRszah /tmp/tmp.fCbNE7IDg3 ++ return 0 + [[ 5Gi == 7Gi ]] + [[ 6 -ge 60 ]] + echo -n . .+ sleep 10 + retry=7 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GNYnRsAiMK +++ mktemp ++ local LAST_ERR=/tmp/tmp.5p74fR6qx2 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.GNYnRsAiMK ++ cat /tmp/tmp.5p74fR6qx2 ++ rm /tmp/tmp.GNYnRsAiMK /tmp/tmp.5p74fR6qx2 ++ return 0 + [[ 7Gi == 7Gi ]] + echo + echo 'PVC mongod-data-some-name-rs0-0 successfully resized to 7Gi' PVC mongod-data-some-name-rs0-0 successfully resized to 7Gi + wait_for_auto_resize mongod-data-some-name-rs0-1 7Gi 60 10 + local pvc=mongod-data-some-name-rs0-1 + local expected_size=7Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-1 to auto-resize to 7Gi' Waiting for PVC mongod-data-some-name-rs0-1 to auto-resize to 7Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TipfNYlE76 +++ mktemp ++ local LAST_ERR=/tmp/tmp.PuhHivXs2B ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.TipfNYlE76 ++ cat /tmp/tmp.PuhHivXs2B ++ rm /tmp/tmp.TipfNYlE76 /tmp/tmp.PuhHivXs2B ++ return 0 + [[ 7Gi == 7Gi ]] + echo + echo 'PVC mongod-data-some-name-rs0-1 successfully resized to 7Gi' PVC mongod-data-some-name-rs0-1 successfully resized to 7Gi + wait_for_auto_resize mongod-data-some-name-rs0-2 7Gi 60 10 + local pvc=mongod-data-some-name-rs0-2 + local expected_size=7Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-2 to auto-resize to 7Gi' Waiting for PVC mongod-data-some-name-rs0-2 to auto-resize to 7Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-2 ++ local pvc=mongod-data-some-name-rs0-2 ++ kubectl_bin get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MeHQxUfAME +++ mktemp ++ local LAST_ERR=/tmp/tmp.X8xHCJy2Ck ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.MeHQxUfAME ++ cat /tmp/tmp.X8xHCJy2Ck ++ rm /tmp/tmp.MeHQxUfAME /tmp/tmp.X8xHCJy2Ck ++ return 0 + [[ 7Gi == 7Gi ]] + echo + echo 'PVC mongod-data-some-name-rs0-2 successfully resized to 7Gi' PVC mongod-data-some-name-rs0-2 successfully resized to 7Gi + fill_disk_to_threshold some-name-rs0-0 50 4000 + local pod=some-name-rs0-0 + local threshold=50 + local size_mb=4000 + echo 'Filling disk on some-name-rs0-0 to approximately 50% (creating 4000MB file)' Filling disk on some-name-rs0-0 to approximately 50% (creating 4000MB file) + kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=4000 2>/dev/null || true' ++ mktemp + local LAST_OUT=/tmp/tmp.sADtDwhb1S ++ mktemp + local LAST_ERR=/tmp/tmp.8v71WSjd9U + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=4000 2>/dev/null || true' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.sADtDwhb1S + cat /tmp/tmp.8v71WSjd9U + rm /tmp/tmp.sADtDwhb1S /tmp/tmp.8v71WSjd9U + return 0 + local usage ++ kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bN54GcWmmE +++ mktemp ++ local LAST_ERR=/tmp/tmp.lLN9fZbxV6 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.bN54GcWmmE ++ cat /tmp/tmp.lLN9fZbxV6 ++ rm /tmp/tmp.bN54GcWmmE /tmp/tmp.lLN9fZbxV6 ++ return 0 + usage=61 + echo 'Current disk usage on some-name-rs0-0: 61%' Current disk usage on some-name-rs0-0: 61% + wait_for_auto_resize mongod-data-some-name-rs0-0 9Gi 60 10 + local pvc=mongod-data-some-name-rs0-0 + local expected_size=9Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 9Gi' Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 9Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mHmYqOVCZx +++ mktemp ++ local LAST_ERR=/tmp/tmp.5QKXYp25hS ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.mHmYqOVCZx ++ cat /tmp/tmp.5QKXYp25hS ++ rm /tmp/tmp.mHmYqOVCZx /tmp/tmp.5QKXYp25hS ++ return 0 + [[ 7Gi == 9Gi ]] + [[ 0 -ge 60 ]] + echo -n . .+ sleep 10 + retry=1 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.k1v06xfjY1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.jUOlsheUTD ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.k1v06xfjY1 ++ cat /tmp/tmp.jUOlsheUTD ++ rm /tmp/tmp.k1v06xfjY1 /tmp/tmp.jUOlsheUTD ++ return 0 + [[ 7Gi == 9Gi ]] + [[ 1 -ge 60 ]] + echo -n . .+ sleep 10 + retry=2 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bRw3vh0q3C +++ mktemp ++ local LAST_ERR=/tmp/tmp.jlRabPhWGv ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.bRw3vh0q3C ++ cat /tmp/tmp.jlRabPhWGv ++ rm /tmp/tmp.bRw3vh0q3C /tmp/tmp.jlRabPhWGv ++ return 0 + [[ 7Gi == 9Gi ]] + [[ 2 -ge 60 ]] + echo -n . .+ sleep 10 + retry=3 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XkR1aR2ZDA +++ mktemp ++ local LAST_ERR=/tmp/tmp.zfqsU8xWdh ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.XkR1aR2ZDA ++ cat /tmp/tmp.zfqsU8xWdh ++ rm /tmp/tmp.XkR1aR2ZDA /tmp/tmp.zfqsU8xWdh ++ return 0 + [[ 9Gi == 9Gi ]] + echo + echo 'PVC mongod-data-some-name-rs0-0 successfully resized to 9Gi' PVC mongod-data-some-name-rs0-0 successfully resized to 9Gi + wait_for_auto_resize mongod-data-some-name-rs0-1 9Gi 60 10 + local pvc=mongod-data-some-name-rs0-1 + local expected_size=9Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-1 to auto-resize to 9Gi' Waiting for PVC mongod-data-some-name-rs0-1 to auto-resize to 9Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pKkaQ4HQa7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.pS7YenbGQp ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.pKkaQ4HQa7 ++ cat /tmp/tmp.pS7YenbGQp ++ rm /tmp/tmp.pKkaQ4HQa7 /tmp/tmp.pS7YenbGQp ++ return 0 + [[ 7Gi == 9Gi ]] + [[ 0 -ge 60 ]] + echo -n . .+ sleep 10 + retry=1 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cuRhbOa7cZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.Zv2wsw98up ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.cuRhbOa7cZ ++ cat /tmp/tmp.Zv2wsw98up ++ rm /tmp/tmp.cuRhbOa7cZ /tmp/tmp.Zv2wsw98up ++ return 0 + [[ 9Gi == 9Gi ]] + echo + echo 'PVC mongod-data-some-name-rs0-1 successfully resized to 9Gi' PVC mongod-data-some-name-rs0-1 successfully resized to 9Gi + wait_for_auto_resize mongod-data-some-name-rs0-2 9Gi 60 10 + local pvc=mongod-data-some-name-rs0-2 + local expected_size=9Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-2 to auto-resize to 9Gi' Waiting for PVC mongod-data-some-name-rs0-2 to auto-resize to 9Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-2 ++ local pvc=mongod-data-some-name-rs0-2 ++ kubectl_bin get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XWD0VIzi4R +++ mktemp ++ local LAST_ERR=/tmp/tmp.j0d4PB8TWq ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.XWD0VIzi4R ++ cat /tmp/tmp.j0d4PB8TWq ++ rm /tmp/tmp.XWD0VIzi4R /tmp/tmp.j0d4PB8TWq ++ return 0 + [[ 7Gi == 9Gi ]] + [[ 0 -ge 60 ]] + echo -n . .+ sleep 10 + retry=1 ++ get_pvc_size mongod-data-some-name-rs0-2 ++ local pvc=mongod-data-some-name-rs0-2 ++ kubectl_bin get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ChKMocU0T1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.7BQ6GVWFzZ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ChKMocU0T1 ++ cat /tmp/tmp.7BQ6GVWFzZ ++ rm /tmp/tmp.ChKMocU0T1 /tmp/tmp.7BQ6GVWFzZ ++ return 0 + [[ 9Gi == 9Gi ]] + echo + echo 'PVC mongod-data-some-name-rs0-2 successfully resized to 9Gi' PVC mongod-data-some-name-rs0-2 successfully resized to 9Gi + fill_disk_to_threshold some-name-rs0-0 50 5200 + local pod=some-name-rs0-0 + local threshold=50 + local size_mb=5200 + echo 'Filling disk on some-name-rs0-0 to approximately 50% (creating 5200MB file)' Filling disk on some-name-rs0-0 to approximately 50% (creating 5200MB file) + kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=5200 2>/dev/null || true' ++ mktemp + local LAST_OUT=/tmp/tmp.C0hKToOA76 ++ mktemp + local LAST_ERR=/tmp/tmp.zvJU3JaZox + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=5200 2>/dev/null || true' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.C0hKToOA76 + cat /tmp/tmp.zvJU3JaZox + rm /tmp/tmp.C0hKToOA76 /tmp/tmp.zvJU3JaZox + return 0 + local usage ++ kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' +++ mktemp ++ local LAST_OUT=/tmp/tmp.w3M8YHpFKc +++ mktemp ++ local LAST_ERR=/tmp/tmp.NKwVbRQpO8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.w3M8YHpFKc ++ cat /tmp/tmp.NKwVbRQpO8 ++ rm /tmp/tmp.w3M8YHpFKc /tmp/tmp.NKwVbRQpO8 ++ return 0 + usage=61 + echo 'Current disk usage on some-name-rs0-0: 61%' Current disk usage on some-name-rs0-0: 61% ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xmIZIys4CG +++ mktemp ++ local LAST_ERR=/tmp/tmp.48a1uZYGSd ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xmIZIys4CG ++ cat /tmp/tmp.48a1uZYGSd ++ rm /tmp/tmp.xmIZIys4CG /tmp/tmp.48a1uZYGSd ++ return 0 + final_size=10Gi + [[ 10Gi == 10Gi ]] + echo 'PVC correctly stopped at maxSize 10Gi' PVC correctly stopped at maxSize 10Gi + desc 'verify final autoscaling status' + set +o xtrace ----------------------------------------------------------------------------------- verify final autoscaling status ----------------------------------------------------------------------------------- + check_autoscaling_status some-name 4 mongod-data-some-name-rs0-0 mongod-data-some-name-rs0-1 mongod-data-some-name-rs0-2 + local cluster=some-name + local expected_resize_count=4 + shift 2 + pvc_names=('mongod-data-some-name-rs0-0' 'mongod-data-some-name-rs0-1' 'mongod-data-some-name-rs0-2') + local pvc_names + local max_retry=30 + local sleep_time=10 + echo 'Checking autoscaling status for PVCs: mongod-data-some-name-rs0-0 mongod-data-some-name-rs0-1 mongod-data-some-name-rs0-2' Checking autoscaling status for PVCs: mongod-data-some-name-rs0-0 mongod-data-some-name-rs0-1 mongod-data-some-name-rs0-2 + echo 'Expected resize count per PVC: 4' Expected resize count per PVC: 4 + for pvc_name in "${pvc_names[@]}" + local retry=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-0.resizeCount}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AB7JKhoQ4f +++ mktemp ++ local LAST_ERR=/tmp/tmp.uIEVLnPEoC ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-0.resizeCount}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.AB7JKhoQ4f ++ cat /tmp/tmp.uIEVLnPEoC ++ rm /tmp/tmp.AB7JKhoQ4f /tmp/tmp.uIEVLnPEoC ++ return 0 + [[ 4 -eq 4 ]] + for pvc_name in "${pvc_names[@]}" + local retry=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-1.resizeCount}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pBg4idXWxv +++ mktemp ++ local LAST_ERR=/tmp/tmp.A9Dw6LtkS8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-1.resizeCount}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.pBg4idXWxv ++ cat /tmp/tmp.A9Dw6LtkS8 ++ rm /tmp/tmp.pBg4idXWxv /tmp/tmp.A9Dw6LtkS8 ++ return 0 + [[ 4 -eq 4 ]] + for pvc_name in "${pvc_names[@]}" + local retry=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-2.resizeCount}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.w1eU3dFSyx +++ mktemp ++ local LAST_ERR=/tmp/tmp.S7W2pcH7Q8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-2.resizeCount}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.w1eU3dFSyx ++ cat /tmp/tmp.S7W2pcH7Q8 ++ rm /tmp/tmp.w1eU3dFSyx /tmp/tmp.S7W2pcH7Q8 ++ return 0 + [[ 4 -eq 4 ]] + wait_cluster_consistency some-name 64 + local cluster_name=some-name + local wait_time=64 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DOaxdsh4Mx +++ mktemp ++ local LAST_ERR=/tmp/tmp.Xs9bbvziR0 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.DOaxdsh4Mx ++ cat /tmp/tmp.Xs9bbvziR0 ++ rm /tmp/tmp.DOaxdsh4Mx /tmp/tmp.Xs9bbvziR0 ++ return 0 + [[ initializing == ready ]] + let retry+=1 + '[' 1 -ge 64 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.F7oBHpiQl0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.LlalSYvgpG ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.F7oBHpiQl0 ++ cat /tmp/tmp.LlalSYvgpG ++ rm /tmp/tmp.F7oBHpiQl0 /tmp/tmp.LlalSYvgpG ++ return 0 + [[ initializing == ready ]] + let retry+=1 + '[' 2 -ge 64 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.R8SCBzLeJL +++ mktemp ++ local LAST_ERR=/tmp/tmp.zqpNretcHA ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.R8SCBzLeJL ++ cat /tmp/tmp.zqpNretcHA ++ rm /tmp/tmp.R8SCBzLeJL /tmp/tmp.zqpNretcHA ++ return 0 + [[ initializing == ready ]] + let retry+=1 + '[' 3 -ge 64 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eBgcc0foui +++ mktemp ++ local LAST_ERR=/tmp/tmp.pK5gG3vUXe ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.eBgcc0foui ++ cat /tmp/tmp.pK5gG3vUXe ++ rm /tmp/tmp.eBgcc0foui /tmp/tmp.pK5gG3vUXe ++ return 0 + [[ initializing == ready ]] + let retry+=1 + '[' 4 -ge 64 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.R3xF4PuZyr +++ mktemp ++ local LAST_ERR=/tmp/tmp.uFgIcRsKyb ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.R3xF4PuZyr ++ cat /tmp/tmp.uFgIcRsKyb ++ rm /tmp/tmp.R3xF4PuZyr /tmp/tmp.uFgIcRsKyb ++ return 0 + [[ ready == ready ]] + echo .OK .OK + destroy pvc-auto-resize-16581 + local namespace=pvc-auto-resize-16581 + local ignore_logs=true + [[ 0 == 1 ]] + desc 'destroy cluster/operator and all other resources' + set +o xtrace ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + '[' true == false ']' + delete_backups + desc 'Delete psmdb-backup' + set +o xtrace ----------------------------------------------------------------------------------- Delete psmdb-backup ----------------------------------------------------------------------------------- ++ kubectl_bin get psmdb-backup --no-headers ++ wc -l +++ mktemp ++ local LAST_OUT=/tmp/tmp.dfMn39ZiN7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.qrRUCeJd1K ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb-backup --no-headers ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.dfMn39ZiN7 ++ cat /tmp/tmp.qrRUCeJd1K No resources found in pvc-auto-resize-16581 namespace. ++ rm /tmp/tmp.dfMn39ZiN7 /tmp/tmp.qrRUCeJd1K ++ return 0 + '[' 0 '!=' 0 ']' + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.LsotXAP4sS ++ mktemp + local LAST_ERR=/tmp/tmp.DR3X7WBR2C + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.LsotXAP4sS customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.DR3X7WBR2C + rm /tmp/tmp.LsotXAP4sS /tmp/tmp.DR3X7WBR2C + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/crd.yaml ++ grep -v '\-\-\-' grep: warning: stray \ before - grep: warning: stray \ before - + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.PdBPlcadTy ++ mktemp + local LAST_ERR=/tmp/tmp.DhD1PPAv6p + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.PdBPlcadTy + cat /tmp/tmp.DhD1PPAv6p + rm /tmp/tmp.PdBPlcadTy /tmp/tmp.DhD1PPAv6p + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.FL6sE8UJLK ++ mktemp + local LAST_ERR=/tmp/tmp.cMbg0h8kef + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.FL6sE8UJLK + cat /tmp/tmp.cMbg0h8kef + rm /tmp/tmp.FL6sE8UJLK /tmp/tmp.cMbg0h8kef + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch perconaservermongodbs.psmdb.percona.com -n pvc-auto-resize-16581 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaservermongodb.psmdb.percona.com/some-name patched + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.G96MCgVJN5 ++ mktemp + local LAST_ERR=/tmp/tmp.qCZEGmSSpD + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.G96MCgVJN5 customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com condition met + cat /tmp/tmp.qCZEGmSSpD + rm /tmp/tmp.G96MCgVJN5 /tmp/tmp.qCZEGmSSpD + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.OuMIMSpyZ1 ++ mktemp + local LAST_ERR=/tmp/tmp.gLtisw13ec + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.OuMIMSpyZ1 clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.gLtisw13ec + rm /tmp/tmp.OuMIMSpyZ1 /tmp/tmp.gLtisw13ec + return 0 + destroy_cert_manager + kubectl_bin delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.rd92RUDkR4 ++ mktemp + local LAST_ERR=/tmp/tmp.U2eq1hfWkT + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.rd92RUDkR4 + cat /tmp/tmp.U2eq1hfWkT Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 0 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.rd92RUDkR4 + cat /tmp/tmp.U2eq1hfWkT Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 4 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.rd92RUDkR4 + cat /tmp/tmp.U2eq1hfWkT Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 8 + cat /tmp/tmp.rd92RUDkR4 + cat /tmp/tmp.U2eq1hfWkT Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + rm /tmp/tmp.rd92RUDkR4 /tmp/tmp.U2eq1hfWkT + return 1 + true + '[' -n '' ']' + '[' -n psmdb-operator ']' + kubectl_bin delete --grace-period=0 --force=true namespace pvc-auto-resize-16581 + rm -rf /tmp/tmp.IJmbDctCRT + kubectl_bin delete --grace-period=0 --force=true namespace psmdb-operator ++ mktemp ++ mktemp + desc 'test passed' + set +o xtrace ----------------------------------------------------------------------------------- test passed ----------------------------------------------------------------------------------- + local LAST_OUT=/tmp/tmp.JsFOvaKTtW + local LAST_OUT=/tmp/tmp.lcJa3FrCKu ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.IrE0oc90pN + local exit_status=0 + local timeout=4 + local LAST_ERR=/tmp/tmp.k1JhNqMTj0 + local exit_status=0 + local timeout=4 ++ seq 0 2 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete --grace-period=0 --force=true namespace pvc-auto-resize-16581 + for i in $(seq 0 2) + set +e + kubectl delete --grace-period=0 --force=true namespace psmdb-operator