Log: /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/e2e-tests/logs/pvc-auto-resize.log Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 + '[' 0 == 1 ']' + ensure_default_sc_allows_expansion + local default_sc ++ get_default_storageclass ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yonZhuXFWT +++ mktemp ++ local LAST_ERR=/tmp/tmp.SobrwChTex ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.yonZhuXFWT ++ cat /tmp/tmp.SobrwChTex ++ rm /tmp/tmp.yonZhuXFWT /tmp/tmp.SobrwChTex ++ return 0 + default_sc=standard-rwo + echo 'Checking if default storageclass standard-rwo allows volume expansion' Checking if default storageclass standard-rwo allows volume expansion + local allowVolumeExpansion ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.name=="standard-rwo")].allowVolumeExpansion}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9EvfCEDoxq +++ mktemp ++ local LAST_ERR=/tmp/tmp.K7Brbublvv ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.name=="standard-rwo")].allowVolumeExpansion}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9EvfCEDoxq ++ cat /tmp/tmp.K7Brbublvv ++ rm /tmp/tmp.9EvfCEDoxq /tmp/tmp.K7Brbublvv ++ return 0 + allowVolumeExpansion=true + [[ true != \t\r\u\e ]] + create_infra pvc-auto-resize-25723 + local ns=pvc-auto-resize-25723 + [[ 1 == 1 ]] + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.w0fG1rVmBe ++ mktemp + local LAST_ERR=/tmp/tmp.Cjksg33FRu + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.w0fG1rVmBe customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.Cjksg33FRu + rm /tmp/tmp.w0fG1rVmBe /tmp/tmp.Cjksg33FRu + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/deploy/crd.yaml ++ grep -v '\-\-\-' grep: warning: stray \ before - grep: warning: stray \ before - + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.Qqn0gDVLVV ++ mktemp + local LAST_ERR=/tmp/tmp.H6OIHu4zjL + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Qqn0gDVLVV + cat /tmp/tmp.H6OIHu4zjL + rm /tmp/tmp.Qqn0gDVLVV /tmp/tmp.H6OIHu4zjL + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.UBh5J5dRkN ++ mktemp + local LAST_ERR=/tmp/tmp.FWbOsWTVdu + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.UBh5J5dRkN + cat /tmp/tmp.FWbOsWTVdu + rm /tmp/tmp.UBh5J5dRkN /tmp/tmp.FWbOsWTVdu + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbs" + kubectl patch perconaservermongodbs.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbs" + : + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.hGONgcD9BN ++ mktemp + local LAST_ERR=/tmp/tmp.XkICWBBbJP + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.hGONgcD9BN + cat /tmp/tmp.XkICWBBbJP + rm /tmp/tmp.hGONgcD9BN /tmp/tmp.XkICWBBbJP + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.pjzg5vKaeu ++ mktemp + local LAST_ERR=/tmp/tmp.ry8LG9mV29 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.pjzg5vKaeu clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.ry8LG9mV29 + rm /tmp/tmp.pjzg5vKaeu /tmp/tmp.ry8LG9mV29 + return 0 + check_crd_for_deletion PR-2226-1e37b178 + local git_tag=PR-2226-1e37b178 ++ curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/PR-2226-1e37b178/deploy/crd.yaml ++ yq eval .metadata.name ++ /usr/sbin/sed s/---//g ++ /usr/sbin/sed ':a;N;$!ba;s/\n/ /g' + for crd_name in $(curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/${git_tag}/deploy/crd.yaml | yq eval '.metadata.name' | $sed 's/---//g' | $sed ':a;N;$!ba;s/\n/ /g') ++ kubectl_bin get crd/null -o 'jsonpath={.status.conditions[-1].type}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bFLIY7OKDY +++ mktemp ++ local LAST_ERR=/tmp/tmp.uMpUu4wweG ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.bFLIY7OKDY ++ cat /tmp/tmp.uMpUu4wweG Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 0 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.bFLIY7OKDY ++ cat /tmp/tmp.uMpUu4wweG Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 4 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.bFLIY7OKDY ++ cat /tmp/tmp.uMpUu4wweG Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 8 ++ cat /tmp/tmp.bFLIY7OKDY ++ cat /tmp/tmp.uMpUu4wweG Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ rm /tmp/tmp.bFLIY7OKDY /tmp/tmp.uMpUu4wweG ++ return 1 + [[ '' == \T\e\r\m\i\n\a\t\i\n\g ]] + '[' -n psmdb-operator ']' + create_namespace psmdb-operator + local namespace=psmdb-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + grep -E -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' + awk '{print$1}' ++ mktemp + '[' -n '' ']' + desc 'cleaned up old namespaces psmdb-operator' + set +o xtrace + xargs kubectl delete ns ----------------------------------------------------------------------------------- cleaned up old namespaces psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace psmdb-operator --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.z9HSJRLCAe ++ mktemp + local LAST_OUT=/tmp/tmp.5w7nkEQAD5 ++ mktemp + local LAST_ERR=/tmp/tmp.3lss3ZOYRH + local exit_status=0 + local timeout=4 ++ seq 0 2 + local LAST_ERR=/tmp/tmp.tdTr6PF9Q1 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get ns + for i in $(seq 0 2) + set +e + kubectl delete namespace psmdb-operator --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.z9HSJRLCAe + cat /tmp/tmp.3lss3ZOYRH + rm /tmp/tmp.z9HSJRLCAe /tmp/tmp.3lss3ZOYRH + return 0 namespace "pvc-auto-resize-3262" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.5w7nkEQAD5 namespace "psmdb-operator" deleted + cat /tmp/tmp.tdTr6PF9Q1 + rm /tmp/tmp.5w7nkEQAD5 /tmp/tmp.tdTr6PF9Q1 + return 0 + kubectl_bin wait --for=delete namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.MXhj9At2MM ++ mktemp + local LAST_ERR=/tmp/tmp.2tVHvQ73Xq + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.MXhj9At2MM + cat /tmp/tmp.2tVHvQ73Xq + rm /tmp/tmp.MXhj9At2MM /tmp/tmp.2tVHvQ73Xq + return 0 + desc 'create namespace psmdb-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.eEyjYQFNw5 ++ mktemp + local LAST_ERR=/tmp/tmp.LWBZQLjFRY + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.eEyjYQFNw5 namespace/psmdb-operator created + cat /tmp/tmp.LWBZQLjFRY + rm /tmp/tmp.eEyjYQFNw5 /tmp/tmp.LWBZQLjFRY + return 0 + set_kube_ctx psmdb-operator + local namespace=psmdb-operator ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.VYZ7ozCOts +++ mktemp ++ local LAST_ERR=/tmp/tmp.mgpcUjOrMm ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.VYZ7ozCOts ++ cat /tmp/tmp.mgpcUjOrMm ++ rm /tmp/tmp.VYZ7ozCOts /tmp/tmp.mgpcUjOrMm ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2226-1e37b178-2-cluster7 --namespace=psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.0p5vI2QSSC ++ mktemp + local LAST_ERR=/tmp/tmp.2KBA8FNNg5 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2226-1e37b178-2-cluster7 --namespace=psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.0p5vI2QSSC Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2226-1e37b178-2-cluster7" modified. + cat /tmp/tmp.2KBA8FNNg5 + rm /tmp/tmp.0p5vI2QSSC /tmp/tmp.2KBA8FNNg5 + return 0 + deploy_operator + desc 'start PSMDB operator: perconalab/percona-server-mongodb-operator:PR-2226-1e37b178' + set +o xtrace ----------------------------------------------------------------------------------- start PSMDB operator: perconalab/percona-server-mongodb-operator:PR-2226-1e37b178 ----------------------------------------------------------------------------------- + local cr_file + '[' -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/e2e-tests/pvc-auto-resize/conf/crd.yaml ']' + cr_file=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/deploy/crd.yaml + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.oyn3GBlVKe ++ mktemp + local LAST_ERR=/tmp/tmp.EUaLOXkiVN + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.oyn3GBlVKe customresourcedefinition.apiextensions.k8s.io/perconaservermongodbbackups.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbrestores.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com serverside-applied + cat /tmp/tmp.EUaLOXkiVN + rm /tmp/tmp.oyn3GBlVKe /tmp/tmp.EUaLOXkiVN + return 0 + '[' -n psmdb-operator ']' + apply_rbac cw-rbac + local operator_namespace=psmdb-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: psmdb-operator^' + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.4YRCqx6Ccg ++ mktemp + local LAST_ERR=/tmp/tmp.YJ1pSer0jd + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.4YRCqx6Ccg clusterrole.rbac.authorization.k8s.io/percona-server-mongodb-operator created serviceaccount/percona-server-mongodb-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-server-mongodb-operator created + cat /tmp/tmp.YJ1pSer0jd + rm /tmp/tmp.4YRCqx6Ccg /tmp/tmp.YJ1pSer0jd + return 0 + yq eval ' (.spec.template.spec.containers[].image = "perconalab/percona-server-mongodb-operator:PR-2226-1e37b178") | ((.. | select(.[] == "DISABLE_TELEMETRY")) |= .value="true") | ((.. | select(.[] == "LOG_LEVEL")) |= .value="DEBUG")' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/deploy/cw-operator.yaml + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.UShwTX5V0l ++ mktemp + local LAST_ERR=/tmp/tmp.2zJO0gRmNN + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.UShwTX5V0l deployment.apps/percona-server-mongodb-operator created + cat /tmp/tmp.2zJO0gRmNN + rm /tmp/tmp.UShwTX5V0l /tmp/tmp.2zJO0gRmNN + return 0 + sleep 20 ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.4lctY0fIdI +++ mktemp ++ local LAST_ERR=/tmp/tmp.EStPDFtXSk ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4lctY0fIdI ++ cat /tmp/tmp.EStPDFtXSk ++ rm /tmp/tmp.4lctY0fIdI /tmp/tmp.EStPDFtXSk ++ return 0 + wait_operator_pod percona-server-mongodb-operator-bff7975f5-tpw72 + local pod=percona-server-mongodb-operator-bff7975f5-tpw72 + set +o xtrace waiting for pod/percona-server-mongodb-operator-bff7975f5-tpw72 to be ready.OK + echo 'Print operator info from log' Print operator info from log + grep 'Manager starting up' ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.aQNuGrVE5K +++ mktemp ++ local LAST_ERR=/tmp/tmp.6EAdrX2osD ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.aQNuGrVE5K ++ cat /tmp/tmp.6EAdrX2osD ++ rm /tmp/tmp.aQNuGrVE5K /tmp/tmp.6EAdrX2osD ++ return 0 + kubectl_bin logs -n psmdb-operator percona-server-mongodb-operator-bff7975f5-tpw72 ++ mktemp + local LAST_OUT=/tmp/tmp.aNTOgbG2hC ++ mktemp + local LAST_ERR=/tmp/tmp.yuffqf3PDN + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl logs -n psmdb-operator percona-server-mongodb-operator-bff7975f5-tpw72 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.aNTOgbG2hC + cat /tmp/tmp.yuffqf3PDN + rm /tmp/tmp.aNTOgbG2hC /tmp/tmp.yuffqf3PDN + return 0 2026-02-03T15:46:26.521Z INFO setup Manager starting up {"gitCommit": "1e37b17888853ae8e125b9de8658b67d796c4f87", "gitBranch": "PR-2226-1e37b178", "buildTime": "", "goVersion": "go1.25.6", "os": "linux", "arch": "amd64"} + create_namespace pvc-auto-resize-25723 + local namespace=pvc-auto-resize-25723 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + grep -E -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' + awk '{print$1}' ++ mktemp + '[' -n '' ']' + desc 'cleaned up old namespaces pvc-auto-resize-25723' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pvc-auto-resize-25723 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pvc-auto-resize-25723 --ignore-not-found + xargs kubectl delete ns + local LAST_OUT=/tmp/tmp.oNDEwFjhD1 ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.xR8IpNGHqd + local exit_status=0 + local timeout=4 ++ seq 0 2 + local LAST_OUT=/tmp/tmp.1BKqmmE2hk + for i in $(seq 0 2) + set +e ++ mktemp + kubectl get ns + local LAST_ERR=/tmp/tmp.SBhdbSg6A8 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete namespace pvc-auto-resize-25723 --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.oNDEwFjhD1 + cat /tmp/tmp.xR8IpNGHqd + rm /tmp/tmp.oNDEwFjhD1 /tmp/tmp.xR8IpNGHqd + return 0 error: resource(s) were provided, but no name was specified + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.1BKqmmE2hk + cat /tmp/tmp.SBhdbSg6A8 + rm /tmp/tmp.1BKqmmE2hk /tmp/tmp.SBhdbSg6A8 + return 0 + kubectl_bin wait --for=delete namespace pvc-auto-resize-25723 ++ mktemp + local LAST_OUT=/tmp/tmp.4qF3Dy3gUs ++ mktemp + local LAST_ERR=/tmp/tmp.izuw1MbilZ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete namespace pvc-auto-resize-25723 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.4qF3Dy3gUs + cat /tmp/tmp.izuw1MbilZ + rm /tmp/tmp.4qF3Dy3gUs /tmp/tmp.izuw1MbilZ + return 0 + desc 'create namespace pvc-auto-resize-25723' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pvc-auto-resize-25723 ----------------------------------------------------------------------------------- + kubectl_bin create namespace pvc-auto-resize-25723 ++ mktemp + local LAST_OUT=/tmp/tmp.rbfsuavItW ++ mktemp + local LAST_ERR=/tmp/tmp.xQSlroQGIQ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace pvc-auto-resize-25723 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.rbfsuavItW namespace/pvc-auto-resize-25723 created + cat /tmp/tmp.xQSlroQGIQ + rm /tmp/tmp.rbfsuavItW /tmp/tmp.xQSlroQGIQ + return 0 + set_kube_ctx pvc-auto-resize-25723 + local namespace=pvc-auto-resize-25723 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.thQnt7yKDM +++ mktemp ++ local LAST_ERR=/tmp/tmp.UNIeSces5O ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.thQnt7yKDM ++ cat /tmp/tmp.UNIeSces5O ++ rm /tmp/tmp.thQnt7yKDM /tmp/tmp.UNIeSces5O ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2226-1e37b178-2-cluster7 --namespace=pvc-auto-resize-25723 ++ mktemp + local LAST_OUT=/tmp/tmp.OkzOWrUQA0 ++ mktemp + local LAST_ERR=/tmp/tmp.8tH55mUClQ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2226-1e37b178-2-cluster7 --namespace=pvc-auto-resize-25723 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.OkzOWrUQA0 Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2226-1e37b178-2-cluster7" modified. + cat /tmp/tmp.8tH55mUClQ + rm /tmp/tmp.OkzOWrUQA0 /tmp/tmp.8tH55mUClQ + return 0 + desc 'create secrets and psmdb client' + set +o xtrace ----------------------------------------------------------------------------------- create secrets and psmdb client ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/e2e-tests/conf/secrets.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.WO7BsqJ1cV ++ mktemp + local LAST_ERR=/tmp/tmp.dSvdjjpi5Y + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/e2e-tests/conf/secrets.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/e2e-tests/conf/client.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.WO7BsqJ1cV secret/some-users created deployment.apps/psmdb-client created + cat /tmp/tmp.dSvdjjpi5Y + rm /tmp/tmp.WO7BsqJ1cV /tmp/tmp.dSvdjjpi5Y + return 0 + desc 'create PSMDB cluster with storageAutoscaling enabled' + set +o xtrace ----------------------------------------------------------------------------------- create PSMDB cluster with storageAutoscaling enabled ----------------------------------------------------------------------------------- + cluster=some-name + spinup_psmdb some-name-rs0 /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/e2e-tests/pvc-auto-resize/conf/some-name.yml + local cluster=some-name-rs0 + local config=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/e2e-tests/pvc-auto-resize/conf/some-name.yml + local size=3 + desc 'create first PSMDB cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PSMDB cluster ----------------------------------------------------------------------------------- + apply_cluster /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/e2e-tests/pvc-auto-resize/conf/some-name.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/e2e-tests/pvc-auto-resize/conf/some-name.yml + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/e2e-tests/pvc-auto-resize/conf/some-name.yml + yq eval '(.spec | select(.image == null)).image = "perconalab/percona-server-mongodb-operator:main-mongod8.0"' ++ mktemp + yq eval '(.spec | select(has("pmm"))).pmm.image = "percona/pmm-client:2.44.1-1"' + yq eval '(.spec | select(has("initImage"))).initImage = "perconalab/percona-server-mongodb-operator:PR-2226-1e37b178"' + local LAST_OUT=/tmp/tmp.kjqQPJY8am ++ mktemp + /usr/sbin/sed -e s/NAME_SPACE/pvc-auto-resize-25723/g + yq eval '.spec.upgradeOptions.apply="Never"' + yq eval '(.spec | select(has("backup"))).backup.image = "perconalab/percona-server-mongodb-operator:main-backup"' + local LAST_ERR=/tmp/tmp.iVAg8JhmAl + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.kjqQPJY8am perconaservermongodb.psmdb.percona.com/some-name created + cat /tmp/tmp.iVAg8JhmAl + rm /tmp/tmp.kjqQPJY8am /tmp/tmp.iVAg8JhmAl + return 0 + desc 'check if Pod is started' + set +o xtrace ----------------------------------------------------------------------------------- check if Pod is started ----------------------------------------------------------------------------------- + wait_for_running some-name-rs0 3 + local name=some-name-rs0 + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=rs0 + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-rs0-0 + local pod=some-name-rs0-0 + set +o xtrace waiting for pod/some-name-rs0-0 to be ready.............OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-rs0-1 + local pod=some-name-rs0-1 + set +o xtrace waiting for pod/some-name-rs0-1 to be ready.............OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XhN6lt2OhF +++ mktemp ++ local LAST_ERR=/tmp/tmp.8HVTomBUbt ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.XhN6lt2OhF ++ cat /tmp/tmp.8HVTomBUbt ++ rm /tmp/tmp.XhN6lt2OhF /tmp/tmp.8HVTomBUbt ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-rs0-2 + local pod=some-name-rs0-2 + set +o xtrace waiting for pod/some-name-rs0-2 to be ready............OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0qsUH7n2dS +++ mktemp ++ local LAST_ERR=/tmp/tmp.sQekch4rIn ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.0qsUH7n2dS ++ cat /tmp/tmp.sQekch4rIn ++ rm /tmp/tmp.0qsUH7n2dS /tmp/tmp.sQekch4rIn ++ return 0 + [[ '' == \t\r\u\e ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jispG92VOr +++ mktemp ++ local LAST_ERR=/tmp/tmp.Dc932NlrvU ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jispG92VOr ++ cat /tmp/tmp.Dc932NlrvU ++ rm /tmp/tmp.jispG92VOr /tmp/tmp.Dc932NlrvU ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness + sleep 20 + compare_kubectl statefulset/some-name-rs0 + local resource=statefulset/some-name-rs0 + local postfix= + local skip_generation_check= + local expected_result=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/e2e-tests/pvc-auto-resize/compare/statefulset_some-name-rs0.yml + local new_result=/tmp/tmp.ltMfQO0SBH/statefulset_some-name-rs0.yml + '[' -n '' -a -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/e2e-tests/pvc-auto-resize/compare/statefulset_some-name-rs0-oc.yml ']' + kubectl_bin get -o yaml statefulset/some-name-rs0 + yq eval ' del(.metadata.ownerReferences[].apiVersion) | del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "NAMESPACE")) | del(.metadata.selfLink) | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | ++ mktemp del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.status) | (.. | select(tag == "!!str")) |= sub("pvc-auto-resize-25723", "NAME_SPACE") | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "extensions/v1beta1")) = "apps/v1" | (.. | select(. == "batch/v1beta1")) = "batch/v1" ' - + local LAST_OUT=/tmp/tmp.PnTxYE3O5T ++ mktemp + local LAST_ERR=/tmp/tmp.CqTIxgV7gk + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get -o yaml statefulset/some-name-rs0 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.PnTxYE3O5T + cat /tmp/tmp.CqTIxgV7gk + rm /tmp/tmp.PnTxYE3O5T /tmp/tmp.CqTIxgV7gk + return 0 + yq -i eval 'del(.spec.persistentVolumeClaimRetentionPolicy)' /tmp/tmp.ltMfQO0SBH/statefulset_some-name-rs0.yml + version_gt 1.22 ++ echo '1.32 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + yq -i eval 'del(.spec.internalTrafficPolicy)' /tmp/tmp.ltMfQO0SBH/statefulset_some-name-rs0.yml + yq -i eval 'del(.spec.allocateLoadBalancerNodePorts)' /tmp/tmp.ltMfQO0SBH/statefulset_some-name-rs0.yml + [[ /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/e2e-tests/pvc-auto-resize/compare/statefulset_some-name-rs0.yml == */cronjob* ]] + '[' -n '' ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/e2e-tests/pvc-auto-resize/compare/statefulset_some-name-rs0.yml /tmp/tmp.ltMfQO0SBH/statefulset_some-name-rs0.yml + log 'compare_kubectl: statefulset/some-name-rs0 OK' + set +o xtrace [2026-02-03T15:49:03+0000] compare_kubectl: statefulset/some-name-rs0 OK + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + run_mongo 'db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})' userAdmin:userAdmin123456@some-name-rs0.pvc-auto-resize-25723 + local 'command=db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})' + local uri=userAdmin:userAdmin123456@some-name-rs0.pvc-auto-resize-25723 + local driver=mongodb+srv + local suffix=.svc.cluster.local + local mongo_flag= + local replica_set=rs0 + [[ userAdmin:userAdmin123456@some-name-rs0.pvc-auto-resize-25723 == *cfg* ]] ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LTmMB7webL +++ mktemp ++ local LAST_ERR=/tmp/tmp.4A8v7O4Psz ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.LTmMB7webL ++ cat /tmp/tmp.4A8v7O4Psz ++ rm /tmp/tmp.LTmMB7webL /tmp/tmp.4A8v7O4Psz ++ return 0 + local client_container=psmdb-client-86cb5d8484-tw2dk + kubectl_bin exec psmdb-client-86cb5d8484-tw2dk -- bash -c 'printf '\''db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})\n'\'' | mongo mongodb+srv://userAdmin:userAdmin123456@some-name-rs0.pvc-auto-resize-25723.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.0sD89cr6z5 ++ mktemp + local LAST_ERR=/tmp/tmp.yuq9NwGByN + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-86cb5d8484-tw2dk -- bash -c 'printf '\''db.createUser({user: "myApp", pwd: "myPass", roles: [{ db: "myApp", role: "readWrite" }]})\n'\'' | mongo mongodb+srv://userAdmin:userAdmin123456@some-name-rs0.pvc-auto-resize-25723.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.0sD89cr6z5 Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-rs0-2.some-name-rs0.pvc-auto-resize-25723.svc.cluster.local:27017,some-name-rs0-1.some-name-rs0.pvc-auto-resize-25723.svc.cluster.local:27017,some-name-rs0-0.some-name-rs0.pvc-auto-resize-25723.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb&replicaSet=rs0&ssl=false Implicit session: session { "id" : UUID("c821b7eb-67b0-4d2c-8dc2-bc0a4b5c0e7d") } Percona Server for MongoDB server version: v8.0.17-6 WARNING: shell and server versions do not match Successfully added user: { "user" : "myApp", "roles" : [ { "db" : "myApp", "role" : "readWrite" } ] } bye + cat /tmp/tmp.yuq9NwGByN + rm /tmp/tmp.0sD89cr6z5 /tmp/tmp.yuq9NwGByN + return 0 + run_mongo 'use myApp\n db.test.insert({ x: 100500 })' myApp:myPass@some-name-rs0.pvc-auto-resize-25723 + local 'command=use myApp\n db.test.insert({ x: 100500 })' + local uri=myApp:myPass@some-name-rs0.pvc-auto-resize-25723 + local driver=mongodb+srv + local suffix=.svc.cluster.local + local mongo_flag= + local replica_set=rs0 + [[ myApp:myPass@some-name-rs0.pvc-auto-resize-25723 == *cfg* ]] ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.70nccTXBTs +++ mktemp ++ local LAST_ERR=/tmp/tmp.IwjMwCdneP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.70nccTXBTs ++ cat /tmp/tmp.IwjMwCdneP ++ rm /tmp/tmp.70nccTXBTs /tmp/tmp.IwjMwCdneP ++ return 0 + local client_container=psmdb-client-86cb5d8484-tw2dk + kubectl_bin exec psmdb-client-86cb5d8484-tw2dk -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb+srv://myApp:myPass@some-name-rs0.pvc-auto-resize-25723.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.Ek7AVSeLMI ++ mktemp + local LAST_ERR=/tmp/tmp.TX1whX7Zup + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-86cb5d8484-tw2dk -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb+srv://myApp:myPass@some-name-rs0.pvc-auto-resize-25723.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Ek7AVSeLMI Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-rs0-2.some-name-rs0.pvc-auto-resize-25723.svc.cluster.local:27017,some-name-rs0-1.some-name-rs0.pvc-auto-resize-25723.svc.cluster.local:27017,some-name-rs0-0.some-name-rs0.pvc-auto-resize-25723.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb&replicaSet=rs0&ssl=false Implicit session: session { "id" : UUID("a98fcaa2-744f-4afe-8196-178a4845340c") } Percona Server for MongoDB server version: v8.0.17-6 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.TX1whX7Zup + rm /tmp/tmp.Ek7AVSeLMI /tmp/tmp.TX1whX7Zup + return 0 + desc 'verify initial PVC size is 1Gi' + set +o xtrace ----------------------------------------------------------------------------------- verify initial PVC size is 1Gi ----------------------------------------------------------------------------------- ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.CEtSI4S0sT +++ mktemp ++ local LAST_ERR=/tmp/tmp.KCopVlJeqp ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.CEtSI4S0sT ++ cat /tmp/tmp.KCopVlJeqp ++ rm /tmp/tmp.CEtSI4S0sT /tmp/tmp.KCopVlJeqp ++ return 0 + initial_size=1Gi + echo 'Initial PVC size: 1Gi' Initial PVC size: 1Gi + [[ 1Gi != \1\G\i ]] + desc 'fill disk to trigger autoscaling threshold 80 percent' + set +o xtrace ----------------------------------------------------------------------------------- fill disk to trigger autoscaling threshold 80 percent ----------------------------------------------------------------------------------- + fill_disk_to_threshold some-name-rs0-0 80 850 + local pod=some-name-rs0-0 + local threshold=80 + local size_mb=850 + echo 'Filling disk on some-name-rs0-0 to approximately 80% (creating 850MB file)' Filling disk on some-name-rs0-0 to approximately 80% (creating 850MB file) + kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=850 2>/dev/null || true' ++ mktemp + local LAST_OUT=/tmp/tmp.p4cd9Aei7V ++ mktemp + local LAST_ERR=/tmp/tmp.xGCGWaWZJt + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=850 2>/dev/null || true' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.p4cd9Aei7V + cat /tmp/tmp.xGCGWaWZJt + rm /tmp/tmp.p4cd9Aei7V /tmp/tmp.xGCGWaWZJt + return 0 + local usage ++ kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sEWumDRPDk +++ mktemp ++ local LAST_ERR=/tmp/tmp.FAmv86sk52 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.sEWumDRPDk ++ cat /tmp/tmp.FAmv86sk52 ++ rm /tmp/tmp.sEWumDRPDk /tmp/tmp.FAmv86sk52 ++ return 0 + usage=100 + echo 'Current disk usage on some-name-rs0-0: 100%' Current disk usage on some-name-rs0-0: 100% + desc 'wait for auto-resize to trigger' + set +o xtrace ----------------------------------------------------------------------------------- wait for auto-resize to trigger ----------------------------------------------------------------------------------- + wait_for_auto_resize mongod-data-some-name-rs0-0 3Gi 60 10 + local pvc=mongod-data-some-name-rs0-0 + local expected_size=3Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 3Gi' Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 3Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zWTiLPtt33 +++ mktemp ++ local LAST_ERR=/tmp/tmp.FtNom8MX7T ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.zWTiLPtt33 ++ cat /tmp/tmp.FtNom8MX7T ++ rm /tmp/tmp.zWTiLPtt33 /tmp/tmp.FtNom8MX7T ++ return 0 + [[ 1Gi == \3\G\i ]] + [[ 0 -ge 60 ]] + echo -n . .+ sleep 10 + retry=1 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wuY9NEavU4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.a0QS8WvsK5 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.wuY9NEavU4 ++ cat /tmp/tmp.a0QS8WvsK5 ++ rm /tmp/tmp.wuY9NEavU4 /tmp/tmp.a0QS8WvsK5 ++ return 0 + [[ 1Gi == \3\G\i ]] + [[ 1 -ge 60 ]] + echo -n . .+ sleep 10 + retry=2 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mXRVDKP4EN +++ mktemp ++ local LAST_ERR=/tmp/tmp.nQvoiBcQt8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.mXRVDKP4EN ++ cat /tmp/tmp.nQvoiBcQt8 ++ rm /tmp/tmp.mXRVDKP4EN /tmp/tmp.nQvoiBcQt8 ++ return 0 + [[ 3Gi == \3\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-0 successfully resized to 3Gi' PVC mongod-data-some-name-rs0-0 successfully resized to 3Gi + wait_for_auto_resize mongod-data-some-name-rs0-1 3Gi 60 10 + local pvc=mongod-data-some-name-rs0-1 + local expected_size=3Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-1 to auto-resize to 3Gi' Waiting for PVC mongod-data-some-name-rs0-1 to auto-resize to 3Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SM1w0Uzag4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.AzWZvzdgFS ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.SM1w0Uzag4 ++ cat /tmp/tmp.AzWZvzdgFS ++ rm /tmp/tmp.SM1w0Uzag4 /tmp/tmp.AzWZvzdgFS ++ return 0 + [[ 1Gi == \3\G\i ]] + [[ 0 -ge 60 ]] + echo -n . .+ sleep 10 + retry=1 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dPWW7isvJd +++ mktemp ++ local LAST_ERR=/tmp/tmp.vw2IvZkI2T ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.dPWW7isvJd ++ cat /tmp/tmp.vw2IvZkI2T ++ rm /tmp/tmp.dPWW7isvJd /tmp/tmp.vw2IvZkI2T ++ return 0 + [[ 1Gi == \3\G\i ]] + [[ 1 -ge 60 ]] + echo -n . .+ sleep 10 + retry=2 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mzZO2qphmB +++ mktemp ++ local LAST_ERR=/tmp/tmp.z3Dhb6MQmi ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.mzZO2qphmB ++ cat /tmp/tmp.z3Dhb6MQmi ++ rm /tmp/tmp.mzZO2qphmB /tmp/tmp.z3Dhb6MQmi ++ return 0 + [[ 1Gi == \3\G\i ]] + [[ 2 -ge 60 ]] + echo -n . .+ sleep 10 + retry=3 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.roBCEr9sDb +++ mktemp ++ local LAST_ERR=/tmp/tmp.xRrN5DLtyI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.roBCEr9sDb ++ cat /tmp/tmp.xRrN5DLtyI ++ rm /tmp/tmp.roBCEr9sDb /tmp/tmp.xRrN5DLtyI ++ return 0 + [[ 1Gi == \3\G\i ]] + [[ 3 -ge 60 ]] + echo -n . .+ sleep 10 + retry=4 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZACfouEiSr +++ mktemp ++ local LAST_ERR=/tmp/tmp.lc4s1lNEDu ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ZACfouEiSr ++ cat /tmp/tmp.lc4s1lNEDu ++ rm /tmp/tmp.ZACfouEiSr /tmp/tmp.lc4s1lNEDu ++ return 0 + [[ 3Gi == \3\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-1 successfully resized to 3Gi' PVC mongod-data-some-name-rs0-1 successfully resized to 3Gi + wait_for_auto_resize mongod-data-some-name-rs0-2 3Gi 60 10 + local pvc=mongod-data-some-name-rs0-2 + local expected_size=3Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-2 to auto-resize to 3Gi' Waiting for PVC mongod-data-some-name-rs0-2 to auto-resize to 3Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-2 ++ local pvc=mongod-data-some-name-rs0-2 ++ kubectl_bin get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.shG2yYcAvy +++ mktemp ++ local LAST_ERR=/tmp/tmp.yv4UwcuXHp ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.shG2yYcAvy ++ cat /tmp/tmp.yv4UwcuXHp ++ rm /tmp/tmp.shG2yYcAvy /tmp/tmp.yv4UwcuXHp ++ return 0 + [[ 3Gi == \3\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-2 successfully resized to 3Gi' PVC mongod-data-some-name-rs0-2 successfully resized to 3Gi + desc 'verify autoscaling status is updated' + set +o xtrace ----------------------------------------------------------------------------------- verify autoscaling status is updated ----------------------------------------------------------------------------------- + check_autoscaling_status some-name 1 mongod-data-some-name-rs0-0 mongod-data-some-name-rs0-1 mongod-data-some-name-rs0-2 + local cluster=some-name + local expected_resize_count=1 + shift 2 + pvc_names=('mongod-data-some-name-rs0-0' 'mongod-data-some-name-rs0-1' 'mongod-data-some-name-rs0-2') + local pvc_names + local max_retry=30 + local sleep_time=10 + echo 'Checking autoscaling status for PVCs: mongod-data-some-name-rs0-0 mongod-data-some-name-rs0-1 mongod-data-some-name-rs0-2' Checking autoscaling status for PVCs: mongod-data-some-name-rs0-0 mongod-data-some-name-rs0-1 mongod-data-some-name-rs0-2 + echo 'Expected resize count per PVC: 1' Expected resize count per PVC: 1 + for pvc_name in "${pvc_names[@]}" + local retry=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-0.resizeCount}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5W4W5HW2TI +++ mktemp ++ local LAST_ERR=/tmp/tmp.ifYjj28SCU ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-0.resizeCount}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5W4W5HW2TI ++ cat /tmp/tmp.ifYjj28SCU ++ rm /tmp/tmp.5W4W5HW2TI /tmp/tmp.ifYjj28SCU ++ return 0 + [[ '' -eq 1 ]] + [[ 0 -ge 30 ]] + echo -n . .+ sleep 10 + retry=1 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-0.resizeCount}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jIbwmQv2Vh +++ mktemp ++ local LAST_ERR=/tmp/tmp.6eYDLRWOaU ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-0.resizeCount}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jIbwmQv2Vh ++ cat /tmp/tmp.6eYDLRWOaU ++ rm /tmp/tmp.jIbwmQv2Vh /tmp/tmp.6eYDLRWOaU ++ return 0 + [[ 1 -eq 1 ]] + for pvc_name in "${pvc_names[@]}" + local retry=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-1.resizeCount}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4cVVEujJ1G +++ mktemp ++ local LAST_ERR=/tmp/tmp.zQwKoQqBoS ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-1.resizeCount}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4cVVEujJ1G ++ cat /tmp/tmp.zQwKoQqBoS ++ rm /tmp/tmp.4cVVEujJ1G /tmp/tmp.zQwKoQqBoS ++ return 0 + [[ 1 -eq 1 ]] + for pvc_name in "${pvc_names[@]}" + local retry=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-2.resizeCount}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.e67P9CatPE +++ mktemp ++ local LAST_ERR=/tmp/tmp.m4AeWYKTgf ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-2.resizeCount}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.e67P9CatPE ++ cat /tmp/tmp.m4AeWYKTgf ++ rm /tmp/tmp.e67P9CatPE /tmp/tmp.m4AeWYKTgf ++ return 0 + [[ 1 -eq 1 ]] + desc 'wait for cluster consistency after resize' + set +o xtrace ----------------------------------------------------------------------------------- wait for cluster consistency after resize ----------------------------------------------------------------------------------- + wait_cluster_consistency some-name 64 + local cluster_name=some-name + local wait_time=64 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cbusx9D5Pu +++ mktemp ++ local LAST_ERR=/tmp/tmp.CpECJqzEgd ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.cbusx9D5Pu ++ cat /tmp/tmp.CpECJqzEgd ++ rm /tmp/tmp.cbusx9D5Pu /tmp/tmp.CpECJqzEgd ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo .OK .OK + desc 'verify cluster is ready' + set +o xtrace ----------------------------------------------------------------------------------- verify cluster is ready ----------------------------------------------------------------------------------- + grep -q ready + kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' ++ mktemp + local LAST_OUT=/tmp/tmp.baU2TPQyvJ ++ mktemp + local LAST_ERR=/tmp/tmp.1XaxfBe0WB + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get psmdb some-name -o 'jsonpath={.status.state}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.baU2TPQyvJ + cat /tmp/tmp.1XaxfBe0WB + rm /tmp/tmp.baU2TPQyvJ /tmp/tmp.1XaxfBe0WB + return 0 + desc 'test namespace resource quota enforcement' + set +o xtrace ----------------------------------------------------------------------------------- test namespace resource quota enforcement ----------------------------------------------------------------------------------- + apply_resourcequota 9Gi + local quota=9Gi + local default_sc ++ get_default_storageclass ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sE8MIOAHiT +++ mktemp ++ local LAST_ERR=/tmp/tmp.6S2cBrnmf2 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.sE8MIOAHiT ++ cat /tmp/tmp.6S2cBrnmf2 ++ rm /tmp/tmp.sE8MIOAHiT /tmp/tmp.6S2cBrnmf2 ++ return 0 + default_sc=standard-rwo + echo 'Applying resourcequota for default storageclass standard-rwo with quota 9Gi' Applying resourcequota for default storageclass standard-rwo with quota 9Gi + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/e2e-tests/pvc-auto-resize/conf/resourcequota.yml + sed s/STORAGECLASS/standard-rwo/ + sed s/QUOTA/9Gi/ + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.B7lBU4hRmM ++ mktemp + local LAST_ERR=/tmp/tmp.Kk2DQibUo7 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.B7lBU4hRmM resourcequota/default-storage-quota created + cat /tmp/tmp.Kk2DQibUo7 + rm /tmp/tmp.B7lBU4hRmM /tmp/tmp.Kk2DQibUo7 + return 0 + echo 'Filling disk again to trigger autoscaling with quota limit' Filling disk again to trigger autoscaling with quota limit + fill_disk_to_threshold some-name-rs0-0 80 2400 + local pod=some-name-rs0-0 + local threshold=80 + local size_mb=2400 + echo 'Filling disk on some-name-rs0-0 to approximately 80% (creating 2400MB file)' Filling disk on some-name-rs0-0 to approximately 80% (creating 2400MB file) + kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=2400 2>/dev/null || true' ++ mktemp + local LAST_OUT=/tmp/tmp.HLPq5i8rpY ++ mktemp + local LAST_ERR=/tmp/tmp.SunPgGWpy1 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=2400 2>/dev/null || true' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.HLPq5i8rpY + cat /tmp/tmp.SunPgGWpy1 + rm /tmp/tmp.HLPq5i8rpY /tmp/tmp.SunPgGWpy1 + return 0 + local usage ++ kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' +++ mktemp ++ local LAST_OUT=/tmp/tmp.U6nE71aLar +++ mktemp ++ local LAST_ERR=/tmp/tmp.WzSfMBfyFK ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.U6nE71aLar ++ cat /tmp/tmp.WzSfMBfyFK ++ rm /tmp/tmp.U6nE71aLar /tmp/tmp.WzSfMBfyFK ++ return 0 + usage=88 + echo 'Current disk usage on some-name-rs0-0: 88%' Current disk usage on some-name-rs0-0: 88% + echo 'Waiting to verify PVC does not resize beyond quota (should stay at 3Gi)' Waiting to verify PVC does not resize beyond quota (should stay at 3Gi) + sleep 30 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IeiLQKe6Sr +++ mktemp ++ local LAST_ERR=/tmp/tmp.UVBhelC1gf ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.IeiLQKe6Sr ++ cat /tmp/tmp.UVBhelC1gf ++ rm /tmp/tmp.IeiLQKe6Sr /tmp/tmp.UVBhelC1gf ++ return 0 + current_size=3Gi + [[ 3Gi != \3\G\i ]] + desc 'increase quota to allow further autoscaling' + set +o xtrace ----------------------------------------------------------------------------------- increase quota to allow further autoscaling ----------------------------------------------------------------------------------- + apply_resourcequota 15Gi + local quota=15Gi + local default_sc ++ get_default_storageclass ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AShoosDYpm +++ mktemp ++ local LAST_ERR=/tmp/tmp.pIEr3eTIbB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.AShoosDYpm ++ cat /tmp/tmp.pIEr3eTIbB ++ rm /tmp/tmp.AShoosDYpm /tmp/tmp.pIEr3eTIbB ++ return 0 + default_sc=standard-rwo + echo 'Applying resourcequota for default storageclass standard-rwo with quota 15Gi' Applying resourcequota for default storageclass standard-rwo with quota 15Gi + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/e2e-tests/pvc-auto-resize/conf/resourcequota.yml + sed s/STORAGECLASS/standard-rwo/ + kubectl_bin apply -f - + sed s/QUOTA/15Gi/ ++ mktemp + local LAST_OUT=/tmp/tmp.qCrzZcDKJ5 ++ mktemp + local LAST_ERR=/tmp/tmp.BmhfvVtJDf + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.qCrzZcDKJ5 resourcequota/default-storage-quota configured + cat /tmp/tmp.BmhfvVtJDf + rm /tmp/tmp.qCrzZcDKJ5 /tmp/tmp.BmhfvVtJDf + return 0 + wait_for_auto_resize mongod-data-some-name-rs0-0 5Gi 60 10 + local pvc=mongod-data-some-name-rs0-0 + local expected_size=5Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 5Gi' Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 5Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8nouBQASYu +++ mktemp ++ local LAST_ERR=/tmp/tmp.1KWCGjvJUK ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.8nouBQASYu ++ cat /tmp/tmp.1KWCGjvJUK ++ rm /tmp/tmp.8nouBQASYu /tmp/tmp.1KWCGjvJUK ++ return 0 + [[ 3Gi == \5\G\i ]] + [[ 0 -ge 60 ]] + echo -n . .+ sleep 10 + retry=1 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.U72juxmkrO +++ mktemp ++ local LAST_ERR=/tmp/tmp.P40X4oLBIz ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.U72juxmkrO ++ cat /tmp/tmp.P40X4oLBIz ++ rm /tmp/tmp.U72juxmkrO /tmp/tmp.P40X4oLBIz ++ return 0 + [[ 3Gi == \5\G\i ]] + [[ 1 -ge 60 ]] + echo -n . .+ sleep 10 + retry=2 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.duzqJBU943 +++ mktemp ++ local LAST_ERR=/tmp/tmp.bnbY7w2ZPm ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.duzqJBU943 ++ cat /tmp/tmp.bnbY7w2ZPm ++ rm /tmp/tmp.duzqJBU943 /tmp/tmp.bnbY7w2ZPm ++ return 0 + [[ 3Gi == \5\G\i ]] + [[ 2 -ge 60 ]] + echo -n . .+ sleep 10 + retry=3 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xPxy37MNxs +++ mktemp ++ local LAST_ERR=/tmp/tmp.IVv0ViWs5n ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xPxy37MNxs ++ cat /tmp/tmp.IVv0ViWs5n ++ rm /tmp/tmp.xPxy37MNxs /tmp/tmp.IVv0ViWs5n ++ return 0 + [[ 5Gi == \5\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-0 successfully resized to 5Gi' PVC mongod-data-some-name-rs0-0 successfully resized to 5Gi + wait_for_auto_resize mongod-data-some-name-rs0-1 5Gi 60 10 + local pvc=mongod-data-some-name-rs0-1 + local expected_size=5Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-1 to auto-resize to 5Gi' Waiting for PVC mongod-data-some-name-rs0-1 to auto-resize to 5Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.C0fcZF7Fek +++ mktemp ++ local LAST_ERR=/tmp/tmp.BY10gHl8YL ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.C0fcZF7Fek ++ cat /tmp/tmp.BY10gHl8YL ++ rm /tmp/tmp.C0fcZF7Fek /tmp/tmp.BY10gHl8YL ++ return 0 + [[ 3Gi == \5\G\i ]] + [[ 0 -ge 60 ]] + echo -n . .+ sleep 10 + retry=1 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5rQeAtvrJc +++ mktemp ++ local LAST_ERR=/tmp/tmp.IYh1PUEhgB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5rQeAtvrJc ++ cat /tmp/tmp.IYh1PUEhgB ++ rm /tmp/tmp.5rQeAtvrJc /tmp/tmp.IYh1PUEhgB ++ return 0 + [[ 3Gi == \5\G\i ]] + [[ 1 -ge 60 ]] + echo -n . .+ sleep 10 + retry=2 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SshBLnZIj6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.pKfgZMW9QW ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.SshBLnZIj6 ++ cat /tmp/tmp.pKfgZMW9QW ++ rm /tmp/tmp.SshBLnZIj6 /tmp/tmp.pKfgZMW9QW ++ return 0 + [[ 5Gi == \5\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-1 successfully resized to 5Gi' PVC mongod-data-some-name-rs0-1 successfully resized to 5Gi + wait_for_auto_resize mongod-data-some-name-rs0-2 5Gi 60 10 + local pvc=mongod-data-some-name-rs0-2 + local expected_size=5Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-2 to auto-resize to 5Gi' Waiting for PVC mongod-data-some-name-rs0-2 to auto-resize to 5Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-2 ++ local pvc=mongod-data-some-name-rs0-2 ++ kubectl_bin get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NLgGpIlevc +++ mktemp ++ local LAST_ERR=/tmp/tmp.A7FW0NlEeU ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.NLgGpIlevc ++ cat /tmp/tmp.A7FW0NlEeU ++ rm /tmp/tmp.NLgGpIlevc /tmp/tmp.A7FW0NlEeU ++ return 0 + [[ 5Gi == \5\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-2 successfully resized to 5Gi' PVC mongod-data-some-name-rs0-2 successfully resized to 5Gi + desc 'verify cluster consistency after quota test' + set +o xtrace ----------------------------------------------------------------------------------- verify cluster consistency after quota test ----------------------------------------------------------------------------------- + wait_cluster_consistency some-name 64 + local cluster_name=some-name + local wait_time=64 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.A4XsRKLLZc +++ mktemp ++ local LAST_ERR=/tmp/tmp.s454jwJ1Ln ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.A4XsRKLLZc ++ cat /tmp/tmp.s454jwJ1Ln ++ rm /tmp/tmp.A4XsRKLLZc /tmp/tmp.s454jwJ1Ln ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 1 -ge 64 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YCVXKTBIOq +++ mktemp ++ local LAST_ERR=/tmp/tmp.7TdZvPK5hp ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.YCVXKTBIOq ++ cat /tmp/tmp.7TdZvPK5hp ++ rm /tmp/tmp.YCVXKTBIOq /tmp/tmp.7TdZvPK5hp ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo .OK .OK + desc 'test maxSize limit' + set +o xtrace ----------------------------------------------------------------------------------- test maxSize limit ----------------------------------------------------------------------------------- + apply_resourcequota 30Gi + local quota=30Gi + local default_sc ++ get_default_storageclass ++ kubectl_bin get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xfJe1oY0BH +++ mktemp ++ local LAST_ERR=/tmp/tmp.RqSO8MQwk6 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sc -o 'jsonpath={.items[?(@.metadata.annotations.storageclass\.kubernetes\.io/is-default-class=="true")].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xfJe1oY0BH ++ cat /tmp/tmp.RqSO8MQwk6 ++ rm /tmp/tmp.xfJe1oY0BH /tmp/tmp.RqSO8MQwk6 ++ return 0 + default_sc=standard-rwo + echo 'Applying resourcequota for default storageclass standard-rwo with quota 30Gi' Applying resourcequota for default storageclass standard-rwo with quota 30Gi + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/e2e-tests/pvc-auto-resize/conf/resourcequota.yml + sed s/STORAGECLASS/standard-rwo/ + sed s/QUOTA/30Gi/ + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.HZ2WL4wH4r ++ mktemp + local LAST_ERR=/tmp/tmp.sCU1AWC6h5 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.HZ2WL4wH4r resourcequota/default-storage-quota configured + cat /tmp/tmp.sCU1AWC6h5 + rm /tmp/tmp.HZ2WL4wH4r /tmp/tmp.sCU1AWC6h5 + return 0 + echo 'Continuing to fill disk to test maxSize enforcement' Continuing to fill disk to test maxSize enforcement + fill_disk_to_threshold some-name-rs0-0 80 4000 + local pod=some-name-rs0-0 + local threshold=80 + local size_mb=4000 + echo 'Filling disk on some-name-rs0-0 to approximately 80% (creating 4000MB file)' Filling disk on some-name-rs0-0 to approximately 80% (creating 4000MB file) + kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=4000 2>/dev/null || true' ++ mktemp + local LAST_OUT=/tmp/tmp.neBeZ7RCp8 ++ mktemp + local LAST_ERR=/tmp/tmp.iMBMBAg2pD + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=4000 2>/dev/null || true' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.neBeZ7RCp8 + cat /tmp/tmp.iMBMBAg2pD + rm /tmp/tmp.neBeZ7RCp8 /tmp/tmp.iMBMBAg2pD + return 0 + local usage ++ kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' +++ mktemp ++ local LAST_OUT=/tmp/tmp.m0GLgepJBw +++ mktemp ++ local LAST_ERR=/tmp/tmp.kgJfBIfUh3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.m0GLgepJBw ++ cat /tmp/tmp.kgJfBIfUh3 ++ rm /tmp/tmp.m0GLgepJBw /tmp/tmp.kgJfBIfUh3 ++ return 0 + usage=85 + echo 'Current disk usage on some-name-rs0-0: 85%' Current disk usage on some-name-rs0-0: 85% + wait_for_auto_resize mongod-data-some-name-rs0-0 7Gi 60 10 + local pvc=mongod-data-some-name-rs0-0 + local expected_size=7Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 7Gi' Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 7Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kyN8fSxzxU +++ mktemp ++ local LAST_ERR=/tmp/tmp.UMsv6kaYBA ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.kyN8fSxzxU ++ cat /tmp/tmp.UMsv6kaYBA ++ rm /tmp/tmp.kyN8fSxzxU /tmp/tmp.UMsv6kaYBA ++ return 0 + [[ 5Gi == \7\G\i ]] + [[ 0 -ge 60 ]] + echo -n . .+ sleep 10 + retry=1 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.S0p5vCf76K +++ mktemp ++ local LAST_ERR=/tmp/tmp.NNGKSrCNDw ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.S0p5vCf76K ++ cat /tmp/tmp.NNGKSrCNDw ++ rm /tmp/tmp.S0p5vCf76K /tmp/tmp.NNGKSrCNDw ++ return 0 + [[ 5Gi == \7\G\i ]] + [[ 1 -ge 60 ]] + echo -n . .+ sleep 10 + retry=2 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oBZXgoCQM7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.YpKSzb7MZB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.oBZXgoCQM7 ++ cat /tmp/tmp.YpKSzb7MZB ++ rm /tmp/tmp.oBZXgoCQM7 /tmp/tmp.YpKSzb7MZB ++ return 0 + [[ 5Gi == \7\G\i ]] + [[ 2 -ge 60 ]] + echo -n . .+ sleep 10 + retry=3 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.iNITV3AFa1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.LErO6ogDPE ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.iNITV3AFa1 ++ cat /tmp/tmp.LErO6ogDPE ++ rm /tmp/tmp.iNITV3AFa1 /tmp/tmp.LErO6ogDPE ++ return 0 + [[ 5Gi == \7\G\i ]] + [[ 3 -ge 60 ]] + echo -n . .+ sleep 10 + retry=4 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Dg5tKC7IlT +++ mktemp ++ local LAST_ERR=/tmp/tmp.mKI8RcvgBP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Dg5tKC7IlT ++ cat /tmp/tmp.mKI8RcvgBP ++ rm /tmp/tmp.Dg5tKC7IlT /tmp/tmp.mKI8RcvgBP ++ return 0 + [[ 5Gi == \7\G\i ]] + [[ 4 -ge 60 ]] + echo -n . .+ sleep 10 + retry=5 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1ZHkVvVazs +++ mktemp ++ local LAST_ERR=/tmp/tmp.3GZMl0DWLO ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.1ZHkVvVazs ++ cat /tmp/tmp.3GZMl0DWLO ++ rm /tmp/tmp.1ZHkVvVazs /tmp/tmp.3GZMl0DWLO ++ return 0 + [[ 5Gi == \7\G\i ]] + [[ 5 -ge 60 ]] + echo -n . .+ sleep 10 + retry=6 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Njpcw73Tmw +++ mktemp ++ local LAST_ERR=/tmp/tmp.lLfciCTNIt ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Njpcw73Tmw ++ cat /tmp/tmp.lLfciCTNIt ++ rm /tmp/tmp.Njpcw73Tmw /tmp/tmp.lLfciCTNIt ++ return 0 + [[ 5Gi == \7\G\i ]] + [[ 6 -ge 60 ]] + echo -n . .+ sleep 10 + retry=7 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IZtVqN1lqq +++ mktemp ++ local LAST_ERR=/tmp/tmp.V8Rzvgmdfl ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.IZtVqN1lqq ++ cat /tmp/tmp.V8Rzvgmdfl ++ rm /tmp/tmp.IZtVqN1lqq /tmp/tmp.V8Rzvgmdfl ++ return 0 + [[ 7Gi == \7\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-0 successfully resized to 7Gi' PVC mongod-data-some-name-rs0-0 successfully resized to 7Gi + fill_disk_to_threshold some-name-rs0-0 80 5600 + local pod=some-name-rs0-0 + local threshold=80 + local size_mb=5600 + echo 'Filling disk on some-name-rs0-0 to approximately 80% (creating 5600MB file)' Filling disk on some-name-rs0-0 to approximately 80% (creating 5600MB file) + kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=5600 2>/dev/null || true' ++ mktemp + local LAST_OUT=/tmp/tmp.hLEwQPv2Da ++ mktemp + local LAST_ERR=/tmp/tmp.fdXwPWvx61 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=5600 2>/dev/null || true' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.hLEwQPv2Da + cat /tmp/tmp.fdXwPWvx61 + rm /tmp/tmp.hLEwQPv2Da /tmp/tmp.fdXwPWvx61 + return 0 + local usage ++ kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' +++ mktemp ++ local LAST_OUT=/tmp/tmp.m7c8Djc6E5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.TG74YPOwnG ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.m7c8Djc6E5 ++ cat /tmp/tmp.TG74YPOwnG ++ rm /tmp/tmp.m7c8Djc6E5 /tmp/tmp.TG74YPOwnG ++ return 0 + usage=83 + echo 'Current disk usage on some-name-rs0-0: 83%' Current disk usage on some-name-rs0-0: 83% + wait_for_auto_resize mongod-data-some-name-rs0-0 9Gi 60 10 + local pvc=mongod-data-some-name-rs0-0 + local expected_size=9Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 9Gi' Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 9Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SeFmxIw3wB +++ mktemp ++ local LAST_ERR=/tmp/tmp.dpJoyrI0iy ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.SeFmxIw3wB ++ cat /tmp/tmp.dpJoyrI0iy ++ rm /tmp/tmp.SeFmxIw3wB /tmp/tmp.dpJoyrI0iy ++ return 0 + [[ 7Gi == \9\G\i ]] + [[ 0 -ge 60 ]] + echo -n . .+ sleep 10 + retry=1 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Vm0HYI7Odf +++ mktemp ++ local LAST_ERR=/tmp/tmp.T1UdrVHdTz ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Vm0HYI7Odf ++ cat /tmp/tmp.T1UdrVHdTz ++ rm /tmp/tmp.Vm0HYI7Odf /tmp/tmp.T1UdrVHdTz ++ return 0 + [[ 7Gi == \9\G\i ]] + [[ 1 -ge 60 ]] + echo -n . .+ sleep 10 + retry=2 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jknFKW4HWA +++ mktemp ++ local LAST_ERR=/tmp/tmp.XoMXqW0C2j ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jknFKW4HWA ++ cat /tmp/tmp.XoMXqW0C2j ++ rm /tmp/tmp.jknFKW4HWA /tmp/tmp.XoMXqW0C2j ++ return 0 + [[ 7Gi == \9\G\i ]] + [[ 2 -ge 60 ]] + echo -n . .+ sleep 10 + retry=3 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KB1QsxNfSX +++ mktemp ++ local LAST_ERR=/tmp/tmp.TILLfc6M8h ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.KB1QsxNfSX ++ cat /tmp/tmp.TILLfc6M8h ++ rm /tmp/tmp.KB1QsxNfSX /tmp/tmp.TILLfc6M8h ++ return 0 + [[ 7Gi == \9\G\i ]] + [[ 3 -ge 60 ]] + echo -n . .+ sleep 10 + retry=4 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OZMw8DDqbr +++ mktemp ++ local LAST_ERR=/tmp/tmp.Arn62Af4XC ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.OZMw8DDqbr ++ cat /tmp/tmp.Arn62Af4XC ++ rm /tmp/tmp.OZMw8DDqbr /tmp/tmp.Arn62Af4XC ++ return 0 + [[ 9Gi == \9\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-0 successfully resized to 9Gi' PVC mongod-data-some-name-rs0-0 successfully resized to 9Gi + fill_disk_to_threshold some-name-rs0-0 80 7200 + local pod=some-name-rs0-0 + local threshold=80 + local size_mb=7200 + echo 'Filling disk on some-name-rs0-0 to approximately 80% (creating 7200MB file)' Filling disk on some-name-rs0-0 to approximately 80% (creating 7200MB file) + kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=7200 2>/dev/null || true' ++ mktemp + local LAST_OUT=/tmp/tmp.BEum9xOnOq ++ mktemp + local LAST_ERR=/tmp/tmp.cFO5jndFHT + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec some-name-rs0-0 -c mongod -- bash -c 'dd if=/dev/zero of=/data/db/fillfile bs=1M count=7200 2>/dev/null || true' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.BEum9xOnOq + cat /tmp/tmp.cFO5jndFHT + rm /tmp/tmp.BEum9xOnOq /tmp/tmp.cFO5jndFHT + return 0 + local usage ++ kubectl_bin exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' +++ mktemp ++ local LAST_OUT=/tmp/tmp.InM1lT2WJI +++ mktemp ++ local LAST_ERR=/tmp/tmp.cgCys5Tqog ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl exec some-name-rs0-0 -c mongod -- bash -c 'df /data/db | tail -1 | awk '\''{print $5}'\'' | sed '\''s/%//'\''' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.InM1lT2WJI ++ cat /tmp/tmp.cgCys5Tqog ++ rm /tmp/tmp.InM1lT2WJI /tmp/tmp.cgCys5Tqog ++ return 0 + usage=83 + echo 'Current disk usage on some-name-rs0-0: 83%' Current disk usage on some-name-rs0-0: 83% ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HO0Ch6DuUL +++ mktemp ++ local LAST_ERR=/tmp/tmp.zNi08qX32I ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.HO0Ch6DuUL ++ cat /tmp/tmp.zNi08qX32I ++ rm /tmp/tmp.HO0Ch6DuUL /tmp/tmp.zNi08qX32I ++ return 0 + final_size=9Gi + [[ 9Gi == \1\0\G\i ]] + [[ 9Gi == \9\G\i ]] + echo 'PVC is at 9Gi, will resize to 10Gi (maxSize)' PVC is at 9Gi, will resize to 10Gi (maxSize) + wait_for_auto_resize mongod-data-some-name-rs0-0 10Gi 60 10 + local pvc=mongod-data-some-name-rs0-0 + local expected_size=10Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 10Gi' Waiting for PVC mongod-data-some-name-rs0-0 to auto-resize to 10Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zF3mbrXm3n +++ mktemp ++ local LAST_ERR=/tmp/tmp.TU6ZOqdGPp ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.zF3mbrXm3n ++ cat /tmp/tmp.TU6ZOqdGPp ++ rm /tmp/tmp.zF3mbrXm3n /tmp/tmp.TU6ZOqdGPp ++ return 0 + [[ 9Gi == \1\0\G\i ]] + [[ 0 -ge 60 ]] + echo -n . .+ sleep 10 + retry=1 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.z2Jk1nWJdI +++ mktemp ++ local LAST_ERR=/tmp/tmp.84uyl5B9hf ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.z2Jk1nWJdI ++ cat /tmp/tmp.84uyl5B9hf ++ rm /tmp/tmp.z2Jk1nWJdI /tmp/tmp.84uyl5B9hf ++ return 0 + [[ 9Gi == \1\0\G\i ]] + [[ 1 -ge 60 ]] + echo -n . .+ sleep 10 + retry=2 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Adn86VuJb4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.BF0XQPu6EZ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Adn86VuJb4 ++ cat /tmp/tmp.BF0XQPu6EZ ++ rm /tmp/tmp.Adn86VuJb4 /tmp/tmp.BF0XQPu6EZ ++ return 0 + [[ 9Gi == \1\0\G\i ]] + [[ 2 -ge 60 ]] + echo -n . .+ sleep 10 + retry=3 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kw3pxlV4DN +++ mktemp ++ local LAST_ERR=/tmp/tmp.SCtoqFwlMm ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.kw3pxlV4DN ++ cat /tmp/tmp.SCtoqFwlMm ++ rm /tmp/tmp.kw3pxlV4DN /tmp/tmp.SCtoqFwlMm ++ return 0 + [[ 9Gi == \1\0\G\i ]] + [[ 3 -ge 60 ]] + echo -n . .+ sleep 10 + retry=4 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.epyRbRKjAW +++ mktemp ++ local LAST_ERR=/tmp/tmp.TehxaBkoHj ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.epyRbRKjAW ++ cat /tmp/tmp.TehxaBkoHj ++ rm /tmp/tmp.epyRbRKjAW /tmp/tmp.TehxaBkoHj ++ return 0 + [[ 9Gi == \1\0\G\i ]] + [[ 4 -ge 60 ]] + echo -n . .+ sleep 10 + retry=5 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HWasXBG7bw +++ mktemp ++ local LAST_ERR=/tmp/tmp.wAaF4FoSjz ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.HWasXBG7bw ++ cat /tmp/tmp.wAaF4FoSjz ++ rm /tmp/tmp.HWasXBG7bw /tmp/tmp.wAaF4FoSjz ++ return 0 + [[ 9Gi == \1\0\G\i ]] + [[ 5 -ge 60 ]] + echo -n . .+ sleep 10 + retry=6 ++ get_pvc_size mongod-data-some-name-rs0-0 ++ local pvc=mongod-data-some-name-rs0-0 ++ kubectl_bin get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9qNzOIwVBr +++ mktemp ++ local LAST_ERR=/tmp/tmp.l1p6Ksczam ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-0 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9qNzOIwVBr ++ cat /tmp/tmp.l1p6Ksczam ++ rm /tmp/tmp.9qNzOIwVBr /tmp/tmp.l1p6Ksczam ++ return 0 + [[ 10Gi == \1\0\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-0 successfully resized to 10Gi' PVC mongod-data-some-name-rs0-0 successfully resized to 10Gi + wait_for_auto_resize mongod-data-some-name-rs0-1 10Gi 60 10 + local pvc=mongod-data-some-name-rs0-1 + local expected_size=10Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-1 to auto-resize to 10Gi' Waiting for PVC mongod-data-some-name-rs0-1 to auto-resize to 10Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-1 ++ local pvc=mongod-data-some-name-rs0-1 ++ kubectl_bin get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.byCNOkDvNT +++ mktemp ++ local LAST_ERR=/tmp/tmp.h3X0mdaTPs ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-1 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.byCNOkDvNT ++ cat /tmp/tmp.h3X0mdaTPs ++ rm /tmp/tmp.byCNOkDvNT /tmp/tmp.h3X0mdaTPs ++ return 0 + [[ 10Gi == \1\0\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-1 successfully resized to 10Gi' PVC mongod-data-some-name-rs0-1 successfully resized to 10Gi + wait_for_auto_resize mongod-data-some-name-rs0-2 10Gi 60 10 + local pvc=mongod-data-some-name-rs0-2 + local expected_size=10Gi + local max_retry=60 + local sleep_time=10 + echo 'Waiting for PVC mongod-data-some-name-rs0-2 to auto-resize to 10Gi' Waiting for PVC mongod-data-some-name-rs0-2 to auto-resize to 10Gi + local retry=0 ++ get_pvc_size mongod-data-some-name-rs0-2 ++ local pvc=mongod-data-some-name-rs0-2 ++ kubectl_bin get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.clDwdc9cft +++ mktemp ++ local LAST_ERR=/tmp/tmp.elOUoz4cNz ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pvc mongod-data-some-name-rs0-2 -o 'jsonpath={.status.capacity.storage}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.clDwdc9cft ++ cat /tmp/tmp.elOUoz4cNz ++ rm /tmp/tmp.clDwdc9cft /tmp/tmp.elOUoz4cNz ++ return 0 + [[ 10Gi == \1\0\G\i ]] + echo + echo 'PVC mongod-data-some-name-rs0-2 successfully resized to 10Gi' PVC mongod-data-some-name-rs0-2 successfully resized to 10Gi + desc 'verify final autoscaling status' + set +o xtrace ----------------------------------------------------------------------------------- verify final autoscaling status ----------------------------------------------------------------------------------- + check_autoscaling_status some-name 4 mongod-data-some-name-rs0-0 mongod-data-some-name-rs0-1 mongod-data-some-name-rs0-2 + local cluster=some-name + local expected_resize_count=4 + shift 2 + pvc_names=('mongod-data-some-name-rs0-0' 'mongod-data-some-name-rs0-1' 'mongod-data-some-name-rs0-2') + local pvc_names + local max_retry=30 + local sleep_time=10 + echo 'Checking autoscaling status for PVCs: mongod-data-some-name-rs0-0 mongod-data-some-name-rs0-1 mongod-data-some-name-rs0-2' Checking autoscaling status for PVCs: mongod-data-some-name-rs0-0 mongod-data-some-name-rs0-1 mongod-data-some-name-rs0-2 + echo 'Expected resize count per PVC: 4' Expected resize count per PVC: 4 + for pvc_name in "${pvc_names[@]}" + local retry=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-0.resizeCount}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.L3u5gupbeG +++ mktemp ++ local LAST_ERR=/tmp/tmp.Be3B7rSdG8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-0.resizeCount}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.L3u5gupbeG ++ cat /tmp/tmp.Be3B7rSdG8 ++ rm /tmp/tmp.L3u5gupbeG /tmp/tmp.Be3B7rSdG8 ++ return 0 + [[ 4 -eq 4 ]] + for pvc_name in "${pvc_names[@]}" + local retry=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-1.resizeCount}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BrdCTAWkfr +++ mktemp ++ local LAST_ERR=/tmp/tmp.BTZNr0kCEh ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-1.resizeCount}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.BrdCTAWkfr ++ cat /tmp/tmp.BTZNr0kCEh ++ rm /tmp/tmp.BrdCTAWkfr /tmp/tmp.BTZNr0kCEh ++ return 0 + [[ 4 -eq 4 ]] + for pvc_name in "${pvc_names[@]}" + local retry=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-2.resizeCount}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GxShrTiF47 +++ mktemp ++ local LAST_ERR=/tmp/tmp.OmAkSmZMOP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.storageAutoscaling.mongod-data-some-name-rs0-2.resizeCount}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.GxShrTiF47 ++ cat /tmp/tmp.OmAkSmZMOP ++ rm /tmp/tmp.GxShrTiF47 /tmp/tmp.OmAkSmZMOP ++ return 0 + [[ 4 -eq 4 ]] + wait_cluster_consistency some-name 64 + local cluster_name=some-name + local wait_time=64 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8FvjBUI4lc +++ mktemp ++ local LAST_ERR=/tmp/tmp.PIwdoQNSJ9 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.8FvjBUI4lc ++ cat /tmp/tmp.PIwdoQNSJ9 ++ rm /tmp/tmp.8FvjBUI4lc /tmp/tmp.PIwdoQNSJ9 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 1 -ge 64 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OfA1locSc5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.oMc9dqGXDX ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.OfA1locSc5 ++ cat /tmp/tmp.oMc9dqGXDX ++ rm /tmp/tmp.OfA1locSc5 /tmp/tmp.oMc9dqGXDX ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo .OK .OK + destroy pvc-auto-resize-25723 + local namespace=pvc-auto-resize-25723 + local ignore_logs=true + [[ 0 == 1 ]] + desc 'destroy cluster/operator and all other resources' + set +o xtrace ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + '[' true == false ']' + delete_backups + desc 'Delete psmdb-backup' + set +o xtrace ----------------------------------------------------------------------------------- Delete psmdb-backup ----------------------------------------------------------------------------------- ++ kubectl_bin get psmdb-backup --no-headers ++ wc -l +++ mktemp ++ local LAST_OUT=/tmp/tmp.9JECi8CzRn +++ mktemp ++ local LAST_ERR=/tmp/tmp.fGEz0yUDFW ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb-backup --no-headers ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9JECi8CzRn ++ cat /tmp/tmp.fGEz0yUDFW No resources found in pvc-auto-resize-25723 namespace. ++ rm /tmp/tmp.9JECi8CzRn /tmp/tmp.fGEz0yUDFW ++ return 0 + '[' 0 '!=' 0 ']' + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.F3E0M3xLp5 ++ mktemp + local LAST_ERR=/tmp/tmp.c6bYFvm1SG + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.F3E0M3xLp5 customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.c6bYFvm1SG + rm /tmp/tmp.F3E0M3xLp5 /tmp/tmp.c6bYFvm1SG + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/deploy/crd.yaml ++ grep -v '\-\-\-' grep: warning: stray \ before - grep: warning: stray \ before - + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.OKxOeWersX ++ mktemp + local LAST_ERR=/tmp/tmp.pIGdvAlrGj + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.OKxOeWersX + cat /tmp/tmp.pIGdvAlrGj + rm /tmp/tmp.OKxOeWersX /tmp/tmp.pIGdvAlrGj + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.UXLUIWJMtd ++ mktemp + local LAST_ERR=/tmp/tmp.GXe0WPdhfG + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.UXLUIWJMtd + cat /tmp/tmp.GXe0WPdhfG + rm /tmp/tmp.UXLUIWJMtd /tmp/tmp.GXe0WPdhfG + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch perconaservermongodbs.psmdb.percona.com -n pvc-auto-resize-25723 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaservermongodb.psmdb.percona.com/some-name patched + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.AN3anjQqF6 ++ mktemp + local LAST_ERR=/tmp/tmp.J2CBLVpGMY + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.AN3anjQqF6 customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com condition met + cat /tmp/tmp.J2CBLVpGMY + rm /tmp/tmp.AN3anjQqF6 /tmp/tmp.J2CBLVpGMY + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.KpinIaJv3P ++ mktemp + local LAST_ERR=/tmp/tmp.jMP476HFNW + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2226/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.KpinIaJv3P clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.jMP476HFNW + rm /tmp/tmp.KpinIaJv3P /tmp/tmp.jMP476HFNW + return 0 + destroy_cert_manager + kubectl_bin delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.xlUOFoz6G3 ++ mktemp + local LAST_ERR=/tmp/tmp.AmSIQSIgpy + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.xlUOFoz6G3 + cat /tmp/tmp.AmSIQSIgpy Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 0 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.xlUOFoz6G3 + cat /tmp/tmp.AmSIQSIgpy Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 4 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.xlUOFoz6G3 + cat /tmp/tmp.AmSIQSIgpy Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 8 + cat /tmp/tmp.xlUOFoz6G3 + cat /tmp/tmp.AmSIQSIgpy Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.1/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + rm /tmp/tmp.xlUOFoz6G3 /tmp/tmp.AmSIQSIgpy + return 1 + true + '[' -n '' ']' + '[' -n psmdb-operator ']' + rm -rf /tmp/tmp.ltMfQO0SBH + kubectl_bin delete --grace-period=0 --force=true namespace pvc-auto-resize-25723 + kubectl_bin delete --grace-period=0 --force=true namespace psmdb-operator ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.x2PFARF4rr + desc 'test passed' + set +o xtrace ----------------------------------------------------------------------------------- test passed ----------------------------------------------------------------------------------- ++ mktemp + local LAST_OUT=/tmp/tmp.DYDGP6kXyw ++ mktemp + local LAST_ERR=/tmp/tmp.RcpeHB92Fx + local exit_status=0 + local timeout=4 + local LAST_ERR=/tmp/tmp.Kp6KRAzCv2 + local exit_status=0 + local timeout=4 ++ seq 0 2 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete --grace-period=0 --force=true namespace pvc-auto-resize-25723 + for i in $(seq 0 2) + set +e + kubectl delete --grace-period=0 --force=true namespace psmdb-operator