Log: /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/logs/data-sharded.log Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 + main + [[ docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0 == *\p\e\r\c\o\n\a\-\s\e\r\v\e\r\-\m\o\n\g\o\d\b\-\o\p\e\r\a\t\o\r* ]] ++ echo -n docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0 ++ /usr/sbin/sed -r 's/.*([0-9].[0-9])$/\1/' + MONGO_VER=8.0 + create_infra data-sharded-19777 + local ns=data-sharded-19777 + [[ 1 == 1 ]] + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.7eeerEZ47P ++ mktemp + local LAST_ERR=/tmp/tmp.2YaeweHwNt + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.7eeerEZ47P customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.2YaeweHwNt + rm /tmp/tmp.7eeerEZ47P /tmp/tmp.2YaeweHwNt + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/crd.yaml ++ grep -v '\-\-\-' grep: warning: stray \ before - grep: warning: stray \ before - + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.4pY3clVfcq ++ mktemp + local LAST_ERR=/tmp/tmp.uWmgPf0vmD + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.4pY3clVfcq + cat /tmp/tmp.uWmgPf0vmD + rm /tmp/tmp.4pY3clVfcq /tmp/tmp.uWmgPf0vmD + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.wCxXms0NWE ++ mktemp + local LAST_ERR=/tmp/tmp.TInnWuWijF + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.wCxXms0NWE + cat /tmp/tmp.TInnWuWijF + rm /tmp/tmp.wCxXms0NWE /tmp/tmp.TInnWuWijF + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbs" + kubectl patch perconaservermongodbs.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbs" + : + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.L69mlQkqZY ++ mktemp + local LAST_ERR=/tmp/tmp.69yMtfbC5Q + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.L69mlQkqZY + cat /tmp/tmp.69yMtfbC5Q + rm /tmp/tmp.L69mlQkqZY /tmp/tmp.69yMtfbC5Q + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.7DowB1YKVZ ++ mktemp + local LAST_ERR=/tmp/tmp.yRI2PVaLXL + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.7DowB1YKVZ clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.yRI2PVaLXL + rm /tmp/tmp.7DowB1YKVZ /tmp/tmp.yRI2PVaLXL + return 0 + check_crd_for_deletion PR-2219-d7e802db + local git_tag=PR-2219-d7e802db ++ curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/PR-2219-d7e802db/deploy/crd.yaml ++ yq eval .metadata.name ++ /usr/sbin/sed s/---//g ++ /usr/sbin/sed ':a;N;$!ba;s/\n/ /g' + for crd_name in $(curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/${git_tag}/deploy/crd.yaml | yq eval '.metadata.name' | $sed 's/---//g' | $sed ':a;N;$!ba;s/\n/ /g') ++ kubectl_bin get crd/null -o 'jsonpath={.status.conditions[-1].type}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XcMuluuzpe +++ mktemp ++ local LAST_ERR=/tmp/tmp.FVgCaJWYi9 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.XcMuluuzpe ++ cat /tmp/tmp.FVgCaJWYi9 Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 0 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.XcMuluuzpe ++ cat /tmp/tmp.FVgCaJWYi9 Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 4 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.XcMuluuzpe ++ cat /tmp/tmp.FVgCaJWYi9 Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 8 ++ cat /tmp/tmp.XcMuluuzpe ++ cat /tmp/tmp.FVgCaJWYi9 Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ rm /tmp/tmp.XcMuluuzpe /tmp/tmp.FVgCaJWYi9 ++ return 1 + [[ '' == \T\e\r\m\i\n\a\t\i\n\g ]] + '[' -n psmdb-operator ']' + create_namespace psmdb-operator + local namespace=psmdb-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ kubectl api-resources ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + grep -E -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' + '[' -n '' ']' + desc 'cleaned up old namespaces psmdb-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace psmdb-operator --ignore-not-found ++ mktemp + awk '{print$1}' ++ mktemp + xargs kubectl delete ns + local LAST_OUT=/tmp/tmp.vPMPHamnkv + local LAST_OUT=/tmp/tmp.81xm4Fg0Ac ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.vpjx3N0SUT + local exit_status=0 + local timeout=4 ++ seq 0 2 + local LAST_ERR=/tmp/tmp.lZpXyPAbXJ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get ns + for i in $(seq 0 2) + set +e + kubectl delete namespace psmdb-operator --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.81xm4Fg0Ac + cat /tmp/tmp.vpjx3N0SUT + rm /tmp/tmp.81xm4Fg0Ac /tmp/tmp.vpjx3N0SUT + return 0 namespace "cert-manager" deleted namespace "data-sharded-5941" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.vPMPHamnkv namespace "psmdb-operator" deleted + cat /tmp/tmp.lZpXyPAbXJ + rm /tmp/tmp.vPMPHamnkv /tmp/tmp.lZpXyPAbXJ + return 0 + kubectl_bin wait --for=delete namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.500V8FfA8F ++ mktemp + local LAST_ERR=/tmp/tmp.YiPf7TB2VK + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.500V8FfA8F + cat /tmp/tmp.YiPf7TB2VK + rm /tmp/tmp.500V8FfA8F /tmp/tmp.YiPf7TB2VK + return 0 + desc 'create namespace psmdb-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.wo6WoyzHDR ++ mktemp + local LAST_ERR=/tmp/tmp.SAptUJbUFr + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.wo6WoyzHDR namespace/psmdb-operator created + cat /tmp/tmp.SAptUJbUFr + rm /tmp/tmp.wo6WoyzHDR /tmp/tmp.SAptUJbUFr + return 0 + set_kube_ctx psmdb-operator + local namespace=psmdb-operator ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.bY8SABmF0g +++ mktemp ++ local LAST_ERR=/tmp/tmp.SWQT6LUEB6 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.bY8SABmF0g ++ cat /tmp/tmp.SWQT6LUEB6 ++ rm /tmp/tmp.bY8SABmF0g /tmp/tmp.SWQT6LUEB6 ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2219-d7e802db-7-cluster9 --namespace=psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.ui8ApdDaN7 ++ mktemp + local LAST_ERR=/tmp/tmp.yejtYDbpTo + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2219-d7e802db-7-cluster9 --namespace=psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ui8ApdDaN7 Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2219-d7e802db-7-cluster9" modified. + cat /tmp/tmp.yejtYDbpTo + rm /tmp/tmp.ui8ApdDaN7 /tmp/tmp.yejtYDbpTo + return 0 + deploy_operator + desc 'start PSMDB operator: docker.io/perconalab/percona-server-mongodb-operator:PR-2219-d7e802db' + set +o xtrace ----------------------------------------------------------------------------------- start PSMDB operator: docker.io/perconalab/percona-server-mongodb-operator:PR-2219-d7e802db ----------------------------------------------------------------------------------- + local cr_file + '[' -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/data-sharded/conf/crd.yaml ']' + cr_file=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/crd.yaml + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.6slUMnvu07 ++ mktemp + local LAST_ERR=/tmp/tmp.oNxCtqdcwH + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.6slUMnvu07 customresourcedefinition.apiextensions.k8s.io/perconaservermongodbbackups.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbrestores.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com serverside-applied + cat /tmp/tmp.oNxCtqdcwH + rm /tmp/tmp.6slUMnvu07 /tmp/tmp.oNxCtqdcwH + return 0 + '[' -n psmdb-operator ']' + apply_rbac cw-rbac + local operator_namespace=psmdb-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: psmdb-operator^' + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.yDncZRNa8V ++ mktemp + local LAST_ERR=/tmp/tmp.OwuCMZbMNd + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.yDncZRNa8V clusterrole.rbac.authorization.k8s.io/percona-server-mongodb-operator created serviceaccount/percona-server-mongodb-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-server-mongodb-operator created + cat /tmp/tmp.OwuCMZbMNd + rm /tmp/tmp.yDncZRNa8V /tmp/tmp.OwuCMZbMNd + return 0 + yq eval ' (.spec.template.spec.containers[].image = "docker.io/perconalab/percona-server-mongodb-operator:PR-2219-d7e802db") | ((.. | select(.[] == "DISABLE_TELEMETRY")) |= .value="true") | ((.. | select(.[] == "LOG_LEVEL")) |= .value="DEBUG")' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/cw-operator.yaml + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.UI2l9LSSjv ++ mktemp + local LAST_ERR=/tmp/tmp.QTWNomAcHH + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.UI2l9LSSjv deployment.apps/percona-server-mongodb-operator created + cat /tmp/tmp.QTWNomAcHH + rm /tmp/tmp.UI2l9LSSjv /tmp/tmp.QTWNomAcHH + return 0 + sleep 20 ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.L9gULRmGUR +++ mktemp ++ local LAST_ERR=/tmp/tmp.AgiH1RPebh ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.L9gULRmGUR ++ cat /tmp/tmp.AgiH1RPebh ++ rm /tmp/tmp.L9gULRmGUR /tmp/tmp.AgiH1RPebh ++ return 0 + wait_operator_pod percona-server-mongodb-operator-76758dcf9-hbrj7 + local pod=percona-server-mongodb-operator-76758dcf9-hbrj7 + set +o xtrace waiting for pod/percona-server-mongodb-operator-76758dcf9-hbrj7 to be ready.OK + echo 'Print operator info from log' Print operator info from log + grep 'Manager starting up' ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.UJK84bwWfq +++ mktemp ++ local LAST_ERR=/tmp/tmp.P5HKxr1qrk ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.UJK84bwWfq ++ cat /tmp/tmp.P5HKxr1qrk ++ rm /tmp/tmp.UJK84bwWfq /tmp/tmp.P5HKxr1qrk ++ return 0 + kubectl_bin logs -n psmdb-operator percona-server-mongodb-operator-76758dcf9-hbrj7 ++ mktemp + local LAST_OUT=/tmp/tmp.0noyp68Evz ++ mktemp + local LAST_ERR=/tmp/tmp.ALP66gPPoM + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl logs -n psmdb-operator percona-server-mongodb-operator-76758dcf9-hbrj7 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.0noyp68Evz + cat /tmp/tmp.ALP66gPPoM + rm /tmp/tmp.0noyp68Evz /tmp/tmp.ALP66gPPoM + return 0 2026-03-10T22:00:59.632Z INFO setup Manager starting up {"gitCommit": "d7e802db10c9b2b2028f56c7b5227cb276fe5878", "gitBranch": "PR-2219-d7e802db", "buildTime": "", "goVersion": "go1.25.8", "os": "linux", "arch": "amd64"} + create_namespace data-sharded-19777 + local namespace=data-sharded-19777 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// ++ helm list --all-namespaces --filter chaos-mesh + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ grep chaos-mesh.org ++ kubectl get crd ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + grep -E -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' + awk '{print$1}' + '[' -n '' ']' + desc 'cleaned up old namespaces data-sharded-19777' + set +o xtrace ----------------------------------------------------------------------------------- + xargs kubectl delete ns ++ mktemp cleaned up old namespaces data-sharded-19777 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace data-sharded-19777 --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.wZZ5dYzJ0i ++ mktemp + local LAST_OUT=/tmp/tmp.izmH0IRTul ++ mktemp + local LAST_ERR=/tmp/tmp.dTnF3wKiy3 + local exit_status=0 + local timeout=4 + local LAST_ERR=/tmp/tmp.h3RZ351fra + local exit_status=0 + local timeout=4 ++ seq 0 2 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get ns + for i in $(seq 0 2) + set +e + kubectl delete namespace data-sharded-19777 --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.wZZ5dYzJ0i + cat /tmp/tmp.dTnF3wKiy3 + rm /tmp/tmp.wZZ5dYzJ0i /tmp/tmp.dTnF3wKiy3 + return 0 error: resource(s) were provided, but no name was specified + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.izmH0IRTul + cat /tmp/tmp.h3RZ351fra + rm /tmp/tmp.izmH0IRTul /tmp/tmp.h3RZ351fra + return 0 + kubectl_bin wait --for=delete namespace data-sharded-19777 ++ mktemp + local LAST_OUT=/tmp/tmp.VROQ5XQud0 ++ mktemp + local LAST_ERR=/tmp/tmp.pf0wBu3o2Z + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete namespace data-sharded-19777 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.VROQ5XQud0 + cat /tmp/tmp.pf0wBu3o2Z + rm /tmp/tmp.VROQ5XQud0 /tmp/tmp.pf0wBu3o2Z + return 0 + desc 'create namespace data-sharded-19777' + set +o xtrace ----------------------------------------------------------------------------------- create namespace data-sharded-19777 ----------------------------------------------------------------------------------- + kubectl_bin create namespace data-sharded-19777 ++ mktemp + local LAST_OUT=/tmp/tmp.o0GJcNxtT3 ++ mktemp + local LAST_ERR=/tmp/tmp.ILHSwhZFnk + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace data-sharded-19777 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.o0GJcNxtT3 namespace/data-sharded-19777 created + cat /tmp/tmp.ILHSwhZFnk + rm /tmp/tmp.o0GJcNxtT3 /tmp/tmp.ILHSwhZFnk + return 0 + set_kube_ctx data-sharded-19777 + local namespace=data-sharded-19777 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.D59rhnLKF1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.yWWCvX2JPM ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.D59rhnLKF1 ++ cat /tmp/tmp.yWWCvX2JPM ++ rm /tmp/tmp.D59rhnLKF1 /tmp/tmp.yWWCvX2JPM ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2219-d7e802db-7-cluster9 --namespace=data-sharded-19777 ++ mktemp + local LAST_OUT=/tmp/tmp.7h4J6743Em ++ mktemp + local LAST_ERR=/tmp/tmp.9j3GA2Sa2k + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2219-d7e802db-7-cluster9 --namespace=data-sharded-19777 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.7h4J6743Em Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2219-d7e802db-7-cluster9" modified. + cat /tmp/tmp.9j3GA2Sa2k + rm /tmp/tmp.7h4J6743Em /tmp/tmp.9j3GA2Sa2k + return 0 + deploy_cert_manager + desc 'deploy cert manager' + set +o xtrace ----------------------------------------------------------------------------------- deploy cert manager ----------------------------------------------------------------------------------- + kubectl_bin create namespace cert-manager ++ mktemp + local LAST_OUT=/tmp/tmp.7KLje7xPXf ++ mktemp + local LAST_ERR=/tmp/tmp.CLnQ6weGzY + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace cert-manager + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.7KLje7xPXf namespace/cert-manager created + cat /tmp/tmp.CLnQ6weGzY + rm /tmp/tmp.7KLje7xPXf /tmp/tmp.CLnQ6weGzY + return 0 + kubectl_bin label namespace cert-manager certmanager.k8s.io/disable-validation=true ++ mktemp + local LAST_OUT=/tmp/tmp.MUI4zvchMa ++ mktemp + local LAST_ERR=/tmp/tmp.abLbiuXI2c + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl label namespace cert-manager certmanager.k8s.io/disable-validation=true + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.MUI4zvchMa namespace/cert-manager labeled + cat /tmp/tmp.abLbiuXI2c + rm /tmp/tmp.MUI4zvchMa /tmp/tmp.abLbiuXI2c + return 0 + kubectl_bin apply -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml --validate=false ++ mktemp + local LAST_OUT=/tmp/tmp.HnGv7Sm6Wm ++ mktemp + local LAST_ERR=/tmp/tmp.zKmx2dOPoC + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml --validate=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.HnGv7Sm6Wm namespace/cert-manager configured customresourcedefinition.apiextensions.k8s.io/challenges.acme.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/orders.acme.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/certificaterequests.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/certificates.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/clusterissuers.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/issuers.cert-manager.io unchanged serviceaccount/cert-manager-cainjector created serviceaccount/cert-manager created serviceaccount/cert-manager-webhook created clusterrole.rbac.authorization.k8s.io/cert-manager-cainjector unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-issuers unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificates unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-orders unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-challenges unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-cluster-view unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-view unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-edit unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-cainjector unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-issuers unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificates unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-orders unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-challenges unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews unchanged role.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection unchanged role.rbac.authorization.k8s.io/cert-manager:leaderelection unchanged role.rbac.authorization.k8s.io/cert-manager-tokenrequest created role.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created rolebinding.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection unchanged rolebinding.rbac.authorization.k8s.io/cert-manager:leaderelection unchanged rolebinding.rbac.authorization.k8s.io/cert-manager-tokenrequest created rolebinding.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created service/cert-manager-cainjector created service/cert-manager created service/cert-manager-webhook created deployment.apps/cert-manager-cainjector created deployment.apps/cert-manager created deployment.apps/cert-manager-webhook created mutatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook configured validatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook configured + cat /tmp/tmp.zKmx2dOPoC Warning: resource namespaces/cert-manager is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. + rm /tmp/tmp.HnGv7Sm6Wm /tmp/tmp.zKmx2dOPoC + return 0 + kubectl_bin -n cert-manager wait pod -l app.kubernetes.io/instance=cert-manager --for=condition=ready ++ mktemp + local LAST_OUT=/tmp/tmp.1NyRi1OyuW ++ mktemp + local LAST_ERR=/tmp/tmp.s7lkLaE1Po + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl -n cert-manager wait pod -l app.kubernetes.io/instance=cert-manager --for=condition=ready + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.1NyRi1OyuW pod/cert-manager-559d798845-bh2dv condition met pod/cert-manager-cainjector-64958d9c7c-mq785 condition met pod/cert-manager-webhook-7fb6f99b56-x8sqp condition met + cat /tmp/tmp.s7lkLaE1Po + rm /tmp/tmp.1NyRi1OyuW /tmp/tmp.s7lkLaE1Po + return 0 + sleep 120 + desc 'create secrets and start client' + set +o xtrace ----------------------------------------------------------------------------------- create secrets and start client ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.eEP7HzXdcv ++ mktemp + local LAST_ERR=/tmp/tmp.lrYYS9kPnJ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/conf/secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.eEP7HzXdcv secret/some-users created + cat /tmp/tmp.lrYYS9kPnJ + rm /tmp/tmp.eEP7HzXdcv /tmp/tmp.lrYYS9kPnJ + return 0 + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/conf/client_with_tls.yml ++ mktemp + local LAST_OUT=/tmp/tmp.4DNXLcTJME ++ mktemp + local LAST_ERR=/tmp/tmp.0NBeANCzRX + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/conf/client_with_tls.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.4DNXLcTJME deployment.apps/psmdb-client created + cat /tmp/tmp.0NBeANCzRX + rm /tmp/tmp.4DNXLcTJME /tmp/tmp.0NBeANCzRX + return 0 + cluster=some-name + desc 'create first PSMDB cluster some-name' + set +o xtrace ----------------------------------------------------------------------------------- create first PSMDB cluster some-name ----------------------------------------------------------------------------------- + apply_cluster /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/data-sharded/conf/some-name.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/data-sharded/conf/some-name.yml + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/data-sharded/conf/some-name.yml + yq eval '(.spec | select(.image == null)).image = "docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0"' + yq eval '(.spec | select(has("pmm"))).pmm.image = "docker.io/percona/pmm-client:2.44.1-1"' + yq eval '(.spec | select(has("initImage"))).initImage = "docker.io/perconalab/percona-server-mongodb-operator:PR-2219-d7e802db"' + yq eval '(.spec | select(has("backup"))).backup.image = "docker.io/perconalab/percona-server-mongodb-operator:main-backup"' ++ mktemp + /usr/sbin/sed -e s/NAME_SPACE/data-sharded-19777/g + yq eval '.spec.upgradeOptions.apply="Never"' + local LAST_OUT=/tmp/tmp.K7XxyR1zQe ++ mktemp + local LAST_ERR=/tmp/tmp.imlHgJMGjQ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.K7XxyR1zQe perconaservermongodb.psmdb.percona.com/some-name created + cat /tmp/tmp.imlHgJMGjQ + rm /tmp/tmp.K7XxyR1zQe /tmp/tmp.imlHgJMGjQ + return 0 + desc 'check if all Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all Pods started ----------------------------------------------------------------------------------- + wait_for_running some-name-cfg 3 + local name=some-name-cfg + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=cfg + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-cfg-0 + local pod=some-name-cfg-0 + set +o xtrace waiting for pod/some-name-cfg-0 to be ready.............OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-cfg-1 + local pod=some-name-cfg-1 + set +o xtrace waiting for pod/some-name-cfg-1 to be ready.............OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ii2TzpU1sX +++ mktemp ++ local LAST_ERR=/tmp/tmp.Lm9YQwWPRl ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Ii2TzpU1sX ++ cat /tmp/tmp.Lm9YQwWPRl ++ rm /tmp/tmp.Ii2TzpU1sX /tmp/tmp.Lm9YQwWPRl ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-cfg-2 + local pod=some-name-cfg-2 + set +o xtrace waiting for pod/some-name-cfg-2 to be ready............OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uLmvLKrT59 +++ mktemp ++ local LAST_ERR=/tmp/tmp.4Epggo9mVU ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.uLmvLKrT59 ++ cat /tmp/tmp.4Epggo9mVU ++ rm /tmp/tmp.uLmvLKrT59 /tmp/tmp.4Epggo9mVU ++ return 0 + [[ '' == \t\r\u\e ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FyAHXC0AKg +++ mktemp ++ local LAST_ERR=/tmp/tmp.7bUE6pDh8A ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.FyAHXC0AKg ++ cat /tmp/tmp.7bUE6pDh8A ++ rm /tmp/tmp.FyAHXC0AKg /tmp/tmp.7bUE6pDh8A ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness.............................................................................. + wait_for_running some-name-rs0 3 + local name=some-name-rs0 + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=rs0 + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-rs0-0 + local pod=some-name-rs0-0 + set +o xtrace waiting for pod/some-name-rs0-0 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-rs0-1 + local pod=some-name-rs0-1 + set +o xtrace waiting for pod/some-name-rs0-1 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YXm0l3ON1p +++ mktemp ++ local LAST_ERR=/tmp/tmp.bxPMZs3eoz ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.YXm0l3ON1p ++ cat /tmp/tmp.bxPMZs3eoz ++ rm /tmp/tmp.YXm0l3ON1p /tmp/tmp.bxPMZs3eoz ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-rs0-2 + local pod=some-name-rs0-2 + set +o xtrace waiting for pod/some-name-rs0-2 to be ready.OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jN8lJPpNYG +++ mktemp ++ local LAST_ERR=/tmp/tmp.sWIEq6Uaul ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jN8lJPpNYG ++ cat /tmp/tmp.sWIEq6Uaul ++ rm /tmp/tmp.jN8lJPpNYG /tmp/tmp.sWIEq6Uaul ++ return 0 + [[ '' == \t\r\u\e ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.V5nou1fLx3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.1YrkNYWjLI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.V5nou1fLx3 ++ cat /tmp/tmp.1YrkNYWjLI ++ rm /tmp/tmp.V5nou1fLx3 /tmp/tmp.1YrkNYWjLI ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness + wait_for_running some-name-rs1 3 + local name=some-name-rs1 + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=rs1 + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-rs1-0 + local pod=some-name-rs1-0 + set +o xtrace waiting for pod/some-name-rs1-0 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-rs1-1 + local pod=some-name-rs1-1 + set +o xtrace waiting for pod/some-name-rs1-1 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs1")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hXZuAheEfN +++ mktemp ++ local LAST_ERR=/tmp/tmp.y04t3sPrho ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs1")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.hXZuAheEfN ++ cat /tmp/tmp.y04t3sPrho ++ rm /tmp/tmp.hXZuAheEfN /tmp/tmp.y04t3sPrho ++ return 0 + [[ true == \t\r\u\e ]] + wait_pod some-name-rs1-arbiter-0 + local pod=some-name-rs1-arbiter-0 + set +o xtrace waiting for pod/some-name-rs1-arbiter-0 to be ready.OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs1")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bbQeZt43h1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.mb6np3IG4e ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs1")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.bbQeZt43h1 ++ cat /tmp/tmp.mb6np3IG4e ++ rm /tmp/tmp.bbQeZt43h1 /tmp/tmp.mb6np3IG4e ++ return 0 + [[ '' == \t\r\u\e ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs1")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eKpW9nNroI +++ mktemp ++ local LAST_ERR=/tmp/tmp.zywd7wBdNL ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs1")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.eKpW9nNroI ++ cat /tmp/tmp.zywd7wBdNL ++ rm /tmp/tmp.eKpW9nNroI /tmp/tmp.zywd7wBdNL ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness + wait_for_running some-name-rs2 3 + local name=some-name-rs2 + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=rs2 + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-rs2-0 + local pod=some-name-rs2-0 + set +o xtrace waiting for pod/some-name-rs2-0 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-rs2-1 + local pod=some-name-rs2-1 + set +o xtrace waiting for pod/some-name-rs2-1 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs2")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.l2NRMPPLiW +++ mktemp ++ local LAST_ERR=/tmp/tmp.FfMU7YIOc1 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs2")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.l2NRMPPLiW ++ cat /tmp/tmp.FfMU7YIOc1 ++ rm /tmp/tmp.l2NRMPPLiW /tmp/tmp.FfMU7YIOc1 ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-rs2-2 + local pod=some-name-rs2-2 + set +o xtrace waiting for pod/some-name-rs2-2 to be ready.OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs2")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3Xo0HmxT7J +++ mktemp ++ local LAST_ERR=/tmp/tmp.D8O9YPcs9V ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs2")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.3Xo0HmxT7J ++ cat /tmp/tmp.D8O9YPcs9V ++ rm /tmp/tmp.3Xo0HmxT7J /tmp/tmp.D8O9YPcs9V ++ return 0 + [[ true == \t\r\u\e ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs2")].nonvoting.size}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xThrxeuVxS +++ mktemp ++ local LAST_ERR=/tmp/tmp.o1n06EcSep ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs2")].nonvoting.size}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xThrxeuVxS ++ cat /tmp/tmp.o1n06EcSep ++ rm /tmp/tmp.xThrxeuVxS /tmp/tmp.o1n06EcSep ++ return 0 + last_pod=2 ++ seq 0 2 + for i in $(seq 0 $last_pod) + wait_pod some-name-rs2-nv-0 + local pod=some-name-rs2-nv-0 + set +o xtrace waiting for pod/some-name-rs2-nv-0 to be ready.OK + for i in $(seq 0 $last_pod) + wait_pod some-name-rs2-nv-1 + local pod=some-name-rs2-nv-1 + set +o xtrace waiting for pod/some-name-rs2-nv-1 to be ready.OK + for i in $(seq 0 $last_pod) + wait_pod some-name-rs2-nv-2 + local pod=some-name-rs2-nv-2 + set +o xtrace waiting for pod/some-name-rs2-nv-2 to be ready.OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs2")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.soWzQ7LTBW +++ mktemp ++ local LAST_ERR=/tmp/tmp.NZHauAUz6S ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs2")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.soWzQ7LTBW ++ cat /tmp/tmp.NZHauAUz6S ++ rm /tmp/tmp.soWzQ7LTBW /tmp/tmp.NZHauAUz6S ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness + wait_for_running some-name-mongos 3 + local name=some-name-mongos + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=mongos + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-mongos-0 + local pod=some-name-mongos-0 + set +o xtrace waiting for pod/some-name-mongos-0 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-mongos-1 + local pod=some-name-mongos-1 + set +o xtrace waiting for pod/some-name-mongos-1 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dwgT2wyQSM +++ mktemp ++ local LAST_ERR=/tmp/tmp.LMgnTfelfj ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.dwgT2wyQSM ++ cat /tmp/tmp.LMgnTfelfj ++ rm /tmp/tmp.dwgT2wyQSM /tmp/tmp.LMgnTfelfj ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-mongos-2 + local pod=some-name-mongos-2 + set +o xtrace waiting for pod/some-name-mongos-2 to be ready.OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vAVcQhobJm +++ mktemp ++ local LAST_ERR=/tmp/tmp.qaxP0IX78h ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.vAVcQhobJm ++ cat /tmp/tmp.qaxP0IX78h ++ rm /tmp/tmp.vAVcQhobJm /tmp/tmp.qaxP0IX78h ++ return 0 + [[ '' == \t\r\u\e ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ahkNocPZfn +++ mktemp ++ local LAST_ERR=/tmp/tmp.TsuK12jC0o ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ahkNocPZfn ++ cat /tmp/tmp.TsuK12jC0o ++ rm /tmp/tmp.ahkNocPZfn /tmp/tmp.TsuK12jC0o ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness + desc 'create user' + set +o xtrace ----------------------------------------------------------------------------------- create user ----------------------------------------------------------------------------------- + run_mongos 'db.createUser({user:"user",pwd:"pass",roles:[{db:"app",role:"readWrite"}]})' userAdmin:userAdmin123456@some-name-mongos.data-sharded-19777 mongodb .svc.cluster.local '--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + local 'command=db.createUser({user:"user",pwd:"pass",roles:[{db:"app",role:"readWrite"}]})' + local uri=userAdmin:userAdmin123456@some-name-mongos.data-sharded-19777 + local driver=mongodb + local suffix=.svc.cluster.local + local 'mongo_flag=--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + local port=27017 + local mongo_bin=mongo ++ echo .svc.cluster.local ++ awk -F: '{print $2}' + suffix_port= + [[ -z '' ]] + suffix=.svc.cluster.local:27017 ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.p5DGKZkP6N +++ mktemp ++ local LAST_ERR=/tmp/tmp.xqvtnvBxE7 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.p5DGKZkP6N ++ cat /tmp/tmp.xqvtnvBxE7 ++ rm /tmp/tmp.p5DGKZkP6N /tmp/tmp.xqvtnvBxE7 ++ return 0 + local client_container=psmdb-client-5649fbb65f-ffgcf + kubectl_bin exec psmdb-client-5649fbb65f-ffgcf -- bash -c 'printf '\''db.createUser({user:"user",pwd:"pass",roles:[{db:"app",role:"readWrite"}]})\n'\'' | mongo mongodb://userAdmin:userAdmin123456@some-name-mongos.data-sharded-19777.svc.cluster.local:27017/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' ++ mktemp + local LAST_OUT=/tmp/tmp.0p7YiFTluv ++ mktemp + local LAST_ERR=/tmp/tmp.VRvYgtQJz6 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-5649fbb65f-ffgcf -- bash -c 'printf '\''db.createUser({user:"user",pwd:"pass",roles:[{db:"app",role:"readWrite"}]})\n'\'' | mongo mongodb://userAdmin:userAdmin123456@some-name-mongos.data-sharded-19777.svc.cluster.local:27017/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.0p7YiFTluv Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.data-sharded-19777.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb {"t":{"$date":"2026-03-10T22:09:56.922Z"},"s":"I", "c":"NETWORK", "id":5490002, "ctx":"thread1","msg":"Started a new thread for the timer service"} Implicit session: session { "id" : UUID("bf85e626-5d48-4f31-b249-687eb7862576") } Percona Server for MongoDB server version: v8.0.19-7 WARNING: shell and server versions do not match Successfully added user: { "user" : "user", "roles" : [ { "db" : "app", "role" : "readWrite" } ] } bye + cat /tmp/tmp.VRvYgtQJz6 + rm /tmp/tmp.0p7YiFTluv /tmp/tmp.VRvYgtQJz6 + return 0 + sleep 2 + desc 'set chunk size to 32 MB' + set +o xtrace ----------------------------------------------------------------------------------- set chunk size to 32 MB ----------------------------------------------------------------------------------- + run_mongos 'use config\n db.settings.save( { _id:"chunksize", value: 32 } )' clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777 mongodb .svc.cluster.local '--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + local 'command=use config\n db.settings.save( { _id:"chunksize", value: 32 } )' + local uri=clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777 + local driver=mongodb + local suffix=.svc.cluster.local + local 'mongo_flag=--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + local port=27017 + local mongo_bin=mongo ++ echo .svc.cluster.local ++ awk -F: '{print $2}' + suffix_port= + [[ -z '' ]] + suffix=.svc.cluster.local:27017 ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mAUzg9aR1T +++ mktemp ++ local LAST_ERR=/tmp/tmp.OIignhTpUb ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.mAUzg9aR1T ++ cat /tmp/tmp.OIignhTpUb ++ rm /tmp/tmp.mAUzg9aR1T /tmp/tmp.OIignhTpUb ++ return 0 + local client_container=psmdb-client-5649fbb65f-ffgcf + kubectl_bin exec psmdb-client-5649fbb65f-ffgcf -- bash -c 'printf '\''use config\n db.settings.save( { _id:"chunksize", value: 32 } )\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777.svc.cluster.local:27017/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' ++ mktemp + local LAST_OUT=/tmp/tmp.L31rxxKg04 ++ mktemp + local LAST_ERR=/tmp/tmp.rll4GTlVhr + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-5649fbb65f-ffgcf -- bash -c 'printf '\''use config\n db.settings.save( { _id:"chunksize", value: 32 } )\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777.svc.cluster.local:27017/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.L31rxxKg04 Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.data-sharded-19777.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb {"t":{"$date":"2026-03-10T22:10:01.680Z"},"s":"I", "c":"NETWORK", "id":5490002, "ctx":"thread1","msg":"Started a new thread for the timer service"} Implicit session: session { "id" : UUID("2fe3cc89-51b5-4630-bf2b-7a993b3bd076") } Percona Server for MongoDB server version: v8.0.19-7 WARNING: shell and server versions do not match switched to db config WriteResult({ "nMatched" : 0, "nUpserted" : 1, "nModified" : 0, "_id" : "chunksize" }) bye + cat /tmp/tmp.rll4GTlVhr + rm /tmp/tmp.L31rxxKg04 /tmp/tmp.rll4GTlVhr + return 0 + sleep 2 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + run_script_mongos /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/data-sharded/data.js user:pass@some-name-mongos.data-sharded-19777 mongodb .svc.cluster.local '--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + local script=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/data-sharded/data.js + local uri=user:pass@some-name-mongos.data-sharded-19777 + local driver=mongodb + local suffix=.svc.cluster.local + local 'mongo_flag=--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + local mongo_bin=mongo ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ApttQKoH4T +++ mktemp ++ local LAST_ERR=/tmp/tmp.9URJXZwixX ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ApttQKoH4T ++ cat /tmp/tmp.9URJXZwixX ++ rm /tmp/tmp.ApttQKoH4T /tmp/tmp.9URJXZwixX ++ return 0 + local client_container=psmdb-client-5649fbb65f-ffgcf ++ basename /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/data-sharded/data.js + name=data.js + kubectl_bin cp /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/data-sharded/data.js data-sharded-19777/psmdb-client-5649fbb65f-ffgcf:/tmp ++ mktemp + local LAST_OUT=/tmp/tmp.qp00aWUD3r ++ mktemp + local LAST_ERR=/tmp/tmp.ybattRkeT1 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl cp /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/e2e-tests/data-sharded/data.js data-sharded-19777/psmdb-client-5649fbb65f-ffgcf:/tmp + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.qp00aWUD3r + cat /tmp/tmp.ybattRkeT1 + rm /tmp/tmp.qp00aWUD3r /tmp/tmp.ybattRkeT1 + return 0 + kubectl_bin exec psmdb-client-5649fbb65f-ffgcf -- bash -c 'mongo mongodb://user:pass@some-name-mongos.data-sharded-19777.svc.cluster.local/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls /tmp/data.js' ++ mktemp + local LAST_OUT=/tmp/tmp.phzfXi5ywL ++ mktemp + local LAST_ERR=/tmp/tmp.bK6pDnuTHB + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-5649fbb65f-ffgcf -- bash -c 'mongo mongodb://user:pass@some-name-mongos.data-sharded-19777.svc.cluster.local/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls /tmp/data.js' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.phzfXi5ywL Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.data-sharded-19777.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb {"t":{"$date":"2026-03-10T22:10:08.830Z"},"s":"I", "c":"NETWORK", "id":5490002, "ctx":"thread1","msg":"Started a new thread for the timer service"} Implicit session: session { "id" : UUID("43bf88a7-1ef0-4de2-8d87-c15e802841ea") } Percona Server for MongoDB server version: v8.0.19-7 WARNING: shell and server versions do not match + cat /tmp/tmp.bK6pDnuTHB + rm /tmp/tmp.phzfXi5ywL /tmp/tmp.bK6pDnuTHB + return 0 + desc 'shard collection' + set +o xtrace ----------------------------------------------------------------------------------- shard collection ----------------------------------------------------------------------------------- + run_mongos 'sh.enableSharding("app")' clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777 mongodb .svc.cluster.local '--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + local 'command=sh.enableSharding("app")' + local uri=clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777 + local driver=mongodb + local suffix=.svc.cluster.local + local 'mongo_flag=--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + local port=27017 + local mongo_bin=mongo ++ echo .svc.cluster.local ++ awk -F: '{print $2}' + suffix_port= + [[ -z '' ]] + suffix=.svc.cluster.local:27017 ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Gy4OMauu79 +++ mktemp ++ local LAST_ERR=/tmp/tmp.anY7eMWHKG ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Gy4OMauu79 ++ cat /tmp/tmp.anY7eMWHKG ++ rm /tmp/tmp.Gy4OMauu79 /tmp/tmp.anY7eMWHKG ++ return 0 + local client_container=psmdb-client-5649fbb65f-ffgcf + kubectl_bin exec psmdb-client-5649fbb65f-ffgcf -- bash -c 'printf '\''sh.enableSharding("app")\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777.svc.cluster.local:27017/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' ++ mktemp + local LAST_OUT=/tmp/tmp.eP3byA0t3G ++ mktemp + local LAST_ERR=/tmp/tmp.jGKPcZPv0V + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-5649fbb65f-ffgcf -- bash -c 'printf '\''sh.enableSharding("app")\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777.svc.cluster.local:27017/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.eP3byA0t3G Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.data-sharded-19777.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb {"t":{"$date":"2026-03-10T22:13:28.155Z"},"s":"I", "c":"NETWORK", "id":5490002, "ctx":"thread1","msg":"Started a new thread for the timer service"} Implicit session: session { "id" : UUID("e19e5235-39c1-4c4b-86f5-c43f4ae4b7d2") } Percona Server for MongoDB server version: v8.0.19-7 WARNING: shell and server versions do not match { "ok" : 1, "$clusterTime" : { "clusterTime" : Timestamp(1773180807, 1), "signature" : { "hash" : BinData(0,"giY/+uKCQHudQYmsn1VZGtvOXJA="), "keyId" : NumberLong("7615751510080618520") } }, "operationTime" : Timestamp(1773180807, 1) } bye + cat /tmp/tmp.jGKPcZPv0V + rm /tmp/tmp.eP3byA0t3G /tmp/tmp.jGKPcZPv0V + return 0 + sleep 2 + run_mongos 'sh.shardCollection("app.city", { _id: 1 } )' clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777 mongodb .svc.cluster.local '--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + local 'command=sh.shardCollection("app.city", { _id: 1 } )' + local uri=clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777 + local driver=mongodb + local suffix=.svc.cluster.local + local 'mongo_flag=--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + local port=27017 + local mongo_bin=mongo ++ echo .svc.cluster.local ++ awk -F: '{print $2}' + suffix_port= + [[ -z '' ]] + suffix=.svc.cluster.local:27017 ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.B7W1VJuWNW +++ mktemp ++ local LAST_ERR=/tmp/tmp.Wyqfxp8TjB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.B7W1VJuWNW ++ cat /tmp/tmp.Wyqfxp8TjB ++ rm /tmp/tmp.B7W1VJuWNW /tmp/tmp.Wyqfxp8TjB ++ return 0 + local client_container=psmdb-client-5649fbb65f-ffgcf + kubectl_bin exec psmdb-client-5649fbb65f-ffgcf -- bash -c 'printf '\''sh.shardCollection("app.city", { _id: 1 } )\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777.svc.cluster.local:27017/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' ++ mktemp + local LAST_OUT=/tmp/tmp.lxKVJbO1ob ++ mktemp + local LAST_ERR=/tmp/tmp.qVW7Sip5ui + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-5649fbb65f-ffgcf -- bash -c 'printf '\''sh.shardCollection("app.city", { _id: 1 } )\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777.svc.cluster.local:27017/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.lxKVJbO1ob Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.data-sharded-19777.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb {"t":{"$date":"2026-03-10T22:13:32.745Z"},"s":"I", "c":"NETWORK", "id":5490002, "ctx":"thread1","msg":"Started a new thread for the timer service"} Implicit session: session { "id" : UUID("96922135-94bf-4118-828f-844bd815d8db") } Percona Server for MongoDB server version: v8.0.19-7 WARNING: shell and server versions do not match { "collectionsharded" : "app.city", "ok" : 1, "$clusterTime" : { "clusterTime" : Timestamp(1773180813, 25), "signature" : { "hash" : BinData(0,"6adQn+DC4lgCiCdquIdurBQqYAo="), "keyId" : NumberLong("7615751510080618520") } }, "operationTime" : Timestamp(1773180813, 25) } bye + cat /tmp/tmp.qVW7Sip5ui + rm /tmp/tmp.lxKVJbO1ob /tmp/tmp.qVW7Sip5ui + return 0 + sleep 120 + desc 'check chunks' + set +o xtrace ----------------------------------------------------------------------------------- check chunks ----------------------------------------------------------------------------------- + chunks_param1=ns + chunks_param2='"app.city"' + [[ 8.0 != \4\.\4 ]] + chunks_param1=uuid ++ run_mongos 'use app\n db.getCollectionInfos({ "name": "city" })[0].info.uuid' user:pass@some-name-mongos.data-sharded-19777 '' '' '--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' ++ local 'command=use app\n db.getCollectionInfos({ "name": "city" })[0].info.uuid' ++ local uri=user:pass@some-name-mongos.data-sharded-19777 ++ local driver=mongodb ++ local suffix=.svc.cluster.local ++ local 'mongo_flag=--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' ++ local port=27017 ++ local mongo_bin=mongo ++ grep 'switched to db app' -A 1 ++ grep -v 'switched to db app' +++ echo .svc.cluster.local +++ awk -F: '{print $2}' ++ suffix_port= ++ [[ -z '' ]] ++ suffix=.svc.cluster.local:27017 +++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.hPrecc74H1 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.8PqiLOozIw +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.hPrecc74H1 +++ cat /tmp/tmp.8PqiLOozIw +++ rm /tmp/tmp.hPrecc74H1 /tmp/tmp.8PqiLOozIw +++ return 0 ++ local client_container=psmdb-client-5649fbb65f-ffgcf ++ kubectl_bin exec psmdb-client-5649fbb65f-ffgcf -- bash -c 'printf '\''use app\n db.getCollectionInfos({ "name": "city" })[0].info.uuid\n'\'' | mongo mongodb://user:pass@some-name-mongos.data-sharded-19777.svc.cluster.local:27017/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' +++ mktemp ++ local LAST_OUT=/tmp/tmp.h04vAGRcN8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.fb6npHxUUE ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl exec psmdb-client-5649fbb65f-ffgcf -- bash -c 'printf '\''use app\n db.getCollectionInfos({ "name": "city" })[0].info.uuid\n'\'' | mongo mongodb://user:pass@some-name-mongos.data-sharded-19777.svc.cluster.local:27017/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.h04vAGRcN8 ++ cat /tmp/tmp.fb6npHxUUE ++ rm /tmp/tmp.h04vAGRcN8 /tmp/tmp.fb6npHxUUE ++ return 0 + chunks_param2='UUID("e33ee60f-7cc5-4f89-bf5c-6cb8c0225d8d")' + shards=0 + for i in "rs0" "rs1" "rs2" ++ run_mongos 'use config\n db.chunks.count({"uuid": UUID("e33ee60f-7cc5-4f89-bf5c-6cb8c0225d8d"), "shard": "rs0"})' clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777 mongodb .svc.cluster.local '--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' ++ local 'command=use config\n db.chunks.count({"uuid": UUID("e33ee60f-7cc5-4f89-bf5c-6cb8c0225d8d"), "shard": "rs0"})' ++ local uri=clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777 ++ local driver=mongodb ++ local suffix=.svc.cluster.local ++ local 'mongo_flag=--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' ++ local port=27017 ++ local mongo_bin=mongo ++ grep 'switched to db config' -A 1 ++ grep -v 'switched to db config' +++ echo .svc.cluster.local +++ awk -F: '{print $2}' ++ suffix_port= ++ [[ -z '' ]] ++ suffix=.svc.cluster.local:27017 +++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.iYh5YKRWAb ++++ mktemp +++ local LAST_ERR=/tmp/tmp.L69eYPNoj7 +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.iYh5YKRWAb +++ cat /tmp/tmp.L69eYPNoj7 +++ rm /tmp/tmp.iYh5YKRWAb /tmp/tmp.L69eYPNoj7 +++ return 0 ++ local client_container=psmdb-client-5649fbb65f-ffgcf ++ kubectl_bin exec psmdb-client-5649fbb65f-ffgcf -- bash -c 'printf '\''use config\n db.chunks.count({"uuid": UUID("e33ee60f-7cc5-4f89-bf5c-6cb8c0225d8d"), "shard": "rs0"})\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777.svc.cluster.local:27017/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0LFIBgfjaV +++ mktemp ++ local LAST_ERR=/tmp/tmp.ACNieEJRnw ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl exec psmdb-client-5649fbb65f-ffgcf -- bash -c 'printf '\''use config\n db.chunks.count({"uuid": UUID("e33ee60f-7cc5-4f89-bf5c-6cb8c0225d8d"), "shard": "rs0"})\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777.svc.cluster.local:27017/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.0LFIBgfjaV ++ cat /tmp/tmp.ACNieEJRnw ++ rm /tmp/tmp.0LFIBgfjaV /tmp/tmp.ACNieEJRnw ++ return 0 + out=1 + desc 'rs0 has 1 chunks' + set +o xtrace ----------------------------------------------------------------------------------- rs0 has 1 chunks ----------------------------------------------------------------------------------- + [[ 1 -ne 0 ]] + (( shards = shards + 1 )) + for i in "rs0" "rs1" "rs2" ++ run_mongos 'use config\n db.chunks.count({"uuid": UUID("e33ee60f-7cc5-4f89-bf5c-6cb8c0225d8d"), "shard": "rs1"})' clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777 mongodb .svc.cluster.local '--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' ++ grep 'switched to db config' -A 1 ++ local 'command=use config\n db.chunks.count({"uuid": UUID("e33ee60f-7cc5-4f89-bf5c-6cb8c0225d8d"), "shard": "rs1"})' ++ local uri=clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777 ++ local driver=mongodb ++ local suffix=.svc.cluster.local ++ local 'mongo_flag=--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' ++ local port=27017 ++ local mongo_bin=mongo ++ grep -v 'switched to db config' +++ echo .svc.cluster.local +++ awk -F: '{print $2}' ++ suffix_port= ++ [[ -z '' ]] ++ suffix=.svc.cluster.local:27017 +++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.e55jzdz9Bq ++++ mktemp +++ local LAST_ERR=/tmp/tmp.iQSF3zgXKM +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.e55jzdz9Bq +++ cat /tmp/tmp.iQSF3zgXKM +++ rm /tmp/tmp.e55jzdz9Bq /tmp/tmp.iQSF3zgXKM +++ return 0 ++ local client_container=psmdb-client-5649fbb65f-ffgcf ++ kubectl_bin exec psmdb-client-5649fbb65f-ffgcf -- bash -c 'printf '\''use config\n db.chunks.count({"uuid": UUID("e33ee60f-7cc5-4f89-bf5c-6cb8c0225d8d"), "shard": "rs1"})\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777.svc.cluster.local:27017/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nH1Tr2jB11 +++ mktemp ++ local LAST_ERR=/tmp/tmp.B9kY0nbWhw ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl exec psmdb-client-5649fbb65f-ffgcf -- bash -c 'printf '\''use config\n db.chunks.count({"uuid": UUID("e33ee60f-7cc5-4f89-bf5c-6cb8c0225d8d"), "shard": "rs1"})\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777.svc.cluster.local:27017/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.nH1Tr2jB11 ++ cat /tmp/tmp.B9kY0nbWhw ++ rm /tmp/tmp.nH1Tr2jB11 /tmp/tmp.B9kY0nbWhw ++ return 0 + out=1 + desc 'rs1 has 1 chunks' + set +o xtrace ----------------------------------------------------------------------------------- rs1 has 1 chunks ----------------------------------------------------------------------------------- + [[ 1 -ne 0 ]] + (( shards = shards + 1 )) + for i in "rs0" "rs1" "rs2" ++ run_mongos 'use config\n db.chunks.count({"uuid": UUID("e33ee60f-7cc5-4f89-bf5c-6cb8c0225d8d"), "shard": "rs2"})' clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777 mongodb .svc.cluster.local '--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' ++ grep 'switched to db config' -A 1 ++ local 'command=use config\n db.chunks.count({"uuid": UUID("e33ee60f-7cc5-4f89-bf5c-6cb8c0225d8d"), "shard": "rs2"})' ++ local uri=clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777 ++ local driver=mongodb ++ local suffix=.svc.cluster.local ++ local 'mongo_flag=--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' ++ local port=27017 ++ local mongo_bin=mongo ++ grep -v 'switched to db config' +++ echo .svc.cluster.local +++ awk -F: '{print $2}' ++ suffix_port= ++ [[ -z '' ]] ++ suffix=.svc.cluster.local:27017 +++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.NFpdlZrvZH ++++ mktemp +++ local LAST_ERR=/tmp/tmp.wmTMzy9S5O +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.NFpdlZrvZH +++ cat /tmp/tmp.wmTMzy9S5O +++ rm /tmp/tmp.NFpdlZrvZH /tmp/tmp.wmTMzy9S5O +++ return 0 ++ local client_container=psmdb-client-5649fbb65f-ffgcf ++ kubectl_bin exec psmdb-client-5649fbb65f-ffgcf -- bash -c 'printf '\''use config\n db.chunks.count({"uuid": UUID("e33ee60f-7cc5-4f89-bf5c-6cb8c0225d8d"), "shard": "rs2"})\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777.svc.cluster.local:27017/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZhE748toTj +++ mktemp ++ local LAST_ERR=/tmp/tmp.G35nGC7swM ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl exec psmdb-client-5649fbb65f-ffgcf -- bash -c 'printf '\''use config\n db.chunks.count({"uuid": UUID("e33ee60f-7cc5-4f89-bf5c-6cb8c0225d8d"), "shard": "rs2"})\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777.svc.cluster.local:27017/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ZhE748toTj ++ cat /tmp/tmp.G35nGC7swM ++ rm /tmp/tmp.ZhE748toTj /tmp/tmp.G35nGC7swM ++ return 0 + out=1 + desc 'rs2 has 1 chunks' + set +o xtrace ----------------------------------------------------------------------------------- rs2 has 1 chunks ----------------------------------------------------------------------------------- + [[ 1 -ne 0 ]] + (( shards = shards + 1 )) + [[ 3 -lt 3 ]] ++ run_mongos 'use app\n db.dropDatabase()' clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777 mongodb .svc.cluster.local '--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' ++ local 'command=use app\n db.dropDatabase()' ++ local uri=clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777 ++ local driver=mongodb ++ local suffix=.svc.cluster.local ++ local 'mongo_flag=--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' ++ local port=27017 ++ local mongo_bin=mongo +++ echo .svc.cluster.local +++ awk -F: '{print $2}' ++ suffix_port= ++ [[ -z '' ]] ++ suffix=.svc.cluster.local:27017 +++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.CFWxVqM6m8 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.JNyucNEcJ6 +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.CFWxVqM6m8 +++ cat /tmp/tmp.JNyucNEcJ6 +++ rm /tmp/tmp.CFWxVqM6m8 /tmp/tmp.JNyucNEcJ6 +++ return 0 ++ local client_container=psmdb-client-5649fbb65f-ffgcf ++ kubectl_bin exec psmdb-client-5649fbb65f-ffgcf -- bash -c 'printf '\''use app\n db.dropDatabase()\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777.svc.cluster.local:27017/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JcTuc2hk3f +++ mktemp ++ local LAST_ERR=/tmp/tmp.1LojsfJZmG ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl exec psmdb-client-5649fbb65f-ffgcf -- bash -c 'printf '\''use app\n db.dropDatabase()\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@some-name-mongos.data-sharded-19777.svc.cluster.local:27017/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.JcTuc2hk3f ++ cat /tmp/tmp.1LojsfJZmG ++ rm /tmp/tmp.JcTuc2hk3f /tmp/tmp.1LojsfJZmG ++ return 0 + res='Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.data-sharded-19777.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb {"t":{"$date":"2026-03-10T22:15:45.240Z"},"s":"I", "c":"NETWORK", "id":5490002, "ctx":"thread1","msg":"Started a new thread for the timer service"} Implicit session: session { "id" : UUID("0ee5836f-9012-4524-af4b-7985e82b2dca") } Percona Server for MongoDB server version: v8.0.19-7 WARNING: shell and server versions do not match switched to db app { "ok" : 1, "$clusterTime" : { "clusterTime" : Timestamp(1773180945, 61), "signature" : { "hash" : BinData(0,"tpt31ejrAN9hPe9+IuZKUT5YiPc="), "keyId" : NumberLong("7615751510080618520") } }, "operationTime" : Timestamp(1773180945, 61) } bye' + echo Percona Server for MongoDB shell version v4.4.29-28 connecting to: 'mongodb://some-name-mongos.data-sharded-19777.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb' '{"t":{"$date":"2026-03-10T22:15:45.240Z"},"s":"I",' '"c":"NETWORK",' '"id":5490002,' '"ctx":"thread1","msg":"Started' a new thread for the timer 'service"}' Implicit session: session '{' '"id"' : 'UUID("0ee5836f-9012-4524-af4b-7985e82b2dca")' '}' Percona Server for MongoDB server version: v8.0.19-7 WARNING: shell and server versions do not match switched to db app '{' '"ok"' : 1, '"$clusterTime"' : '{' '"clusterTime"' : 'Timestamp(1773180945,' '61),' '"signature"' : '{' '"hash"' : 'BinData(0,"tpt31ejrAN9hPe9+IuZKUT5YiPc="),' '"keyId"' : 'NumberLong("7615751510080618520")' '}' '},' '"operationTime"' : 'Timestamp(1773180945,' '61)' '}' bye + grep -q '"ok" : 1' + desc 'check if rs1 and all its related stateful sets are properly removed' + set +o xtrace ----------------------------------------------------------------------------------- check if rs1 and all its related stateful sets are properly removed ----------------------------------------------------------------------------------- + check_rs_proper_component_deletion some-name rs1 + local cluster=some-name + local rs_name=rs1 ++ kubectl_bin get psmdb some-name -ojson ++ jq --arg RS rs1 '.spec.replsets | map(.name == $RS) | index(true)' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ox3xOfCwIf +++ mktemp ++ local LAST_ERR=/tmp/tmp.5i0wvnW1gz ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -ojson ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ox3xOfCwIf ++ cat /tmp/tmp.5i0wvnW1gz ++ rm /tmp/tmp.ox3xOfCwIf /tmp/tmp.5i0wvnW1gz ++ return 0 + rs_idx=1 + kubectl_bin patch psmdb some-name --type=json '-p=[{'\''op'\'': '\''remove'\'', '\''path'\'': '\''/spec/replsets/1'\''}]' ++ mktemp + local LAST_OUT=/tmp/tmp.8S0GwoJ91H ++ mktemp + local LAST_ERR=/tmp/tmp.PJ9lfODKCP + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl patch psmdb some-name --type=json '-p=[{'\''op'\'': '\''remove'\'', '\''path'\'': '\''/spec/replsets/1'\''}]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.8S0GwoJ91H perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.PJ9lfODKCP + rm /tmp/tmp.8S0GwoJ91H /tmp/tmp.PJ9lfODKCP + return 0 + echo -n 'Deleting replset rs1' Deleting replset rs1++ kubectl_bin get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs1 -ojson ++ jq '.items | length' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6Zl4AOf2oi +++ mktemp ++ local LAST_ERR=/tmp/tmp.vsqJdnW2n0 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs1 -ojson ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.6Zl4AOf2oi ++ cat /tmp/tmp.vsqJdnW2n0 ++ rm /tmp/tmp.6Zl4AOf2oi /tmp/tmp.vsqJdnW2n0 ++ return 0 + [[ 1 -eq 0 ]] + let retry+=1 + '[' 1 -ge 70 ']' + echo -n . .+ sleep 30 ++ kubectl_bin get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs1 -ojson ++ jq '.items | length' +++ mktemp ++ local LAST_OUT=/tmp/tmp.v8JrUbEanL +++ mktemp ++ local LAST_ERR=/tmp/tmp.rsDY0WYoLD ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs1 -ojson ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.v8JrUbEanL ++ cat /tmp/tmp.rsDY0WYoLD ++ rm /tmp/tmp.v8JrUbEanL /tmp/tmp.rsDY0WYoLD ++ return 0 + [[ 0 -eq 0 ]] + echo OK OK + desc 'check if rs2 and all its related stateful sets are properly removed' + set +o xtrace ----------------------------------------------------------------------------------- check if rs2 and all its related stateful sets are properly removed ----------------------------------------------------------------------------------- + check_rs_proper_component_deletion some-name rs2 + local cluster=some-name + local rs_name=rs2 ++ kubectl_bin get psmdb some-name -ojson ++ jq --arg RS rs2 '.spec.replsets | map(.name == $RS) | index(true)' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QLn1M3kWcy +++ mktemp ++ local LAST_ERR=/tmp/tmp.gDa3pwxIQ5 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -ojson ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.QLn1M3kWcy ++ cat /tmp/tmp.gDa3pwxIQ5 ++ rm /tmp/tmp.QLn1M3kWcy /tmp/tmp.gDa3pwxIQ5 ++ return 0 + rs_idx=1 + kubectl_bin patch psmdb some-name --type=json '-p=[{'\''op'\'': '\''remove'\'', '\''path'\'': '\''/spec/replsets/1'\''}]' ++ mktemp + local LAST_OUT=/tmp/tmp.mdxtsCcvly ++ mktemp + local LAST_ERR=/tmp/tmp.89iYuMhJxA + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl patch psmdb some-name --type=json '-p=[{'\''op'\'': '\''remove'\'', '\''path'\'': '\''/spec/replsets/1'\''}]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.mdxtsCcvly perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.89iYuMhJxA + rm /tmp/tmp.mdxtsCcvly /tmp/tmp.89iYuMhJxA + return 0 + echo -n 'Deleting replset rs2' Deleting replset rs2++ kubectl_bin get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ jq '.items | length' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lA1kYrKyhl +++ mktemp ++ local LAST_ERR=/tmp/tmp.vl2QMb7ky7 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.lA1kYrKyhl ++ cat /tmp/tmp.vl2QMb7ky7 ++ rm /tmp/tmp.lA1kYrKyhl /tmp/tmp.vl2QMb7ky7 ++ return 0 + [[ 2 -eq 0 ]] + let retry+=1 + '[' 2 -ge 70 ']' + echo -n . .+ sleep 30 ++ kubectl_bin get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ jq '.items | length' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uj2bH8a3aC +++ mktemp ++ local LAST_ERR=/tmp/tmp.YcxvurAAjc ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.uj2bH8a3aC ++ cat /tmp/tmp.YcxvurAAjc ++ rm /tmp/tmp.uj2bH8a3aC /tmp/tmp.YcxvurAAjc ++ return 0 + [[ 2 -eq 0 ]] + let retry+=1 + '[' 3 -ge 70 ']' + echo -n . .+ sleep 30 ++ kubectl_bin get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ jq '.items | length' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nxQzvEcTse +++ mktemp ++ local LAST_ERR=/tmp/tmp.x8iFrGcN6B ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.nxQzvEcTse ++ cat /tmp/tmp.x8iFrGcN6B ++ rm /tmp/tmp.nxQzvEcTse /tmp/tmp.x8iFrGcN6B ++ return 0 + [[ 2 -eq 0 ]] + let retry+=1 + '[' 4 -ge 70 ']' + echo -n . .+ sleep 30 ++ kubectl_bin get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ jq '.items | length' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4VgFN9ut4g +++ mktemp ++ local LAST_ERR=/tmp/tmp.PNBkUbMHl0 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4VgFN9ut4g ++ cat /tmp/tmp.PNBkUbMHl0 ++ rm /tmp/tmp.4VgFN9ut4g /tmp/tmp.PNBkUbMHl0 ++ return 0 + [[ 2 -eq 0 ]] + let retry+=1 + '[' 5 -ge 70 ']' + echo -n . .+ sleep 30 ++ kubectl_bin get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ jq '.items | length' +++ mktemp ++ local LAST_OUT=/tmp/tmp.O5UQHT9DZf +++ mktemp ++ local LAST_ERR=/tmp/tmp.sbJSSqHUc5 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.O5UQHT9DZf ++ cat /tmp/tmp.sbJSSqHUc5 ++ rm /tmp/tmp.O5UQHT9DZf /tmp/tmp.sbJSSqHUc5 ++ return 0 + [[ 2 -eq 0 ]] + let retry+=1 + '[' 6 -ge 70 ']' + echo -n . .+ sleep 30 ++ kubectl_bin get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ jq '.items | length' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zp8tmbM1PP +++ mktemp ++ local LAST_ERR=/tmp/tmp.LOJKAXPyXW ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.zp8tmbM1PP ++ cat /tmp/tmp.LOJKAXPyXW ++ rm /tmp/tmp.zp8tmbM1PP /tmp/tmp.LOJKAXPyXW ++ return 0 + [[ 2 -eq 0 ]] + let retry+=1 + '[' 7 -ge 70 ']' + echo -n . .+ sleep 30 ++ kubectl_bin get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ jq '.items | length' +++ mktemp ++ local LAST_OUT=/tmp/tmp.P1Py231Kxo +++ mktemp ++ local LAST_ERR=/tmp/tmp.YVDNJBqMxN ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.P1Py231Kxo ++ cat /tmp/tmp.YVDNJBqMxN ++ rm /tmp/tmp.P1Py231Kxo /tmp/tmp.YVDNJBqMxN ++ return 0 + [[ 2 -eq 0 ]] + let retry+=1 + '[' 8 -ge 70 ']' + echo -n . .+ sleep 30 ++ kubectl_bin get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ jq '.items | length' +++ mktemp ++ local LAST_OUT=/tmp/tmp.73r1MPsKYN +++ mktemp ++ local LAST_ERR=/tmp/tmp.Fmcvg17dCx ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.73r1MPsKYN ++ cat /tmp/tmp.Fmcvg17dCx ++ rm /tmp/tmp.73r1MPsKYN /tmp/tmp.Fmcvg17dCx ++ return 0 + [[ 2 -eq 0 ]] + let retry+=1 + '[' 9 -ge 70 ']' + echo -n . .+ sleep 30 ++ kubectl_bin get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ jq '.items | length' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qXZD0zutMU +++ mktemp ++ local LAST_ERR=/tmp/tmp.doZGBEE6XZ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.qXZD0zutMU ++ cat /tmp/tmp.doZGBEE6XZ ++ rm /tmp/tmp.qXZD0zutMU /tmp/tmp.doZGBEE6XZ ++ return 0 + [[ 2 -eq 0 ]] + let retry+=1 + '[' 10 -ge 70 ']' + echo -n . .+ sleep 30 ++ kubectl_bin get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ jq '.items | length' +++ mktemp ++ local LAST_OUT=/tmp/tmp.O9S8Gb84BG +++ mktemp ++ local LAST_ERR=/tmp/tmp.6lln8HhVvl ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.O9S8Gb84BG ++ cat /tmp/tmp.6lln8HhVvl ++ rm /tmp/tmp.O9S8Gb84BG /tmp/tmp.6lln8HhVvl ++ return 0 + [[ 2 -eq 0 ]] + let retry+=1 + '[' 11 -ge 70 ']' + echo -n . .+ sleep 30 ++ kubectl_bin get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ jq '.items | length' +++ mktemp ++ local LAST_OUT=/tmp/tmp.henIizUBgH +++ mktemp ++ local LAST_ERR=/tmp/tmp.mWXnOEeWPS ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.henIizUBgH ++ cat /tmp/tmp.mWXnOEeWPS ++ rm /tmp/tmp.henIizUBgH /tmp/tmp.mWXnOEeWPS ++ return 0 + [[ 2 -eq 0 ]] + let retry+=1 + '[' 12 -ge 70 ']' + echo -n . .+ sleep 30 ++ kubectl_bin get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ jq '.items | length' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nV0tMb8Ybm +++ mktemp ++ local LAST_ERR=/tmp/tmp.jm9m5Fvfwe ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.nV0tMb8Ybm ++ cat /tmp/tmp.jm9m5Fvfwe ++ rm /tmp/tmp.nV0tMb8Ybm /tmp/tmp.jm9m5Fvfwe ++ return 0 + [[ 2 -eq 0 ]] + let retry+=1 + '[' 13 -ge 70 ']' + echo -n . .+ sleep 30 ++ kubectl_bin get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ jq '.items | length' +++ mktemp ++ local LAST_OUT=/tmp/tmp.r1E2C9Mo0Z +++ mktemp ++ local LAST_ERR=/tmp/tmp.aDB9VuDJAI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.r1E2C9Mo0Z ++ cat /tmp/tmp.aDB9VuDJAI ++ rm /tmp/tmp.r1E2C9Mo0Z /tmp/tmp.aDB9VuDJAI ++ return 0 + [[ 2 -eq 0 ]] + let retry+=1 + '[' 14 -ge 70 ']' + echo -n . .+ sleep 30 ++ kubectl_bin get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ jq '.items | length' +++ mktemp ++ local LAST_OUT=/tmp/tmp.X9BkKmIVTe +++ mktemp ++ local LAST_ERR=/tmp/tmp.WP02xotdZt ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.X9BkKmIVTe ++ cat /tmp/tmp.WP02xotdZt ++ rm /tmp/tmp.X9BkKmIVTe /tmp/tmp.WP02xotdZt ++ return 0 + [[ 1 -eq 0 ]] + let retry+=1 + '[' 15 -ge 70 ']' + echo -n . .+ sleep 30 ++ kubectl_bin get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ jq '.items | length' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rPhtcmtoJK +++ mktemp ++ local LAST_ERR=/tmp/tmp.SD7UQEzDzp ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get sts -l app.kubernetes.io/instance=some-name,app.kubernetes.io/replset=rs2 -ojson ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.rPhtcmtoJK ++ cat /tmp/tmp.SD7UQEzDzp ++ rm /tmp/tmp.rPhtcmtoJK /tmp/tmp.SD7UQEzDzp ++ return 0 + [[ 0 -eq 0 ]] + echo OK OK + destroy data-sharded-19777 + local namespace=data-sharded-19777 + local ignore_logs=true + [[ 0 == 1 ]] + desc 'destroy cluster/operator and all other resources' + set +o xtrace ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + '[' true == false ']' + delete_backups + desc 'Delete psmdb-backup' + set +o xtrace ----------------------------------------------------------------------------------- Delete psmdb-backup ----------------------------------------------------------------------------------- ++ kubectl_bin get psmdb-backup --no-headers ++ wc -l +++ mktemp ++ local LAST_OUT=/tmp/tmp.j76iIYSvz8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.obC4v8xx2s ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb-backup --no-headers ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.j76iIYSvz8 ++ cat /tmp/tmp.obC4v8xx2s No resources found in data-sharded-19777 namespace. ++ rm /tmp/tmp.j76iIYSvz8 /tmp/tmp.obC4v8xx2s ++ return 0 + '[' 0 '!=' 0 ']' + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.9zgqUhm5LM ++ mktemp + local LAST_ERR=/tmp/tmp.58el3C9AA4 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.9zgqUhm5LM customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.58el3C9AA4 + rm /tmp/tmp.9zgqUhm5LM /tmp/tmp.58el3C9AA4 + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/crd.yaml ++ grep -v '\-\-\-' grep: warning: stray \ before - grep: warning: stray \ before - + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.6nKeNjHxsR ++ mktemp + local LAST_ERR=/tmp/tmp.a1thkc76Va + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.6nKeNjHxsR + cat /tmp/tmp.a1thkc76Va + rm /tmp/tmp.6nKeNjHxsR /tmp/tmp.a1thkc76Va + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.kdCKuyeaGK ++ mktemp + local LAST_ERR=/tmp/tmp.6BO1XP0gRG + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.kdCKuyeaGK + cat /tmp/tmp.6BO1XP0gRG + rm /tmp/tmp.kdCKuyeaGK /tmp/tmp.6BO1XP0gRG + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbs" + kubectl patch perconaservermongodbs.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbs" + : + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.EC4gxBG9lb ++ mktemp + local LAST_ERR=/tmp/tmp.CJnpcmjDvr + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.EC4gxBG9lb + cat /tmp/tmp.CJnpcmjDvr + rm /tmp/tmp.EC4gxBG9lb /tmp/tmp.CJnpcmjDvr + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.ls4gVSX9BH ++ mktemp + local LAST_ERR=/tmp/tmp.h2jE5T9c97 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2219/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ls4gVSX9BH clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.h2jE5T9c97 + rm /tmp/tmp.ls4gVSX9BH /tmp/tmp.h2jE5T9c97 + return 0 + destroy_cert_manager + kubectl_bin delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.aqyfBmVCsg ++ mktemp + local LAST_ERR=/tmp/tmp.rRvV3VuKjq + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.aqyfBmVCsg namespace "cert-manager" deleted customresourcedefinition.apiextensions.k8s.io "challenges.acme.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "orders.acme.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "certificaterequests.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "certificates.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "clusterissuers.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "issuers.cert-manager.io" deleted serviceaccount "cert-manager-cainjector" deleted from cert-manager namespace serviceaccount "cert-manager" deleted from cert-manager namespace serviceaccount "cert-manager-webhook" deleted from cert-manager namespace clusterrole.rbac.authorization.k8s.io "cert-manager-cainjector" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-issuers" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-certificates" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-orders" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-challenges" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-cluster-view" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-view" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-edit" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-cainjector" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-issuers" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-certificates" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-orders" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-challenges" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" deleted role.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" deleted from kube-system namespace role.rbac.authorization.k8s.io "cert-manager:leaderelection" deleted from kube-system namespace rolebinding.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" deleted from kube-system namespace rolebinding.rbac.authorization.k8s.io "cert-manager:leaderelection" deleted from kube-system namespace mutatingwebhookconfiguration.admissionregistration.k8s.io "cert-manager-webhook" deleted validatingwebhookconfiguration.admissionregistration.k8s.io "cert-manager-webhook" deleted + cat /tmp/tmp.rRvV3VuKjq Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found + sleep 0 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.aqyfBmVCsg namespace "cert-manager" deleted + cat /tmp/tmp.rRvV3VuKjq Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 4 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.aqyfBmVCsg + cat /tmp/tmp.rRvV3VuKjq Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 8 + cat /tmp/tmp.aqyfBmVCsg + cat /tmp/tmp.rRvV3VuKjq Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + rm /tmp/tmp.aqyfBmVCsg /tmp/tmp.rRvV3VuKjq + return 1 + true + '[' -n '' ']' + '[' -n psmdb-operator ']' + kubectl_bin delete --grace-period=0 --force=true namespace data-sharded-19777 + rm -rf /tmp/tmp.m6dyp9GRir + kubectl_bin delete --grace-period=0 --force=true namespace psmdb-operator ++ mktemp ++ mktemp + desc 'test passed' + set +o xtrace ----------------------------------------------------------------------------------- test passed ----------------------------------------------------------------------------------- + local LAST_OUT=/tmp/tmp.CvpgwssrjQ ++ mktemp + local LAST_OUT=/tmp/tmp.QSnv8kkiQN ++ mktemp + local LAST_ERR=/tmp/tmp.knuPfb2U1Q + local exit_status=0 + local timeout=4 ++ seq 0 2 + local LAST_ERR=/tmp/tmp.nc6V3o7iNd + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete --grace-period=0 --force=true namespace data-sharded-19777 + for i in $(seq 0 2) + set +e + kubectl delete --grace-period=0 --force=true namespace psmdb-operator