++ echo 'Log: /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/logs/mongod-major-upgrade-sharded.log' Log: /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/logs/mongod-major-upgrade-sharded.log ++ '[' -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/conf/cloud-secret.yml ']' ++ SKIP_BACKUPS_TO_AWS_GCP_AZURE= ++ oc get projects ++ kubectl get nodes ++ grep '^minikube' +++ kubectl version -o json +++ jq -r .serverVersion.gitVersion +++ grep -eks- Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 ++ '[' ']' ++ EKS=0 +++ kubectl version -o json +++ jq -r .serverVersion.gitVersion +++ grep gke Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 ++ '[' v1.32.12-gke.1127000 ']' ++ GKE=1 +++ kubectl version -o json +++ jq -r '.serverVersion.major + "." + .serverVersion.minor' +++ /usr/sbin/sed -r 's/[^0-9.]+//g' Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 ++ KUBE_VERSION=1.32 + set_debug + [[ 1 == 1 ]] + set -o xtrace + main + rbac=rbac + '[' -n psmdb-operator ']' + rbac=cw-rbac + create_infra mongod-major-upgrade-sharded-7355 + local ns=mongod-major-upgrade-sharded-7355 + [[ 1 == 1 ]] + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.YlH5FSRl83 ++ mktemp + local LAST_ERR=/tmp/tmp.ZRKape1op2 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.YlH5FSRl83 customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.ZRKape1op2 + rm /tmp/tmp.YlH5FSRl83 /tmp/tmp.ZRKape1op2 + return 0 ++ grep -v '\-\-\-' ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/crd.yaml grep: warning: stray \ before - grep: warning: stray \ before - + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + grep -v NAMESPACE + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.bLgK4CWxzZ ++ mktemp + local LAST_ERR=/tmp/tmp.7mrxjjDTO4 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.bLgK4CWxzZ + cat /tmp/tmp.7mrxjjDTO4 + rm /tmp/tmp.bLgK4CWxzZ /tmp/tmp.7mrxjjDTO4 + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.gLFzumvvXe ++ mktemp + local LAST_ERR=/tmp/tmp.sORsnJRYe8 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.gLFzumvvXe + cat /tmp/tmp.sORsnJRYe8 + rm /tmp/tmp.gLFzumvvXe /tmp/tmp.sORsnJRYe8 + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbs" + kubectl patch perconaservermongodbs.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbs" + : + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.fMsFcAxjLv ++ mktemp + local LAST_ERR=/tmp/tmp.DKfoeDpZJw + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.fMsFcAxjLv + cat /tmp/tmp.DKfoeDpZJw + rm /tmp/tmp.fMsFcAxjLv /tmp/tmp.DKfoeDpZJw + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.uAhi8Ndb5M ++ mktemp + local LAST_ERR=/tmp/tmp.9Aqvhejb5k + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.uAhi8Ndb5M clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.9Aqvhejb5k + rm /tmp/tmp.uAhi8Ndb5M /tmp/tmp.9Aqvhejb5k + return 0 + check_crd_for_deletion PR-2266-f9fc55604 + local git_tag=PR-2266-f9fc55604 ++ curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/PR-2266-f9fc55604/deploy/crd.yaml ++ yq eval .metadata.name ++ /usr/sbin/sed ':a;N;$!ba;s/\n/ /g' ++ /usr/sbin/sed s/---//g + for crd_name in $(curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/${git_tag}/deploy/crd.yaml | yq eval '.metadata.name' | $sed 's/---//g' | $sed ':a;N;$!ba;s/\n/ /g') ++ kubectl_bin get crd/null -o 'jsonpath={.status.conditions[-1].type}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NQPAfqthfW +++ mktemp ++ local LAST_ERR=/tmp/tmp.aLgd1Da2ZH ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.NQPAfqthfW ++ cat /tmp/tmp.aLgd1Da2ZH Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 0 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.NQPAfqthfW ++ cat /tmp/tmp.aLgd1Da2ZH Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 4 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.NQPAfqthfW ++ cat /tmp/tmp.aLgd1Da2ZH Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 8 ++ cat /tmp/tmp.NQPAfqthfW ++ cat /tmp/tmp.aLgd1Da2ZH Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ rm /tmp/tmp.NQPAfqthfW /tmp/tmp.aLgd1Da2ZH ++ return 1 + [[ '' == \T\e\r\m\i\n\a\t\i\n\g ]] + '[' -n psmdb-operator ']' + create_namespace psmdb-operator + local namespace=psmdb-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl api-resources ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + grep -E -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' + '[' -n '' ']' + desc 'cleaned up old namespaces psmdb-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace psmdb-operator --ignore-not-found + awk '{print$1}' + xargs kubectl delete ns ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.cLSvAbXrXy ++ mktemp + local LAST_ERR=/tmp/tmp.BQWrl25N6R + local exit_status=0 + local timeout=4 ++ seq 0 2 + local LAST_OUT=/tmp/tmp.XEjb6CWUhi + for i in $(seq 0 2) + set +e + kubectl delete namespace psmdb-operator --ignore-not-found ++ mktemp + local LAST_ERR=/tmp/tmp.wwZ8aoUNtS + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get ns + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.XEjb6CWUhi + cat /tmp/tmp.wwZ8aoUNtS + rm /tmp/tmp.XEjb6CWUhi /tmp/tmp.wwZ8aoUNtS + return 0 namespace "mongod-major-upgrade-sharded-11828" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.cLSvAbXrXy namespace "psmdb-operator" deleted + cat /tmp/tmp.BQWrl25N6R + rm /tmp/tmp.cLSvAbXrXy /tmp/tmp.BQWrl25N6R + return 0 + kubectl_bin wait --for=delete namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.DqmDYDcAXR ++ mktemp + local LAST_ERR=/tmp/tmp.8eD1t1nAD5 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.DqmDYDcAXR + cat /tmp/tmp.8eD1t1nAD5 + rm /tmp/tmp.DqmDYDcAXR /tmp/tmp.8eD1t1nAD5 + return 0 + desc 'create namespace psmdb-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.PiGhTySJB3 ++ mktemp + local LAST_ERR=/tmp/tmp.Y0dyiDEqg5 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.PiGhTySJB3 namespace/psmdb-operator created + cat /tmp/tmp.Y0dyiDEqg5 + rm /tmp/tmp.PiGhTySJB3 /tmp/tmp.Y0dyiDEqg5 + return 0 + set_kube_ctx psmdb-operator + local namespace=psmdb-operator ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.EsnLRbu4S3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.bnBwPdUTF3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.EsnLRbu4S3 ++ cat /tmp/tmp.bnBwPdUTF3 ++ rm /tmp/tmp.EsnLRbu4S3 /tmp/tmp.bnBwPdUTF3 ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2266-f9fc55604-9-cluster4 --namespace=psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.urtqxfIoe8 ++ mktemp + local LAST_ERR=/tmp/tmp.Y3dN1D0jJ3 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2266-f9fc55604-9-cluster4 --namespace=psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.urtqxfIoe8 Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2266-f9fc55604-9-cluster4" modified. + cat /tmp/tmp.Y3dN1D0jJ3 + rm /tmp/tmp.urtqxfIoe8 /tmp/tmp.Y3dN1D0jJ3 + return 0 + deploy_operator + desc 'start PSMDB operator: docker.io/perconalab/percona-server-mongodb-operator:PR-2266-f9fc55604' + set +o xtrace ----------------------------------------------------------------------------------- start PSMDB operator: docker.io/perconalab/percona-server-mongodb-operator:PR-2266-f9fc55604 ----------------------------------------------------------------------------------- + local cr_file + '[' -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/mongod-major-upgrade-sharded/conf/crd.yaml ']' + cr_file=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/crd.yaml + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.W9jalpLCI6 ++ mktemp + local LAST_ERR=/tmp/tmp.qRU23l0WnK + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.W9jalpLCI6 customresourcedefinition.apiextensions.k8s.io/perconaservermongodbbackups.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbrestores.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com serverside-applied + cat /tmp/tmp.qRU23l0WnK + rm /tmp/tmp.W9jalpLCI6 /tmp/tmp.qRU23l0WnK + return 0 + '[' -n psmdb-operator ']' + apply_rbac cw-rbac + local operator_namespace=psmdb-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: psmdb-operator^' + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.cxHXbU1a0E ++ mktemp + local LAST_ERR=/tmp/tmp.8zOdGMMn73 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.cxHXbU1a0E clusterrole.rbac.authorization.k8s.io/percona-server-mongodb-operator created serviceaccount/percona-server-mongodb-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-server-mongodb-operator created + cat /tmp/tmp.8zOdGMMn73 + rm /tmp/tmp.cxHXbU1a0E /tmp/tmp.8zOdGMMn73 + return 0 + yq eval ' (.spec.template.spec.containers[].image = "docker.io/perconalab/percona-server-mongodb-operator:PR-2266-f9fc55604") | ((.. | select(.[] == "DISABLE_TELEMETRY")) |= .value="true") | ((.. | select(.[] == "LOG_LEVEL")) |= .value="DEBUG")' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/cw-operator.yaml + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.cp5aQU13mU ++ mktemp + local LAST_ERR=/tmp/tmp.EQpX2IAfIW + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.cp5aQU13mU deployment.apps/percona-server-mongodb-operator created + cat /tmp/tmp.EQpX2IAfIW + rm /tmp/tmp.cp5aQU13mU /tmp/tmp.EQpX2IAfIW + return 0 + sleep 20 ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.LxSbQOaWLy +++ mktemp ++ local LAST_ERR=/tmp/tmp.WAmCZdp58v ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.LxSbQOaWLy ++ cat /tmp/tmp.WAmCZdp58v ++ rm /tmp/tmp.LxSbQOaWLy /tmp/tmp.WAmCZdp58v ++ return 0 + wait_operator_pod percona-server-mongodb-operator-5554c4df68-sn28n + local pod=percona-server-mongodb-operator-5554c4df68-sn28n + set +o xtrace waiting for pod/percona-server-mongodb-operator-5554c4df68-sn28n to be ready.OK + echo 'Print operator info from log' Print operator info from log ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp + grep 'Manager starting up' ++ local LAST_OUT=/tmp/tmp.YlKVpJzquH +++ mktemp ++ local LAST_ERR=/tmp/tmp.XumBN9luKH ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.YlKVpJzquH ++ cat /tmp/tmp.XumBN9luKH ++ rm /tmp/tmp.YlKVpJzquH /tmp/tmp.XumBN9luKH ++ return 0 + kubectl_bin logs -n psmdb-operator percona-server-mongodb-operator-5554c4df68-sn28n ++ mktemp + local LAST_OUT=/tmp/tmp.4cYexnrFQh ++ mktemp + local LAST_ERR=/tmp/tmp.mgRJe0RTuH + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl logs -n psmdb-operator percona-server-mongodb-operator-5554c4df68-sn28n + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.4cYexnrFQh + cat /tmp/tmp.mgRJe0RTuH + rm /tmp/tmp.4cYexnrFQh /tmp/tmp.mgRJe0RTuH + return 0 2026-03-10T07:28:42.857Z INFO setup Manager starting up {"gitCommit": "f9fc5560468cd2df79998ec06ba830d1104a31a3", "gitBranch": "PR-2266-f9fc55604", "buildTime": "", "goVersion": "go1.25.8", "os": "linux", "arch": "amd64"} + create_namespace mongod-major-upgrade-sharded-7355 + local namespace=mongod-major-upgrade-sharded-7355 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get ValidatingWebhookConfiguration + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + grep -E -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' + '[' -n '' ']' + desc 'cleaned up old namespaces mongod-major-upgrade-sharded-7355' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces mongod-major-upgrade-sharded-7355 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace mongod-major-upgrade-sharded-7355 --ignore-not-found + awk '{print$1}' + xargs kubectl delete ns ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.tOoS3qCzsb ++ mktemp + local LAST_ERR=/tmp/tmp.wBGyZKABjo + local exit_status=0 + local timeout=4 ++ seq 0 2 + local LAST_OUT=/tmp/tmp.C3xvCe5efx + for i in $(seq 0 2) + set +e + kubectl get ns ++ mktemp + local LAST_ERR=/tmp/tmp.X4pZVuGJrf + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete namespace mongod-major-upgrade-sharded-7355 --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.tOoS3qCzsb + cat /tmp/tmp.wBGyZKABjo + rm /tmp/tmp.tOoS3qCzsb /tmp/tmp.wBGyZKABjo + return 0 error: resource(s) were provided, but no name was specified + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.C3xvCe5efx + cat /tmp/tmp.X4pZVuGJrf + rm /tmp/tmp.C3xvCe5efx /tmp/tmp.X4pZVuGJrf + return 0 + kubectl_bin wait --for=delete namespace mongod-major-upgrade-sharded-7355 ++ mktemp + local LAST_OUT=/tmp/tmp.A77FIaeGy1 ++ mktemp + local LAST_ERR=/tmp/tmp.QaeYyucbYg + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete namespace mongod-major-upgrade-sharded-7355 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.A77FIaeGy1 + cat /tmp/tmp.QaeYyucbYg + rm /tmp/tmp.A77FIaeGy1 /tmp/tmp.QaeYyucbYg + return 0 + desc 'create namespace mongod-major-upgrade-sharded-7355' + set +o xtrace ----------------------------------------------------------------------------------- create namespace mongod-major-upgrade-sharded-7355 ----------------------------------------------------------------------------------- + kubectl_bin create namespace mongod-major-upgrade-sharded-7355 ++ mktemp + local LAST_OUT=/tmp/tmp.lkIk6u3ukn ++ mktemp + local LAST_ERR=/tmp/tmp.UQVsqZszFl + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace mongod-major-upgrade-sharded-7355 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.lkIk6u3ukn namespace/mongod-major-upgrade-sharded-7355 created + cat /tmp/tmp.UQVsqZszFl + rm /tmp/tmp.lkIk6u3ukn /tmp/tmp.UQVsqZszFl + return 0 + set_kube_ctx mongod-major-upgrade-sharded-7355 + local namespace=mongod-major-upgrade-sharded-7355 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.HFeAeoh2k4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Pmq9iz8aNa ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.HFeAeoh2k4 ++ cat /tmp/tmp.Pmq9iz8aNa ++ rm /tmp/tmp.HFeAeoh2k4 /tmp/tmp.Pmq9iz8aNa ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2266-f9fc55604-9-cluster4 --namespace=mongod-major-upgrade-sharded-7355 ++ mktemp + local LAST_OUT=/tmp/tmp.GYa9LGHytf ++ mktemp + local LAST_ERR=/tmp/tmp.OE1eNfdKjf + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2266-f9fc55604-9-cluster4 --namespace=mongod-major-upgrade-sharded-7355 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.GYa9LGHytf Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2266-f9fc55604-9-cluster4" modified. + cat /tmp/tmp.OE1eNfdKjf + rm /tmp/tmp.GYa9LGHytf /tmp/tmp.OE1eNfdKjf + return 0 + apply_s3_storage_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.1N1tGy51ls ++ mktemp + local LAST_ERR=/tmp/tmp.J4i8TqQach + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.1N1tGy51ls secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created secret/gcp-cs-sa-key-secret created + cat /tmp/tmp.J4i8TqQach + rm /tmp/tmp.1N1tGy51ls /tmp/tmp.J4i8TqQach + return 0 + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/conf/client.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.UKODQBXNFh ++ mktemp + local LAST_ERR=/tmp/tmp.RYHHwur7DW + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/conf/client.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/conf/secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.UKODQBXNFh deployment.apps/psmdb-client created secret/some-users created + cat /tmp/tmp.RYHHwur7DW + rm /tmp/tmp.UKODQBXNFh /tmp/tmp.RYHHwur7DW + return 0 + desc 'install version service' + set +o xtrace ----------------------------------------------------------------------------------- install version service ----------------------------------------------------------------------------------- + cp /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/mongod-major-upgrade-sharded/conf/operator.main.psmdb-operator.dep.json /tmp/tmp.out8E8NpB2/operator.1.23.0.psmdb-operator.dep.json + generate_vs_json /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/mongod-major-upgrade-sharded/conf/operator.main.psmdb-operator.json /tmp/tmp.out8E8NpB2/operator.1.23.0.psmdb-operator.json + local template_path=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/mongod-major-upgrade-sharded/conf/operator.main.psmdb-operator.json + local target_path=/tmp/tmp.out8E8NpB2/operator.1.23.0.psmdb-operator.json ++ jq '.versions[0].operator="1.23.0"' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/mongod-major-upgrade-sharded/conf/operator.main.psmdb-operator.json + local 'version_service_source={ "versions": [ { "operator": "1.23.0", "product": "psmdb-operator", "matrix": { "mongod": {}, "pmm": { "2.27.0": { "image_path": "percona/pmm-client:2.27.0", "image_hash": "4f4d4508afe4ef3b5d72e2ebec6485be84204902b0b2b23c3a9e7c2fe4726884", "status": "recommended", "critical": false } }, "backup": {}, "operator": {} } } ] }' + for image_mongod in ${IMAGE_MONGOD_CHAIN[@]} ++ get_mongod_ver_from_image docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0 ++ local image=docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0 +++ run_simple_cli_inside_image docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0 'mongod --version' +++ local image=docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0 +++ local 'cli=mongod --version' +++ local pod_name=29190 +++ kubectl_bin -n default run 29190 --image=docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0 --restart=Never --command -- sleep infinity +++ /usr/sbin/sed -r 's/^.*db version v(([0-9]+\.){2}[0-9]+-[0-9]+).*$/\1/g' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.84SUC2TtcL ++++ mktemp +++ local LAST_ERR=/tmp/tmp.j4m07XnhMM +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl -n default run 29190 --image=docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0 --restart=Never --command -- sleep infinity +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.84SUC2TtcL +++ cat /tmp/tmp.j4m07XnhMM +++ rm /tmp/tmp.84SUC2TtcL /tmp/tmp.j4m07XnhMM +++ return 0 +++ kubectl_bin -n default wait --for=condition=Ready pod/29190 ++++ mktemp +++ local LAST_OUT=/tmp/tmp.CxphBVJiaJ ++++ mktemp +++ local LAST_ERR=/tmp/tmp.igawPHtl7z +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl -n default wait --for=condition=Ready pod/29190 +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.CxphBVJiaJ +++ cat /tmp/tmp.igawPHtl7z +++ rm /tmp/tmp.CxphBVJiaJ /tmp/tmp.igawPHtl7z +++ return 0 ++++ kubectl_bin -n default exec 29190 -- bash -c 'mongod --version 2>&1' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.91Pmj6G4sM +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.NFKpaCIkyK ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl -n default exec 29190 -- bash -c 'mongod --version 2>&1' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.91Pmj6G4sM ++++ cat /tmp/tmp.NFKpaCIkyK ++++ rm /tmp/tmp.91Pmj6G4sM /tmp/tmp.NFKpaCIkyK ++++ return 0 +++ local 'output=db version v6.0.27-21 Build Info: { "version": "6.0.27-21", "gitVersion": "7b47bbb85ce39c999dc6f02ec3ebe4c45644e5b4", "openSSLVersion": "OpenSSL 3.5.1 1 Jul 2025", "modules": [], "perconaFeatures": [ "MemoryEngine", "HotBackup", "BackupCursorAggregationStage", "BackupCursorExtendAggregationStage", "AWSIAM", "Kerberos", "LDAP", "TDE", "FIPSMode", "Auditing", "ProfilingRateLimit", "LogRedaction", "ngram" ], "allocator": "tcmalloc", "environment": { "distarch": "x86_64", "target_arch": "x86_64" } }' +++ kubectl_bin -n default delete pod/29190 --grace-period=0 --force ++++ mktemp +++ local LAST_OUT=/tmp/tmp.SY5WyC5983 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.irJPss9pWQ +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl -n default delete pod/29190 --grace-period=0 --force +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.SY5WyC5983 +++ cat /tmp/tmp.irJPss9pWQ Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. +++ rm /tmp/tmp.SY5WyC5983 /tmp/tmp.irJPss9pWQ +++ return 0 +++ echo db version v6.0.27-21 Build Info: '{' '"version":' '"6.0.27-21",' '"gitVersion":' '"7b47bbb85ce39c999dc6f02ec3ebe4c45644e5b4",' '"openSSLVersion":' '"OpenSSL' 3.5.1 1 Jul '2025",' '"modules":' '[],' '"perconaFeatures":' '[' '"MemoryEngine",' '"HotBackup",' '"BackupCursorAggregationStage",' '"BackupCursorExtendAggregationStage",' '"AWSIAM",' '"Kerberos",' '"LDAP",' '"TDE",' '"FIPSMode",' '"Auditing",' '"ProfilingRateLimit",' '"LogRedaction",' '"ngram"' '],' '"allocator":' '"tcmalloc",' '"environment":' '{' '"distarch":' '"x86_64",' '"target_arch":' '"x86_64"' '}' '}' ++ version_info=6.0.27-21 ++ [[ ! 6.0.27-21 =~ ^([0-9]+\.){2}[0-9]+-[0-9]+$ ]] ++ echo 6.0.27-21 + current_mongod_version=6.0.27-21 ++ echo '{' '"versions":' '[' '{' '"operator":' '"1.23.0",' '"product":' '"psmdb-operator",' '"matrix":' '{' '"mongod":' '{},' '"pmm":' '{' '"2.27.0":' '{' '"image_path":' '"percona/pmm-client:2.27.0",' '"image_hash":' '"4f4d4508afe4ef3b5d72e2ebec6485be84204902b0b2b23c3a9e7c2fe4726884",' '"status":' '"recommended",' '"critical":' false '}' '},' '"backup":' '{},' '"operator":' '{}' '}' '}' ']' '}' ++ jq '.versions[0].matrix.mongod += {"6.0.27-21": {"image_path":"docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0","status":"recommended"}}' + version_service_source='{ "versions": [ { "operator": "1.23.0", "product": "psmdb-operator", "matrix": { "mongod": { "6.0.27-21": { "image_path": "docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0", "status": "recommended" } }, "pmm": { "2.27.0": { "image_path": "percona/pmm-client:2.27.0", "image_hash": "4f4d4508afe4ef3b5d72e2ebec6485be84204902b0b2b23c3a9e7c2fe4726884", "status": "recommended", "critical": false } }, "backup": {}, "operator": {} } } ] }' + for image_mongod in ${IMAGE_MONGOD_CHAIN[@]} ++ get_mongod_ver_from_image docker.io/perconalab/percona-server-mongodb-operator:main-mongod7.0 ++ local image=docker.io/perconalab/percona-server-mongodb-operator:main-mongod7.0 +++ run_simple_cli_inside_image docker.io/perconalab/percona-server-mongodb-operator:main-mongod7.0 'mongod --version' +++ local image=docker.io/perconalab/percona-server-mongodb-operator:main-mongod7.0 +++ local 'cli=mongod --version' +++ local pod_name=31034 +++ kubectl_bin -n default run 31034 --image=docker.io/perconalab/percona-server-mongodb-operator:main-mongod7.0 --restart=Never --command -- sleep infinity +++ /usr/sbin/sed -r 's/^.*db version v(([0-9]+\.){2}[0-9]+-[0-9]+).*$/\1/g' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.wYUzGjXrsx ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ko4i1whV8A +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl -n default run 31034 --image=docker.io/perconalab/percona-server-mongodb-operator:main-mongod7.0 --restart=Never --command -- sleep infinity +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.wYUzGjXrsx +++ cat /tmp/tmp.ko4i1whV8A +++ rm /tmp/tmp.wYUzGjXrsx /tmp/tmp.ko4i1whV8A +++ return 0 +++ kubectl_bin -n default wait --for=condition=Ready pod/31034 ++++ mktemp +++ local LAST_OUT=/tmp/tmp.dGVIjjo83R ++++ mktemp +++ local LAST_ERR=/tmp/tmp.vmyEa8uE3q +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl -n default wait --for=condition=Ready pod/31034 +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.dGVIjjo83R +++ cat /tmp/tmp.vmyEa8uE3q +++ rm /tmp/tmp.dGVIjjo83R /tmp/tmp.vmyEa8uE3q +++ return 0 ++++ kubectl_bin -n default exec 31034 -- bash -c 'mongod --version 2>&1' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.TeH5ox1tEQ +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.umZKKpNuF3 ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl -n default exec 31034 -- bash -c 'mongod --version 2>&1' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.TeH5ox1tEQ ++++ cat /tmp/tmp.umZKKpNuF3 ++++ rm /tmp/tmp.TeH5ox1tEQ /tmp/tmp.umZKKpNuF3 ++++ return 0 +++ local 'output=db version v7.0.30-16 Build Info: { "version": "7.0.30-16", "gitVersion": "cdfac0aedc386df60083285302f18ecef4c5382e", "openSSLVersion": "OpenSSL 3.5.1 1 Jul 2025", "modules": [], "perconaFeatures": [ "MemoryEngine", "HotBackup", "BackupCursorAggregationStage", "BackupCursorExtendAggregationStage", "AWSIAM", "Kerberos", "LDAP", "OIDC", "TDE", "FIPSMode", "FCBIS", "Auditing", "ProfilingRateLimit", "LogRedaction", "ngram" ], "allocator": "tcmalloc", "environment": { "distarch": "x86_64", "target_arch": "x86_64" } }' +++ kubectl_bin -n default delete pod/31034 --grace-period=0 --force ++++ mktemp +++ local LAST_OUT=/tmp/tmp.MRHOSWyqGE ++++ mktemp +++ local LAST_ERR=/tmp/tmp.5fj3i93I7f +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl -n default delete pod/31034 --grace-period=0 --force +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.MRHOSWyqGE +++ cat /tmp/tmp.5fj3i93I7f Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. +++ rm /tmp/tmp.MRHOSWyqGE /tmp/tmp.5fj3i93I7f +++ return 0 +++ echo db version v7.0.30-16 Build Info: '{' '"version":' '"7.0.30-16",' '"gitVersion":' '"cdfac0aedc386df60083285302f18ecef4c5382e",' '"openSSLVersion":' '"OpenSSL' 3.5.1 1 Jul '2025",' '"modules":' '[],' '"perconaFeatures":' '[' '"MemoryEngine",' '"HotBackup",' '"BackupCursorAggregationStage",' '"BackupCursorExtendAggregationStage",' '"AWSIAM",' '"Kerberos",' '"LDAP",' '"OIDC",' '"TDE",' '"FIPSMode",' '"FCBIS",' '"Auditing",' '"ProfilingRateLimit",' '"LogRedaction",' '"ngram"' '],' '"allocator":' '"tcmalloc",' '"environment":' '{' '"distarch":' '"x86_64",' '"target_arch":' '"x86_64"' '}' '}' ++ version_info=7.0.30-16 ++ [[ ! 7.0.30-16 =~ ^([0-9]+\.){2}[0-9]+-[0-9]+$ ]] ++ echo 7.0.30-16 + current_mongod_version=7.0.30-16 ++ echo '{' '"versions":' '[' '{' '"operator":' '"1.23.0",' '"product":' '"psmdb-operator",' '"matrix":' '{' '"mongod":' '{' '"6.0.27-21":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0",' '"status":' '"recommended"' '}' '},' '"pmm":' '{' '"2.27.0":' '{' '"image_path":' '"percona/pmm-client:2.27.0",' '"image_hash":' '"4f4d4508afe4ef3b5d72e2ebec6485be84204902b0b2b23c3a9e7c2fe4726884",' '"status":' '"recommended",' '"critical":' false '}' '},' '"backup":' '{},' '"operator":' '{}' '}' '}' ']' '}' ++ jq '.versions[0].matrix.mongod += {"7.0.30-16": {"image_path":"docker.io/perconalab/percona-server-mongodb-operator:main-mongod7.0","status":"recommended"}}' + version_service_source='{ "versions": [ { "operator": "1.23.0", "product": "psmdb-operator", "matrix": { "mongod": { "6.0.27-21": { "image_path": "docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0", "status": "recommended" }, "7.0.30-16": { "image_path": "docker.io/perconalab/percona-server-mongodb-operator:main-mongod7.0", "status": "recommended" } }, "pmm": { "2.27.0": { "image_path": "percona/pmm-client:2.27.0", "image_hash": "4f4d4508afe4ef3b5d72e2ebec6485be84204902b0b2b23c3a9e7c2fe4726884", "status": "recommended", "critical": false } }, "backup": {}, "operator": {} } } ] }' + for image_mongod in ${IMAGE_MONGOD_CHAIN[@]} ++ get_mongod_ver_from_image docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0 ++ local image=docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0 +++ run_simple_cli_inside_image docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0 'mongod --version' +++ /usr/sbin/sed -r 's/^.*db version v(([0-9]+\.){2}[0-9]+-[0-9]+).*$/\1/g' +++ local image=docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0 +++ local 'cli=mongod --version' +++ local pod_name=16983 +++ kubectl_bin -n default run 16983 --image=docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0 --restart=Never --command -- sleep infinity ++++ mktemp +++ local LAST_OUT=/tmp/tmp.UVpYAI7nPd ++++ mktemp +++ local LAST_ERR=/tmp/tmp.MBMEc5QAzA +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl -n default run 16983 --image=docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0 --restart=Never --command -- sleep infinity +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.UVpYAI7nPd +++ cat /tmp/tmp.MBMEc5QAzA +++ rm /tmp/tmp.UVpYAI7nPd /tmp/tmp.MBMEc5QAzA +++ return 0 +++ kubectl_bin -n default wait --for=condition=Ready pod/16983 ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Ng9p1CISCA ++++ mktemp +++ local LAST_ERR=/tmp/tmp.wexS8nzk9Q +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl -n default wait --for=condition=Ready pod/16983 +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.Ng9p1CISCA +++ cat /tmp/tmp.wexS8nzk9Q +++ rm /tmp/tmp.Ng9p1CISCA /tmp/tmp.wexS8nzk9Q +++ return 0 ++++ kubectl_bin -n default exec 16983 -- bash -c 'mongod --version 2>&1' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.fe3bDZr6pC +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.QzoagU5CCA ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl -n default exec 16983 -- bash -c 'mongod --version 2>&1' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.fe3bDZr6pC ++++ cat /tmp/tmp.QzoagU5CCA ++++ rm /tmp/tmp.fe3bDZr6pC /tmp/tmp.QzoagU5CCA ++++ return 0 +++ local 'output=db version v8.0.19-7 Build Info: { "version": "8.0.19-7", "gitVersion": "902b12dbea631904a8175499a56c8bb95b708483", "openSSLVersion": "OpenSSL 3.5.1 1 Jul 2025", "modules": [], "perconaFeatures": [ "MemoryEngine", "HotBackup", "BackupCursorAggregationStage", "BackupCursorExtendAggregationStage", "AWSIAM", "Kerberos", "LDAP", "OIDC", "TDE", "FIPSMode", "FCBIS", "Auditing", "ProfilingRateLimit", "LogRedaction", "ngram" ], "allocator": "tcmalloc-google", "environment": { "distarch": "x86_64", "target_arch": "x86_64" } }' +++ kubectl_bin -n default delete pod/16983 --grace-period=0 --force ++++ mktemp +++ local LAST_OUT=/tmp/tmp.15txusBak9 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.tsB96n4Sgy +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl -n default delete pod/16983 --grace-period=0 --force +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.15txusBak9 +++ cat /tmp/tmp.tsB96n4Sgy Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. +++ rm /tmp/tmp.15txusBak9 /tmp/tmp.tsB96n4Sgy +++ return 0 +++ echo db version v8.0.19-7 Build Info: '{' '"version":' '"8.0.19-7",' '"gitVersion":' '"902b12dbea631904a8175499a56c8bb95b708483",' '"openSSLVersion":' '"OpenSSL' 3.5.1 1 Jul '2025",' '"modules":' '[],' '"perconaFeatures":' '[' '"MemoryEngine",' '"HotBackup",' '"BackupCursorAggregationStage",' '"BackupCursorExtendAggregationStage",' '"AWSIAM",' '"Kerberos",' '"LDAP",' '"OIDC",' '"TDE",' '"FIPSMode",' '"FCBIS",' '"Auditing",' '"ProfilingRateLimit",' '"LogRedaction",' '"ngram"' '],' '"allocator":' '"tcmalloc-google",' '"environment":' '{' '"distarch":' '"x86_64",' '"target_arch":' '"x86_64"' '}' '}' ++ version_info=8.0.19-7 ++ [[ ! 8.0.19-7 =~ ^([0-9]+\.){2}[0-9]+-[0-9]+$ ]] ++ echo 8.0.19-7 + current_mongod_version=8.0.19-7 ++ echo '{' '"versions":' '[' '{' '"operator":' '"1.23.0",' '"product":' '"psmdb-operator",' '"matrix":' '{' '"mongod":' '{' '"6.0.27-21":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0",' '"status":' '"recommended"' '},' '"7.0.30-16":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod7.0",' '"status":' '"recommended"' '}' '},' '"pmm":' '{' '"2.27.0":' '{' '"image_path":' '"percona/pmm-client:2.27.0",' '"image_hash":' '"4f4d4508afe4ef3b5d72e2ebec6485be84204902b0b2b23c3a9e7c2fe4726884",' '"status":' '"recommended",' '"critical":' false '}' '},' '"backup":' '{},' '"operator":' '{}' '}' '}' ']' '}' ++ jq '.versions[0].matrix.mongod += {"8.0.19-7": {"image_path":"docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0","status":"recommended"}}' + version_service_source='{ "versions": [ { "operator": "1.23.0", "product": "psmdb-operator", "matrix": { "mongod": { "6.0.27-21": { "image_path": "docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0", "status": "recommended" }, "7.0.30-16": { "image_path": "docker.io/perconalab/percona-server-mongodb-operator:main-mongod7.0", "status": "recommended" }, "8.0.19-7": { "image_path": "docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0", "status": "recommended" } }, "pmm": { "2.27.0": { "image_path": "percona/pmm-client:2.27.0", "image_hash": "4f4d4508afe4ef3b5d72e2ebec6485be84204902b0b2b23c3a9e7c2fe4726884", "status": "recommended", "critical": false } }, "backup": {}, "operator": {} } } ] }' ++ echo '{' '"versions":' '[' '{' '"operator":' '"1.23.0",' '"product":' '"psmdb-operator",' '"matrix":' '{' '"mongod":' '{' '"6.0.27-21":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0",' '"status":' '"recommended"' '},' '"7.0.30-16":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod7.0",' '"status":' '"recommended"' '},' '"8.0.19-7":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0",' '"status":' '"recommended"' '}' '},' '"pmm":' '{' '"2.27.0":' '{' '"image_path":' '"percona/pmm-client:2.27.0",' '"image_hash":' '"4f4d4508afe4ef3b5d72e2ebec6485be84204902b0b2b23c3a9e7c2fe4726884",' '"status":' '"recommended",' '"critical":' false '}' '},' '"backup":' '{},' '"operator":' '{}' '}' '}' ']' '}' +++ get_pbm_version docker.io/perconalab/percona-server-mongodb-operator:main-backup +++ local image=docker.io/perconalab/percona-server-mongodb-operator:main-backup ++++ run_simple_cli_inside_image docker.io/perconalab/percona-server-mongodb-operator:main-backup 'pbm-agent version' ++++ local image=docker.io/perconalab/percona-server-mongodb-operator:main-backup ++++ local 'cli=pbm-agent version' ++++ local pod_name=17069 ++++ /usr/sbin/sed -r 's/^Version:\ (([0-9]+\.){2}[0-9]+)\ .*/\1/g' ++++ kubectl_bin -n default run 17069 --image=docker.io/perconalab/percona-server-mongodb-operator:main-backup --restart=Never --command -- sleep infinity +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.hoqRxVZDWH +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.0uLbtYcJAk ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl -n default run 17069 --image=docker.io/perconalab/percona-server-mongodb-operator:main-backup --restart=Never --command -- sleep infinity ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.hoqRxVZDWH ++++ cat /tmp/tmp.0uLbtYcJAk ++++ rm /tmp/tmp.hoqRxVZDWH /tmp/tmp.0uLbtYcJAk ++++ return 0 ++++ kubectl_bin -n default wait --for=condition=Ready pod/17069 +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.q5GpkfKCBw +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.pyJBrN5HdJ ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl -n default wait --for=condition=Ready pod/17069 ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.q5GpkfKCBw ++++ cat /tmp/tmp.pyJBrN5HdJ ++++ rm /tmp/tmp.q5GpkfKCBw /tmp/tmp.pyJBrN5HdJ ++++ return 0 +++++ kubectl_bin -n default exec 17069 -- bash -c 'pbm-agent version 2>&1' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.IiwwK9GNNP ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.T3ngjfE0fj +++++ local exit_status=0 +++++ local timeout=4 ++++++ seq 0 2 +++++ for i in $(seq 0 2) +++++ set +e +++++ kubectl -n default exec 17069 -- bash -c 'pbm-agent version 2>&1' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 -a -n 1 ']' +++++ break +++++ cat /tmp/tmp.IiwwK9GNNP +++++ cat /tmp/tmp.T3ngjfE0fj +++++ rm /tmp/tmp.IiwwK9GNNP /tmp/tmp.T3ngjfE0fj +++++ return 0 ++++ local 'output=Version: 2.13.0 Platform: linux/amd64 GitCommit: e9b57e1395fe2602bd9637d5e4ef19b1df2bb3b7 GitBranch: release-2.13.0 BuildTime: 2026-02-26_12:41_UTC GoVersion: go1.25.7' ++++ kubectl_bin -n default delete pod/17069 --grace-period=0 --force +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.PhdCX11yo5 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.zVqjaIOszv ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl -n default delete pod/17069 --grace-period=0 --force ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.PhdCX11yo5 ++++ cat /tmp/tmp.zVqjaIOszv Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. ++++ rm /tmp/tmp.PhdCX11yo5 /tmp/tmp.zVqjaIOszv ++++ return 0 ++++ echo Version: 2.13.0 Platform: linux/amd64 GitCommit: e9b57e1395fe2602bd9637d5e4ef19b1df2bb3b7 GitBranch: release-2.13.0 BuildTime: 2026-02-26_12:41_UTC GoVersion: go1.25.7 +++ local version_info=2.13.0 +++ [[ ! 2.13.0 =~ ^([0-9]+\.){2}[0-9]+$ ]] +++ echo 2.13.0 ++ jq '.versions[0].matrix.backup += {"2.13.0": {"image_path":"docker.io/perconalab/percona-server-mongodb-operator:main-backup","status":"recommended"}}' + version_service_source='{ "versions": [ { "operator": "1.23.0", "product": "psmdb-operator", "matrix": { "mongod": { "6.0.27-21": { "image_path": "docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0", "status": "recommended" }, "7.0.30-16": { "image_path": "docker.io/perconalab/percona-server-mongodb-operator:main-mongod7.0", "status": "recommended" }, "8.0.19-7": { "image_path": "docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0", "status": "recommended" } }, "pmm": { "2.27.0": { "image_path": "percona/pmm-client:2.27.0", "image_hash": "4f4d4508afe4ef3b5d72e2ebec6485be84204902b0b2b23c3a9e7c2fe4726884", "status": "recommended", "critical": false } }, "backup": { "2.13.0": { "image_path": "docker.io/perconalab/percona-server-mongodb-operator:main-backup", "status": "recommended" } }, "operator": {} } } ] }' ++ echo '{' '"versions":' '[' '{' '"operator":' '"1.23.0",' '"product":' '"psmdb-operator",' '"matrix":' '{' '"mongod":' '{' '"6.0.27-21":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0",' '"status":' '"recommended"' '},' '"7.0.30-16":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod7.0",' '"status":' '"recommended"' '},' '"8.0.19-7":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0",' '"status":' '"recommended"' '}' '},' '"pmm":' '{' '"2.27.0":' '{' '"image_path":' '"percona/pmm-client:2.27.0",' '"image_hash":' '"4f4d4508afe4ef3b5d72e2ebec6485be84204902b0b2b23c3a9e7c2fe4726884",' '"status":' '"recommended",' '"critical":' false '}' '},' '"backup":' '{' '"2.13.0":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-backup",' '"status":' '"recommended"' '}' '},' '"operator":' '{}' '}' '}' ']' '}' ++ jq '.versions[0].matrix.operator += {"1.23.0": {"image_path":"docker.io/perconalab/percona-server-mongodb-operator:PR-2266-f9fc55604","status":"recommended"}}' + version_service_source='{ "versions": [ { "operator": "1.23.0", "product": "psmdb-operator", "matrix": { "mongod": { "6.0.27-21": { "image_path": "docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0", "status": "recommended" }, "7.0.30-16": { "image_path": "docker.io/perconalab/percona-server-mongodb-operator:main-mongod7.0", "status": "recommended" }, "8.0.19-7": { "image_path": "docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0", "status": "recommended" } }, "pmm": { "2.27.0": { "image_path": "percona/pmm-client:2.27.0", "image_hash": "4f4d4508afe4ef3b5d72e2ebec6485be84204902b0b2b23c3a9e7c2fe4726884", "status": "recommended", "critical": false } }, "backup": { "2.13.0": { "image_path": "docker.io/perconalab/percona-server-mongodb-operator:main-backup", "status": "recommended" } }, "operator": { "1.23.0": { "image_path": "docker.io/perconalab/percona-server-mongodb-operator:PR-2266-f9fc55604", "status": "recommended" } } } } ] }' + echo '{' '"versions":' '[' '{' '"operator":' '"1.23.0",' '"product":' '"psmdb-operator",' '"matrix":' '{' '"mongod":' '{' '"6.0.27-21":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0",' '"status":' '"recommended"' '},' '"7.0.30-16":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod7.0",' '"status":' '"recommended"' '},' '"8.0.19-7":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0",' '"status":' '"recommended"' '}' '},' '"pmm":' '{' '"2.27.0":' '{' '"image_path":' '"percona/pmm-client:2.27.0",' '"image_hash":' '"4f4d4508afe4ef3b5d72e2ebec6485be84204902b0b2b23c3a9e7c2fe4726884",' '"status":' '"recommended",' '"critical":' false '}' '},' '"backup":' '{' '"2.13.0":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-backup",' '"status":' '"recommended"' '}' '},' '"operator":' '{' '"1.23.0":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:PR-2266-f9fc55604",' '"status":' '"recommended"' '}' '}' '}' '}' ']' '}' + jq . + kubectl_bin create configmap -n psmdb-operator versions --from-file /tmp/tmp.out8E8NpB2/operator.1.23.0.psmdb-operator.dep.json --from-file /tmp/tmp.out8E8NpB2/operator.1.23.0.psmdb-operator.json ++ mktemp + local LAST_OUT=/tmp/tmp.aenjBih0kC ++ mktemp + local LAST_ERR=/tmp/tmp.3KSYMRwnPQ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create configmap -n psmdb-operator versions --from-file /tmp/tmp.out8E8NpB2/operator.1.23.0.psmdb-operator.dep.json --from-file /tmp/tmp.out8E8NpB2/operator.1.23.0.psmdb-operator.json + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.aenjBih0kC configmap/versions created + cat /tmp/tmp.3KSYMRwnPQ + rm /tmp/tmp.aenjBih0kC /tmp/tmp.3KSYMRwnPQ + return 0 + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/mongod-major-upgrade-sharded/conf/vs.yml + /usr/sbin/sed -r s#operator.9.9.9.psmdb-operator#operator.1.23.0.psmdb-operator#g + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.nt8VCZO060 ++ mktemp + local LAST_ERR=/tmp/tmp.FPA7CvxzOi + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.nt8VCZO060 deployment.apps/version-service created service/version-service created + cat /tmp/tmp.FPA7CvxzOi + rm /tmp/tmp.nt8VCZO060 /tmp/tmp.FPA7CvxzOi + return 0 ++ jq '.[] | .[] |.matrix.mongod' /tmp/tmp.out8E8NpB2/operator.1.23.0.psmdb-operator.json + version_matrix='{ "6.0.27-21": { "image_path": "docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0", "status": "recommended" }, "7.0.30-16": { "image_path": "docker.io/perconalab/percona-server-mongodb-operator:main-mongod7.0", "status": "recommended" }, "8.0.19-7": { "image_path": "docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0", "status": "recommended" } }' ++ get_mongod_ver_from_image docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0 ++ local image=docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0 +++ /usr/sbin/sed -r 's/^.*db version v(([0-9]+\.){2}[0-9]+-[0-9]+).*$/\1/g' +++ run_simple_cli_inside_image docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0 'mongod --version' +++ local image=docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0 +++ local 'cli=mongod --version' +++ local pod_name=23409 +++ kubectl_bin -n default run 23409 --image=docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0 --restart=Never --command -- sleep infinity ++++ mktemp +++ local LAST_OUT=/tmp/tmp.aYNPUm2hzE ++++ mktemp +++ local LAST_ERR=/tmp/tmp.iS5KcoQdyp +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl -n default run 23409 --image=docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0 --restart=Never --command -- sleep infinity +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.aYNPUm2hzE +++ cat /tmp/tmp.iS5KcoQdyp +++ rm /tmp/tmp.aYNPUm2hzE /tmp/tmp.iS5KcoQdyp +++ return 0 +++ kubectl_bin -n default wait --for=condition=Ready pod/23409 ++++ mktemp +++ local LAST_OUT=/tmp/tmp.QMj0PojiaF ++++ mktemp +++ local LAST_ERR=/tmp/tmp.7DnIu1nUa0 +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl -n default wait --for=condition=Ready pod/23409 +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.QMj0PojiaF +++ cat /tmp/tmp.7DnIu1nUa0 +++ rm /tmp/tmp.QMj0PojiaF /tmp/tmp.7DnIu1nUa0 +++ return 0 ++++ kubectl_bin -n default exec 23409 -- bash -c 'mongod --version 2>&1' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.zcdCG0IgA6 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.7mw79sg2PL ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl -n default exec 23409 -- bash -c 'mongod --version 2>&1' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.zcdCG0IgA6 ++++ cat /tmp/tmp.7mw79sg2PL ++++ rm /tmp/tmp.zcdCG0IgA6 /tmp/tmp.7mw79sg2PL ++++ return 0 +++ local 'output=db version v8.0.19-7 Build Info: { "version": "8.0.19-7", "gitVersion": "902b12dbea631904a8175499a56c8bb95b708483", "openSSLVersion": "OpenSSL 3.5.1 1 Jul 2025", "modules": [], "perconaFeatures": [ "MemoryEngine", "HotBackup", "BackupCursorAggregationStage", "BackupCursorExtendAggregationStage", "AWSIAM", "Kerberos", "LDAP", "OIDC", "TDE", "FIPSMode", "FCBIS", "Auditing", "ProfilingRateLimit", "LogRedaction", "ngram" ], "allocator": "tcmalloc-google", "environment": { "distarch": "x86_64", "target_arch": "x86_64" } }' +++ kubectl_bin -n default delete pod/23409 --grace-period=0 --force ++++ mktemp +++ local LAST_OUT=/tmp/tmp.1CQuOfYGBY ++++ mktemp +++ local LAST_ERR=/tmp/tmp.A2qkxo2V7l +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl -n default delete pod/23409 --grace-period=0 --force +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.1CQuOfYGBY +++ cat /tmp/tmp.A2qkxo2V7l Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. +++ rm /tmp/tmp.1CQuOfYGBY /tmp/tmp.A2qkxo2V7l +++ return 0 +++ echo db version v8.0.19-7 Build Info: '{' '"version":' '"8.0.19-7",' '"gitVersion":' '"902b12dbea631904a8175499a56c8bb95b708483",' '"openSSLVersion":' '"OpenSSL' 3.5.1 1 Jul '2025",' '"modules":' '[],' '"perconaFeatures":' '[' '"MemoryEngine",' '"HotBackup",' '"BackupCursorAggregationStage",' '"BackupCursorExtendAggregationStage",' '"AWSIAM",' '"Kerberos",' '"LDAP",' '"OIDC",' '"TDE",' '"FIPSMode",' '"FCBIS",' '"Auditing",' '"ProfilingRateLimit",' '"LogRedaction",' '"ngram"' '],' '"allocator":' '"tcmalloc-google",' '"environment":' '{' '"distarch":' '"x86_64",' '"target_arch":' '"x86_64"' '}' '}' ++ version_info=8.0.19-7 ++ [[ ! 8.0.19-7 =~ ^([0-9]+\.){2}[0-9]+-[0-9]+$ ]] ++ echo 8.0.19-7 + current_mongod_version=8.0.19-7 ++ echo '{' '"6.0.27-21":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0",' '"status":' '"recommended"' '},' '"7.0.30-16":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod7.0",' '"status":' '"recommended"' '},' '"8.0.19-7":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0",' '"status":' '"recommended"' '}' '}' ++ jq '. += {"8.0.19-7":{"image_path":"docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0","status": "recommended"}}' + version_matrix='{ "6.0.27-21": { "image_path": "docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0", "status": "recommended" }, "7.0.30-16": { "image_path": "docker.io/perconalab/percona-server-mongodb-operator:main-mongod7.0", "status": "recommended" }, "8.0.19-7": { "image_path": "docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0", "status": "recommended" } }' ++ echo '{' '"6.0.27-21":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0",' '"status":' '"recommended"' '},' '"7.0.30-16":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod7.0",' '"status":' '"recommended"' '},' '"8.0.19-7":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0",' '"status":' '"recommended"' '}' '}' ++ jq 'to_entries | sort_by( .key | split("[[:punct:]]";"g") | map(tonumber) ) | map({(.key): .value}) ' + version_matrix='[ { "6.0.27-21": { "image_path": "docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0", "status": "recommended" } }, { "7.0.30-16": { "image_path": "docker.io/perconalab/percona-server-mongodb-operator:main-mongod7.0", "status": "recommended" } }, { "8.0.19-7": { "image_path": "docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0", "status": "recommended" } } ]' ++ echo '[' '{' '"6.0.27-21":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0",' '"status":' '"recommended"' '}' '},' '{' '"7.0.30-16":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod7.0",' '"status":' '"recommended"' '}' '},' '{' '"8.0.19-7":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0",' '"status":' '"recommended"' '}' '}' ']' ++ jq -r '.[] | keys | .[] | split(".") | .[:2] | join(".")' ++ uniq ++ tail -n +2 + versions_to_verify='7.0 8.0' + cluster=some-name ++ echo '[' '{' '"6.0.27-21":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0",' '"status":' '"recommended"' '}' '},' '{' '"7.0.30-16":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod7.0",' '"status":' '"recommended"' '}' '},' '{' '"8.0.19-7":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0",' '"status":' '"recommended"' '}' '}' ']' ++ jq '.[0] | keys | .[0]' + desc 'Starting the cluster with IMAGE_MONGOD "6.0.27-21"' + set +o xtrace ----------------------------------------------------------------------------------- Starting the cluster with IMAGE_MONGOD "6.0.27-21" ----------------------------------------------------------------------------------- ++ echo '[' '{' '"6.0.27-21":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0",' '"status":' '"recommended"' '}' '},' '{' '"7.0.30-16":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod7.0",' '"status":' '"recommended"' '}' '},' '{' '"8.0.19-7":' '{' '"image_path":' '"docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0",' '"status":' '"recommended"' '}' '}' ']' ++ jq -r 'to_entries | .[0].value | .[].image_path' + export IMAGE_MONGOD=docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0 + IMAGE_MONGOD=docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0 + qualify_image_var IMAGE_MONGOD + local var=IMAGE_MONGOD + local val=docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0 + local out= + local ref + [[ -z docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0 ]] + [[ docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0 == *\ * ]] ++ qualify_image_ref docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0 ++ local ref=docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0 ++ local first=docker.io ++ [[ docker.io == *.* ]] ++ printf %s docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0 ++ return 0 + printf -v IMAGE_MONGOD %s docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0 + apply_cluster /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/mongod-major-upgrade-sharded/conf/some-name.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/mongod-major-upgrade-sharded/conf/some-name.yml + kubectl_bin apply -f - ++ mktemp + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/mongod-major-upgrade-sharded/conf/some-name.yml + yq eval '(.spec | select(.image == null)).image = "docker.io/perconalab/percona-server-mongodb-operator:main-mongod6.0"' + yq eval '(.spec | select(has("pmm"))).pmm.image = "docker.io/percona/pmm-client:2.44.1-1"' + yq eval '(.spec | select(has("initImage"))).initImage = "docker.io/perconalab/percona-server-mongodb-operator:PR-2266-f9fc55604"' + yq eval '(.spec | select(has("backup"))).backup.image = "docker.io/perconalab/percona-server-mongodb-operator:main-backup"' + /usr/sbin/sed -e s/NAME_SPACE/mongod-major-upgrade-sharded-7355/g + local LAST_OUT=/tmp/tmp.eCDs5dYvhe + yq eval '.spec.upgradeOptions.apply="Never"' ++ mktemp + local LAST_ERR=/tmp/tmp.WiGShEMu2g + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.eCDs5dYvhe perconaservermongodb.psmdb.percona.com/some-name created + cat /tmp/tmp.WiGShEMu2g + rm /tmp/tmp.eCDs5dYvhe /tmp/tmp.WiGShEMu2g + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- + wait_for_running some-name-cfg 3 + local name=some-name-cfg + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=cfg + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-cfg-0 + local pod=some-name-cfg-0 + set +o xtrace waiting for pod/some-name-cfg-0 to be ready..............OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-cfg-1 + local pod=some-name-cfg-1 + set +o xtrace waiting for pod/some-name-cfg-1 to be ready............OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZqOhMURUCc +++ mktemp ++ local LAST_ERR=/tmp/tmp.dnsOHL81IT ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ZqOhMURUCc ++ cat /tmp/tmp.dnsOHL81IT ++ rm /tmp/tmp.ZqOhMURUCc /tmp/tmp.dnsOHL81IT ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-cfg-2 + local pod=some-name-cfg-2 + set +o xtrace waiting for pod/some-name-cfg-2 to be ready...........OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.veWP3UJV4n +++ mktemp ++ local LAST_ERR=/tmp/tmp.EZWM1xlaQF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.veWP3UJV4n ++ cat /tmp/tmp.EZWM1xlaQF ++ rm /tmp/tmp.veWP3UJV4n /tmp/tmp.EZWM1xlaQF ++ return 0 + [[ '' == \t\r\u\e ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.58KmACXBIR +++ mktemp ++ local LAST_ERR=/tmp/tmp.wFn0FERd6R ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.58KmACXBIR ++ cat /tmp/tmp.wFn0FERd6R ++ rm /tmp/tmp.58KmACXBIR /tmp/tmp.wFn0FERd6R ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness......................... + wait_for_running some-name-rs0 3 + local name=some-name-rs0 + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=rs0 + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-rs0-0 + local pod=some-name-rs0-0 + set +o xtrace waiting for pod/some-name-rs0-0 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-rs0-1 + local pod=some-name-rs0-1 + set +o xtrace waiting for pod/some-name-rs0-1 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GfKdOARQV9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.APu3KHkfS4 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.GfKdOARQV9 ++ cat /tmp/tmp.APu3KHkfS4 ++ rm /tmp/tmp.GfKdOARQV9 /tmp/tmp.APu3KHkfS4 ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-rs0-2 + local pod=some-name-rs0-2 + set +o xtrace waiting for pod/some-name-rs0-2 to be ready.OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.siu51qGvIE +++ mktemp ++ local LAST_ERR=/tmp/tmp.kV893pjgRT ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.siu51qGvIE ++ cat /tmp/tmp.kV893pjgRT ++ rm /tmp/tmp.siu51qGvIE /tmp/tmp.kV893pjgRT ++ return 0 + [[ '' == \t\r\u\e ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xAfvsZPZT9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.3obZnHHl15 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xAfvsZPZT9 ++ cat /tmp/tmp.3obZnHHl15 ++ rm /tmp/tmp.xAfvsZPZT9 /tmp/tmp.3obZnHHl15 ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness + wait_for_running some-name-mongos 3 + local name=some-name-mongos + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=mongos + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-mongos-0 + local pod=some-name-mongos-0 + set +o xtrace waiting for pod/some-name-mongos-0 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-mongos-1 + local pod=some-name-mongos-1 + set +o xtrace waiting for pod/some-name-mongos-1 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AjTCi3kkoL +++ mktemp ++ local LAST_ERR=/tmp/tmp.Th5XTL4YsW ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.AjTCi3kkoL ++ cat /tmp/tmp.Th5XTL4YsW ++ rm /tmp/tmp.AjTCi3kkoL /tmp/tmp.Th5XTL4YsW ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-mongos-2 + local pod=some-name-mongos-2 + set +o xtrace waiting for pod/some-name-mongos-2 to be ready.OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pMqTvx1hn4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.LC2FnCgpsI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.pMqTvx1hn4 ++ cat /tmp/tmp.LC2FnCgpsI ++ rm /tmp/tmp.pMqTvx1hn4 /tmp/tmp.LC2FnCgpsI ++ return 0 + [[ '' == \t\r\u\e ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oCZGWhf3sJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.0WNk2V5I3F ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.oCZGWhf3sJ ++ cat /tmp/tmp.0WNk2V5I3F ++ rm /tmp/tmp.oCZGWhf3sJ /tmp/tmp.0WNk2V5I3F ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.clWunu4X56 +++ mktemp ++ local LAST_ERR=/tmp/tmp.AXutHCyffD ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.clWunu4X56 ++ cat /tmp/tmp.AXutHCyffD ++ rm /tmp/tmp.clWunu4X56 /tmp/tmp.AXutHCyffD ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo .OK .OK + desc 'write data, read from all' + set +o xtrace ----------------------------------------------------------------------------------- write data, read from all ----------------------------------------------------------------------------------- + run_mongos 'db.createUser({user:"myApp",pwd:"myPass",roles:[{db:"myApp",role:"readWrite"}]})' userAdmin:userAdmin123456@some-name-mongos.mongod-major-upgrade-sharded-7355 + local 'command=db.createUser({user:"myApp",pwd:"myPass",roles:[{db:"myApp",role:"readWrite"}]})' + local uri=userAdmin:userAdmin123456@some-name-mongos.mongod-major-upgrade-sharded-7355 + local driver=mongodb + local suffix=.svc.cluster.local + local mongo_flag= + local port=27017 + local mongo_bin=mongo ++ echo .svc.cluster.local ++ awk -F: '{print $2}' + suffix_port= + [[ -z '' ]] + suffix=.svc.cluster.local:27017 ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ztQcISTMuv +++ mktemp ++ local LAST_ERR=/tmp/tmp.OYga197tXz ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ztQcISTMuv ++ cat /tmp/tmp.OYga197tXz ++ rm /tmp/tmp.ztQcISTMuv /tmp/tmp.OYga197tXz ++ return 0 + local client_container=psmdb-client-bb8b97679-zfsvh + kubectl_bin exec psmdb-client-bb8b97679-zfsvh -- bash -c 'printf '\''db.createUser({user:"myApp",pwd:"myPass",roles:[{db:"myApp",role:"readWrite"}]})\n'\'' | mongo mongodb://userAdmin:userAdmin123456@some-name-mongos.mongod-major-upgrade-sharded-7355.svc.cluster.local:27017/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.E3nN5It0Ol ++ mktemp + local LAST_ERR=/tmp/tmp.V1vwsQzyqr + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-zfsvh -- bash -c 'printf '\''db.createUser({user:"myApp",pwd:"myPass",roles:[{db:"myApp",role:"readWrite"}]})\n'\'' | mongo mongodb://userAdmin:userAdmin123456@some-name-mongos.mongod-major-upgrade-sharded-7355.svc.cluster.local:27017/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.E3nN5It0Ol Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.mongod-major-upgrade-sharded-7355.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("d422a2ef-4f76-4ce9-8675-6e03c27228ca") } Percona Server for MongoDB server version: v6.0.27-21 WARNING: shell and server versions do not match Successfully added user: { "user" : "myApp", "roles" : [ { "db" : "myApp", "role" : "readWrite" } ] } bye + cat /tmp/tmp.V1vwsQzyqr + rm /tmp/tmp.E3nN5It0Ol /tmp/tmp.V1vwsQzyqr + return 0 + sleep 2 + run_mongos 'use myApp\n db.test.insert({ x: 100501 })' myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355 + local 'command=use myApp\n db.test.insert({ x: 100501 })' + local uri=myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355 + local driver=mongodb + local suffix=.svc.cluster.local + local mongo_flag= + local port=27017 + local mongo_bin=mongo ++ echo .svc.cluster.local ++ awk -F: '{print $2}' + suffix_port= + [[ -z '' ]] + suffix=.svc.cluster.local:27017 ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QDYXpmjlBz +++ mktemp ++ local LAST_ERR=/tmp/tmp.SmShIaUTe3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.QDYXpmjlBz ++ cat /tmp/tmp.SmShIaUTe3 ++ rm /tmp/tmp.QDYXpmjlBz /tmp/tmp.SmShIaUTe3 ++ return 0 + local client_container=psmdb-client-bb8b97679-zfsvh + kubectl_bin exec psmdb-client-bb8b97679-zfsvh -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355.svc.cluster.local:27017/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.smwcg0aI1A ++ mktemp + local LAST_ERR=/tmp/tmp.mRzGzG4DUu + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-zfsvh -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355.svc.cluster.local:27017/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.smwcg0aI1A Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.mongod-major-upgrade-sharded-7355.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("2911e36e-8123-4d86-93d5-a1b51cc091a0") } Percona Server for MongoDB server version: v6.0.27-21 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.mRzGzG4DUu + rm /tmp/tmp.smwcg0aI1A /tmp/tmp.mRzGzG4DUu + return 0 + compare_mongos_cmd find myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355 + local command=find + local uri=myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355 + local postfix= + local suffix= + local database=myApp + local collection=test + local port=27017 + local tls=false + [[ false == \t\r\u\e ]] + mongos_command=run_mongos + log 'running db.test.command() in myApp' + set +o xtrace [2026-03-10T07:33:03+0000] running db.test.command() in myApp + run_mongos 'use myApp\n db.test.find()' myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355 mongodb '' '' 27017 + local 'command=use myApp\n db.test.find()' + local uri=myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355 + local driver=mongodb + local suffix=.svc.cluster.local + local mongo_flag= + local port=27017 + local mongo_bin=mongo + grep -E -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:|Started a new thread for the timer service' + /usr/sbin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ echo .svc.cluster.local ++ awk -F: '{print $2}' + suffix_port= + [[ -z '' ]] + suffix=.svc.cluster.local:27017 ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dKlBKRQGyT +++ mktemp ++ local LAST_ERR=/tmp/tmp.z0Zx8VryC0 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.dKlBKRQGyT ++ cat /tmp/tmp.z0Zx8VryC0 ++ rm /tmp/tmp.dKlBKRQGyT /tmp/tmp.z0Zx8VryC0 ++ return 0 + local client_container=psmdb-client-bb8b97679-zfsvh + kubectl_bin exec psmdb-client-bb8b97679-zfsvh -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355.svc.cluster.local:27017/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.ve4I0xH6qJ ++ mktemp + local LAST_ERR=/tmp/tmp.c6cQLQbKfI + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-zfsvh -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355.svc.cluster.local:27017/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ve4I0xH6qJ + cat /tmp/tmp.c6cQLQbKfI + rm /tmp/tmp.ve4I0xH6qJ /tmp/tmp.c6cQLQbKfI + return 0 + [[ 0 -eq 0 ]] + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/mongod-major-upgrade-sharded/compare/find.json /tmp/tmp.out8E8NpB2/find + desc 'Starting to follow mongod upgrade images chain' + set +o xtrace ----------------------------------------------------------------------------------- Starting to follow mongod upgrade images chain ----------------------------------------------------------------------------------- + target_generation=2 + for version in ${versions_to_verify[@]} + desc 'Testing upgrade to version: 7.0' + set +o xtrace ----------------------------------------------------------------------------------- Testing upgrade to version: 7.0 ----------------------------------------------------------------------------------- + kubectl_bin patch psmdb/some-name --type=json '-p=[ {"op":"replace","path":"/spec/upgradeOptions/apply","value": "7.0-recommended"}, ]' ++ mktemp + local LAST_OUT=/tmp/tmp.u1aXFlLup4 ++ mktemp + local LAST_ERR=/tmp/tmp.RigMvAofix + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl patch psmdb/some-name --type=json '-p=[ {"op":"replace","path":"/spec/upgradeOptions/apply","value": "7.0-recommended"}, ]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.u1aXFlLup4 perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.RigMvAofix + rm /tmp/tmp.u1aXFlLup4 /tmp/tmp.RigMvAofix + return 0 + sleep 70 + wait_for_running some-name-cfg 3 + local name=some-name-cfg + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=cfg + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-cfg-0 + local pod=some-name-cfg-0 + set +o xtrace waiting for pod/some-name-cfg-0 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-cfg-1 + local pod=some-name-cfg-1 + set +o xtrace waiting for pod/some-name-cfg-1 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.v0qE9728rZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.qzpKUuVcxw ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.v0qE9728rZ ++ cat /tmp/tmp.qzpKUuVcxw ++ rm /tmp/tmp.v0qE9728rZ /tmp/tmp.qzpKUuVcxw ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-cfg-2 + local pod=some-name-cfg-2 + set +o xtrace waiting for pod/some-name-cfg-2 to be ready.OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PtXoOJwob9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.LphA62M2UD ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.PtXoOJwob9 ++ cat /tmp/tmp.LphA62M2UD ++ rm /tmp/tmp.PtXoOJwob9 /tmp/tmp.LphA62M2UD ++ return 0 + [[ '' == \t\r\u\e ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ixTOhf2gIc +++ mktemp ++ local LAST_ERR=/tmp/tmp.71UKvZJf8X ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ixTOhf2gIc ++ cat /tmp/tmp.71UKvZJf8X ++ rm /tmp/tmp.ixTOhf2gIc /tmp/tmp.71UKvZJf8X ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness.......................................................................................................................................... + wait_for_running some-name-rs0 3 + local name=some-name-rs0 + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=rs0 + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-rs0-0 + local pod=some-name-rs0-0 + set +o xtrace waiting for pod/some-name-rs0-0 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-rs0-1 + local pod=some-name-rs0-1 + set +o xtrace waiting for pod/some-name-rs0-1 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xDPZHPBtjY +++ mktemp ++ local LAST_ERR=/tmp/tmp.CHuxtlUucO ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xDPZHPBtjY ++ cat /tmp/tmp.CHuxtlUucO ++ rm /tmp/tmp.xDPZHPBtjY /tmp/tmp.CHuxtlUucO ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-rs0-2 + local pod=some-name-rs0-2 + set +o xtrace waiting for pod/some-name-rs0-2 to be ready.OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kPY7dDi6PI +++ mktemp ++ local LAST_ERR=/tmp/tmp.kzJY24jOzB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.kPY7dDi6PI ++ cat /tmp/tmp.kzJY24jOzB ++ rm /tmp/tmp.kPY7dDi6PI /tmp/tmp.kzJY24jOzB ++ return 0 + [[ '' == \t\r\u\e ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4wcSVZ3UvA +++ mktemp ++ local LAST_ERR=/tmp/tmp.jaUUE9VwGF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4wcSVZ3UvA ++ cat /tmp/tmp.jaUUE9VwGF ++ rm /tmp/tmp.4wcSVZ3UvA /tmp/tmp.jaUUE9VwGF ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness + wait_for_running some-name-mongos 3 + local name=some-name-mongos + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=mongos + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-mongos-0 + local pod=some-name-mongos-0 + set +o xtrace waiting for pod/some-name-mongos-0 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-mongos-1 + local pod=some-name-mongos-1 + set +o xtrace waiting for pod/some-name-mongos-1 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LME5CYdOK5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.8ZDGICK1Br ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.LME5CYdOK5 ++ cat /tmp/tmp.8ZDGICK1Br ++ rm /tmp/tmp.LME5CYdOK5 /tmp/tmp.8ZDGICK1Br ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-mongos-2 + local pod=some-name-mongos-2 + set +o xtrace waiting for pod/some-name-mongos-2 to be ready.OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.d4XZ71Ykts +++ mktemp ++ local LAST_ERR=/tmp/tmp.HWn3jPY7wQ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.d4XZ71Ykts ++ cat /tmp/tmp.HWn3jPY7wQ ++ rm /tmp/tmp.d4XZ71Ykts /tmp/tmp.HWn3jPY7wQ ++ return 0 + [[ '' == \t\r\u\e ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HUA7pNB1XP +++ mktemp ++ local LAST_ERR=/tmp/tmp.xu1BbA8IXw ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.HUA7pNB1XP ++ cat /tmp/tmp.xu1BbA8IXw ++ rm /tmp/tmp.HUA7pNB1XP /tmp/tmp.xu1BbA8IXw ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.enN4ZPtdrZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.Km3JBUqvO0 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.enN4ZPtdrZ ++ cat /tmp/tmp.Km3JBUqvO0 ++ rm /tmp/tmp.enN4ZPtdrZ /tmp/tmp.Km3JBUqvO0 ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo .OK .OK + compare_generation 2 statefulset some-name-cfg + local generation=2 + local resource_type=statefulset + local resource_name=some-name-cfg + local current_generation ++ kubectl_bin get statefulset some-name-cfg -o 'jsonpath={.metadata.generation}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YudHHiDswz +++ mktemp ++ local LAST_ERR=/tmp/tmp.fu7Bvi1wt1 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get statefulset some-name-cfg -o 'jsonpath={.metadata.generation}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.YudHHiDswz ++ cat /tmp/tmp.fu7Bvi1wt1 ++ rm /tmp/tmp.YudHHiDswz /tmp/tmp.fu7Bvi1wt1 ++ return 0 + current_generation=2 + [[ 2 != \2 ]] + compare_generation 2 statefulset some-name-rs0 + local generation=2 + local resource_type=statefulset + local resource_name=some-name-rs0 + local current_generation ++ kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={.metadata.generation}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SrUIV4ZH7A +++ mktemp ++ local LAST_ERR=/tmp/tmp.2BmYhh5jYz ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get statefulset some-name-rs0 -o 'jsonpath={.metadata.generation}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.SrUIV4ZH7A ++ cat /tmp/tmp.2BmYhh5jYz ++ rm /tmp/tmp.SrUIV4ZH7A /tmp/tmp.2BmYhh5jYz ++ return 0 + current_generation=2 + [[ 2 != \2 ]] + compare_generation 2 statefulset some-name-mongos + local generation=2 + local resource_type=statefulset + local resource_name=some-name-mongos + local current_generation ++ kubectl_bin get statefulset some-name-mongos -o 'jsonpath={.metadata.generation}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fLrDg4dLLK +++ mktemp ++ local LAST_ERR=/tmp/tmp.bnANe8gvWh ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get statefulset some-name-mongos -o 'jsonpath={.metadata.generation}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.fLrDg4dLLK ++ cat /tmp/tmp.bnANe8gvWh ++ rm /tmp/tmp.fLrDg4dLLK /tmp/tmp.bnANe8gvWh ++ return 0 + current_generation=2 + [[ 2 != \2 ]] + kubectl_bin patch psmdb/some-name --type=json '-p=[ {"op":"replace","path":"/spec/upgradeOptions/setFCV","value": true} ]' ++ mktemp + local LAST_OUT=/tmp/tmp.h5AK6D2jZc ++ mktemp + local LAST_ERR=/tmp/tmp.kGMQIiictE + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl patch psmdb/some-name --type=json '-p=[ {"op":"replace","path":"/spec/upgradeOptions/setFCV","value": true} ]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.h5AK6D2jZc perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.kGMQIiictE + rm /tmp/tmp.h5AK6D2jZc /tmp/tmp.kGMQIiictE + return 0 + sleep 10 + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ux8rIJmVx8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.qHTtoO9cTG ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ux8rIJmVx8 ++ cat /tmp/tmp.qHTtoO9cTG ++ rm /tmp/tmp.ux8rIJmVx8 /tmp/tmp.qHTtoO9cTG ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo .OK .OK ++ run_mongos 'JSON.stringify(db.adminCommand({getParameter:1,featureCompatibilityVersion:1}))' clusterAdmin:clusterAdmin123456@some-name-cfg.mongod-major-upgrade-sharded-7355 ++ local 'command=JSON.stringify(db.adminCommand({getParameter:1,featureCompatibilityVersion:1}))' ++ local uri=clusterAdmin:clusterAdmin123456@some-name-cfg.mongod-major-upgrade-sharded-7355 ++ local driver=mongodb ++ local suffix=.svc.cluster.local ++ local mongo_flag= ++ local port=27017 ++ local mongo_bin=mongo ++ grep -E '^\{.*\}$' ++ jq -r .featureCompatibilityVersion.version +++ echo .svc.cluster.local +++ awk -F: '{print $2}' ++ suffix_port= ++ [[ -z '' ]] ++ suffix=.svc.cluster.local:27017 +++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.mMxHQOAJA1 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.TlSUQZvKbR +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.mMxHQOAJA1 +++ cat /tmp/tmp.TlSUQZvKbR +++ rm /tmp/tmp.mMxHQOAJA1 /tmp/tmp.TlSUQZvKbR +++ return 0 ++ local client_container=psmdb-client-bb8b97679-zfsvh ++ kubectl_bin exec psmdb-client-bb8b97679-zfsvh -- bash -c 'printf '\''JSON.stringify(db.adminCommand({getParameter:1,featureCompatibilityVersion:1}))\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@some-name-cfg.mongod-major-upgrade-sharded-7355.svc.cluster.local:27017/admin ' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8z7W1ZIPut +++ mktemp ++ local LAST_ERR=/tmp/tmp.OkQFPnbNcQ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl exec psmdb-client-bb8b97679-zfsvh -- bash -c 'printf '\''JSON.stringify(db.adminCommand({getParameter:1,featureCompatibilityVersion:1}))\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@some-name-cfg.mongod-major-upgrade-sharded-7355.svc.cluster.local:27017/admin ' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.8z7W1ZIPut ++ cat /tmp/tmp.OkQFPnbNcQ ++ rm /tmp/tmp.8z7W1ZIPut /tmp/tmp.OkQFPnbNcQ ++ return 0 + currentFCV=7.0 + [[ 7.0 != 7.0 ]] + run_mongos 'use myApp\n db.test.insert({ x: 100502 })' myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355 + local 'command=use myApp\n db.test.insert({ x: 100502 })' + local uri=myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355 + local driver=mongodb + local suffix=.svc.cluster.local + local mongo_flag= + local port=27017 + local mongo_bin=mongo ++ echo .svc.cluster.local ++ awk -F: '{print $2}' + suffix_port= + [[ -z '' ]] + suffix=.svc.cluster.local:27017 ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xljxrx8wNi +++ mktemp ++ local LAST_ERR=/tmp/tmp.buKTxvIxeW ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xljxrx8wNi ++ cat /tmp/tmp.buKTxvIxeW ++ rm /tmp/tmp.xljxrx8wNi /tmp/tmp.buKTxvIxeW ++ return 0 + local client_container=psmdb-client-bb8b97679-zfsvh + kubectl_bin exec psmdb-client-bb8b97679-zfsvh -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100502 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355.svc.cluster.local:27017/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.CF3mhGeNRY ++ mktemp + local LAST_ERR=/tmp/tmp.UPAQJejj9h + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-zfsvh -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100502 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355.svc.cluster.local:27017/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.CF3mhGeNRY Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.mongod-major-upgrade-sharded-7355.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("bf7d4793-55ce-45c3-8c95-46b27fd40323") } Percona Server for MongoDB server version: v7.0.30-16 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.UPAQJejj9h + rm /tmp/tmp.CF3mhGeNRY /tmp/tmp.UPAQJejj9h + return 0 + compare_mongos_cmd find myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355 -2 + local command=find + local uri=myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355 + local postfix=-2 + local suffix= + local database=myApp + local collection=test + local port=27017 + local tls=false + [[ false == \t\r\u\e ]] + mongos_command=run_mongos + log 'running db.test.command() in myApp' + set +o xtrace [2026-03-10T07:40:23+0000] running db.test.command() in myApp + run_mongos 'use myApp\n db.test.find()' myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355 mongodb '' '' 27017 + local 'command=use myApp\n db.test.find()' + local uri=myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355 + local driver=mongodb + local suffix=.svc.cluster.local + local mongo_flag= + local port=27017 + local mongo_bin=mongo + grep -E -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:|Started a new thread for the timer service' + /usr/sbin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ echo .svc.cluster.local ++ awk -F: '{print $2}' + suffix_port= + [[ -z '' ]] + suffix=.svc.cluster.local:27017 ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eX5yohsvMu +++ mktemp ++ local LAST_ERR=/tmp/tmp.maP6FPMxET ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.eX5yohsvMu ++ cat /tmp/tmp.maP6FPMxET ++ rm /tmp/tmp.eX5yohsvMu /tmp/tmp.maP6FPMxET ++ return 0 + local client_container=psmdb-client-bb8b97679-zfsvh + kubectl_bin exec psmdb-client-bb8b97679-zfsvh -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355.svc.cluster.local:27017/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.8kNU9dGqQr ++ mktemp + local LAST_ERR=/tmp/tmp.sBYlG3MvFE + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-zfsvh -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355.svc.cluster.local:27017/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.8kNU9dGqQr + cat /tmp/tmp.sBYlG3MvFE + rm /tmp/tmp.8kNU9dGqQr /tmp/tmp.sBYlG3MvFE + return 0 + [[ 0 -eq 0 ]] + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/mongod-major-upgrade-sharded/compare/find-2.json /tmp/tmp.out8E8NpB2/find-2 + target_generation=3 + for version in ${versions_to_verify[@]} + desc 'Testing upgrade to version: 8.0' + set +o xtrace ----------------------------------------------------------------------------------- Testing upgrade to version: 8.0 ----------------------------------------------------------------------------------- + kubectl_bin patch psmdb/some-name --type=json '-p=[ {"op":"replace","path":"/spec/upgradeOptions/apply","value": "8.0-recommended"}, ]' ++ mktemp + local LAST_OUT=/tmp/tmp.RVuXzvGywB ++ mktemp + local LAST_ERR=/tmp/tmp.SVgwy6LIde + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl patch psmdb/some-name --type=json '-p=[ {"op":"replace","path":"/spec/upgradeOptions/apply","value": "8.0-recommended"}, ]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.RVuXzvGywB perconaservermongodb.psmdb.percona.com/some-name patched + cat /tmp/tmp.SVgwy6LIde + rm /tmp/tmp.RVuXzvGywB /tmp/tmp.SVgwy6LIde + return 0 + sleep 70 + wait_for_running some-name-cfg 3 + local name=some-name-cfg + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=cfg + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-cfg-0 + local pod=some-name-cfg-0 + set +o xtrace waiting for pod/some-name-cfg-0 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-cfg-1 + local pod=some-name-cfg-1 + set +o xtrace waiting for pod/some-name-cfg-1 to be ready.....OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rCqAzeVlEb +++ mktemp ++ local LAST_ERR=/tmp/tmp.roRNEtt4cK ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.rCqAzeVlEb ++ cat /tmp/tmp.roRNEtt4cK ++ rm /tmp/tmp.rCqAzeVlEb /tmp/tmp.roRNEtt4cK ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-cfg-2 + local pod=some-name-cfg-2 + set +o xtrace waiting for pod/some-name-cfg-2 to be ready.OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.E3ZOy7flXj +++ mktemp ++ local LAST_ERR=/tmp/tmp.4VyAPJN3se ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.E3ZOy7flXj ++ cat /tmp/tmp.4VyAPJN3se ++ rm /tmp/tmp.E3ZOy7flXj /tmp/tmp.4VyAPJN3se ++ return 0 + [[ '' == \t\r\u\e ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7pubqOHpsS +++ mktemp ++ local LAST_ERR=/tmp/tmp.qVLMiROnts ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.7pubqOHpsS ++ cat /tmp/tmp.qVLMiROnts ++ rm /tmp/tmp.7pubqOHpsS /tmp/tmp.qVLMiROnts ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness............................................................................................................................... + wait_for_running some-name-rs0 3 + local name=some-name-rs0 + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=rs0 + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-rs0-0 + local pod=some-name-rs0-0 + set +o xtrace waiting for pod/some-name-rs0-0 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-rs0-1 + local pod=some-name-rs0-1 + set +o xtrace waiting for pod/some-name-rs0-1 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.m7N8mZKbYD +++ mktemp ++ local LAST_ERR=/tmp/tmp.ikzyBp4GwN ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.m7N8mZKbYD ++ cat /tmp/tmp.ikzyBp4GwN ++ rm /tmp/tmp.m7N8mZKbYD /tmp/tmp.ikzyBp4GwN ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-rs0-2 + local pod=some-name-rs0-2 + set +o xtrace waiting for pod/some-name-rs0-2 to be ready.OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QImc3fsPGC +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZR7ku7luEB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.QImc3fsPGC ++ cat /tmp/tmp.ZR7ku7luEB ++ rm /tmp/tmp.QImc3fsPGC /tmp/tmp.ZR7ku7luEB ++ return 0 + [[ '' == \t\r\u\e ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.v0KAM9kpRy +++ mktemp ++ local LAST_ERR=/tmp/tmp.mPNy6ZNWoy ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.v0KAM9kpRy ++ cat /tmp/tmp.mPNy6ZNWoy ++ rm /tmp/tmp.v0KAM9kpRy /tmp/tmp.mPNy6ZNWoy ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness + wait_for_running some-name-mongos 3 + local name=some-name-mongos + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=mongos + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-mongos-0 + local pod=some-name-mongos-0 + set +o xtrace waiting for pod/some-name-mongos-0 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-mongos-1 + local pod=some-name-mongos-1 + set +o xtrace waiting for pod/some-name-mongos-1 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qYFEOYnove +++ mktemp ++ local LAST_ERR=/tmp/tmp.xyCGwUuu3l ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.qYFEOYnove ++ cat /tmp/tmp.xyCGwUuu3l ++ rm /tmp/tmp.qYFEOYnove /tmp/tmp.xyCGwUuu3l ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-mongos-2 + local pod=some-name-mongos-2 + set +o xtrace waiting for pod/some-name-mongos-2 to be ready.OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wSJ6nzgtbm +++ mktemp ++ local LAST_ERR=/tmp/tmp.RQWK8ldeCq ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.wSJ6nzgtbm ++ cat /tmp/tmp.RQWK8ldeCq ++ rm /tmp/tmp.wSJ6nzgtbm /tmp/tmp.RQWK8ldeCq ++ return 0 + [[ '' == \t\r\u\e ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DlIitoMwUp +++ mktemp ++ local LAST_ERR=/tmp/tmp.mVCyr10EVh ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.DlIitoMwUp ++ cat /tmp/tmp.mVCyr10EVh ++ rm /tmp/tmp.DlIitoMwUp /tmp/tmp.mVCyr10EVh ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6R5QKWpYmb +++ mktemp ++ local LAST_ERR=/tmp/tmp.JzDs1Q3ePe ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.6R5QKWpYmb ++ cat /tmp/tmp.JzDs1Q3ePe ++ rm /tmp/tmp.6R5QKWpYmb /tmp/tmp.JzDs1Q3ePe ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo .OK .OK + compare_generation 3 statefulset some-name-cfg + local generation=3 + local resource_type=statefulset + local resource_name=some-name-cfg + local current_generation ++ kubectl_bin get statefulset some-name-cfg -o 'jsonpath={.metadata.generation}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VVRkI5ET2H +++ mktemp ++ local LAST_ERR=/tmp/tmp.ndVD2gaP7v ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get statefulset some-name-cfg -o 'jsonpath={.metadata.generation}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.VVRkI5ET2H ++ cat /tmp/tmp.ndVD2gaP7v ++ rm /tmp/tmp.VVRkI5ET2H /tmp/tmp.ndVD2gaP7v ++ return 0 + current_generation=3 + [[ 3 != \3 ]] + compare_generation 3 statefulset some-name-rs0 + local generation=3 + local resource_type=statefulset + local resource_name=some-name-rs0 + local current_generation ++ kubectl_bin get statefulset some-name-rs0 -o 'jsonpath={.metadata.generation}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eJ1fPFrLS2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.7cX2qmRNvx ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get statefulset some-name-rs0 -o 'jsonpath={.metadata.generation}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.eJ1fPFrLS2 ++ cat /tmp/tmp.7cX2qmRNvx ++ rm /tmp/tmp.eJ1fPFrLS2 /tmp/tmp.7cX2qmRNvx ++ return 0 + current_generation=3 + [[ 3 != \3 ]] + compare_generation 3 statefulset some-name-mongos + local generation=3 + local resource_type=statefulset + local resource_name=some-name-mongos + local current_generation ++ kubectl_bin get statefulset some-name-mongos -o 'jsonpath={.metadata.generation}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.We6i3BHNm5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.9jkBY6o11e ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get statefulset some-name-mongos -o 'jsonpath={.metadata.generation}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.We6i3BHNm5 ++ cat /tmp/tmp.9jkBY6o11e ++ rm /tmp/tmp.We6i3BHNm5 /tmp/tmp.9jkBY6o11e ++ return 0 + current_generation=3 + [[ 3 != \3 ]] + kubectl_bin patch psmdb/some-name --type=json '-p=[ {"op":"replace","path":"/spec/upgradeOptions/setFCV","value": true} ]' ++ mktemp + local LAST_OUT=/tmp/tmp.mdoE2aKuhQ ++ mktemp + local LAST_ERR=/tmp/tmp.5uDBdJ30U1 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl patch psmdb/some-name --type=json '-p=[ {"op":"replace","path":"/spec/upgradeOptions/setFCV","value": true} ]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.mdoE2aKuhQ perconaservermongodb.psmdb.percona.com/some-name patched (no change) + cat /tmp/tmp.5uDBdJ30U1 + rm /tmp/tmp.mdoE2aKuhQ /tmp/tmp.5uDBdJ30U1 + return 0 + sleep 10 + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1dgt7ogt5n +++ mktemp ++ local LAST_ERR=/tmp/tmp.C5gexN9scf ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.1dgt7ogt5n ++ cat /tmp/tmp.C5gexN9scf ++ rm /tmp/tmp.1dgt7ogt5n /tmp/tmp.C5gexN9scf ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo .OK .OK ++ run_mongos 'JSON.stringify(db.adminCommand({getParameter:1,featureCompatibilityVersion:1}))' clusterAdmin:clusterAdmin123456@some-name-cfg.mongod-major-upgrade-sharded-7355 ++ local 'command=JSON.stringify(db.adminCommand({getParameter:1,featureCompatibilityVersion:1}))' ++ local uri=clusterAdmin:clusterAdmin123456@some-name-cfg.mongod-major-upgrade-sharded-7355 ++ local driver=mongodb ++ local suffix=.svc.cluster.local ++ local mongo_flag= ++ local port=27017 ++ local mongo_bin=mongo ++ grep -E '^\{.*\}$' ++ jq -r .featureCompatibilityVersion.version +++ echo .svc.cluster.local +++ awk -F: '{print $2}' ++ suffix_port= ++ [[ -z '' ]] ++ suffix=.svc.cluster.local:27017 +++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.yMT2ACjq8a ++++ mktemp +++ local LAST_ERR=/tmp/tmp.hapoinMUn2 +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.yMT2ACjq8a +++ cat /tmp/tmp.hapoinMUn2 +++ rm /tmp/tmp.yMT2ACjq8a /tmp/tmp.hapoinMUn2 +++ return 0 ++ local client_container=psmdb-client-bb8b97679-zfsvh ++ kubectl_bin exec psmdb-client-bb8b97679-zfsvh -- bash -c 'printf '\''JSON.stringify(db.adminCommand({getParameter:1,featureCompatibilityVersion:1}))\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@some-name-cfg.mongod-major-upgrade-sharded-7355.svc.cluster.local:27017/admin ' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4ZKP5H8qZX +++ mktemp ++ local LAST_ERR=/tmp/tmp.qxm8sI5bN3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl exec psmdb-client-bb8b97679-zfsvh -- bash -c 'printf '\''JSON.stringify(db.adminCommand({getParameter:1,featureCompatibilityVersion:1}))\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@some-name-cfg.mongod-major-upgrade-sharded-7355.svc.cluster.local:27017/admin ' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4ZKP5H8qZX ++ cat /tmp/tmp.qxm8sI5bN3 ++ rm /tmp/tmp.4ZKP5H8qZX /tmp/tmp.qxm8sI5bN3 ++ return 0 + currentFCV=8.0 + [[ 8.0 != 8.0 ]] + run_mongos 'use myApp\n db.test.insert({ x: 100503 })' myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355 + local 'command=use myApp\n db.test.insert({ x: 100503 })' + local uri=myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355 + local driver=mongodb + local suffix=.svc.cluster.local + local mongo_flag= + local port=27017 + local mongo_bin=mongo ++ echo .svc.cluster.local ++ awk -F: '{print $2}' + suffix_port= + [[ -z '' ]] + suffix=.svc.cluster.local:27017 ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MTzjVSEKgn +++ mktemp ++ local LAST_ERR=/tmp/tmp.GPFWh7IlBu ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.MTzjVSEKgn ++ cat /tmp/tmp.GPFWh7IlBu ++ rm /tmp/tmp.MTzjVSEKgn /tmp/tmp.GPFWh7IlBu ++ return 0 + local client_container=psmdb-client-bb8b97679-zfsvh + kubectl_bin exec psmdb-client-bb8b97679-zfsvh -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100503 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355.svc.cluster.local:27017/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.EKuS2YznZW ++ mktemp + local LAST_ERR=/tmp/tmp.m2knFGFR2V + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-zfsvh -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100503 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355.svc.cluster.local:27017/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.EKuS2YznZW Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.mongod-major-upgrade-sharded-7355.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("6f6b215a-a15d-4e72-bf2d-e444eb5138f9") } Percona Server for MongoDB server version: v8.0.19-7 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.m2knFGFR2V + rm /tmp/tmp.EKuS2YznZW /tmp/tmp.m2knFGFR2V + return 0 + compare_mongos_cmd find myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355 -3 + local command=find + local uri=myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355 + local postfix=-3 + local suffix= + local database=myApp + local collection=test + local port=27017 + local tls=false + [[ false == \t\r\u\e ]] + mongos_command=run_mongos + log 'running db.test.command() in myApp' + set +o xtrace [2026-03-10T07:47:30+0000] running db.test.command() in myApp + run_mongos 'use myApp\n db.test.find()' myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355 mongodb '' '' 27017 + local 'command=use myApp\n db.test.find()' + local uri=myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355 + local driver=mongodb + local suffix=.svc.cluster.local + local mongo_flag= + local port=27017 + local mongo_bin=mongo + grep -E -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:|Started a new thread for the timer service' + /usr/sbin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ echo .svc.cluster.local ++ awk -F: '{print $2}' + suffix_port= + [[ -z '' ]] + suffix=.svc.cluster.local:27017 ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qwBES5rAjS +++ mktemp ++ local LAST_ERR=/tmp/tmp.UWL3LGbvMa ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.qwBES5rAjS ++ cat /tmp/tmp.UWL3LGbvMa ++ rm /tmp/tmp.qwBES5rAjS /tmp/tmp.UWL3LGbvMa ++ return 0 + local client_container=psmdb-client-bb8b97679-zfsvh + kubectl_bin exec psmdb-client-bb8b97679-zfsvh -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355.svc.cluster.local:27017/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.KMAC6SJvGg ++ mktemp + local LAST_ERR=/tmp/tmp.4XVdmMgTxW + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-bb8b97679-zfsvh -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.mongod-major-upgrade-sharded-7355.svc.cluster.local:27017/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.KMAC6SJvGg + cat /tmp/tmp.4XVdmMgTxW + rm /tmp/tmp.KMAC6SJvGg /tmp/tmp.4XVdmMgTxW + return 0 + [[ 0 -eq 0 ]] + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/e2e-tests/mongod-major-upgrade-sharded/compare/find-3.json /tmp/tmp.out8E8NpB2/find-3 + target_generation=4 + destroy mongod-major-upgrade-sharded-7355 + local namespace=mongod-major-upgrade-sharded-7355 + local ignore_logs=true + [[ 0 == 1 ]] + desc 'destroy cluster/operator and all other resources' + set +o xtrace ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + '[' true == false ']' + delete_backups + desc 'Delete psmdb-backup' + set +o xtrace ----------------------------------------------------------------------------------- Delete psmdb-backup ----------------------------------------------------------------------------------- ++ kubectl_bin get psmdb-backup --no-headers ++ wc -l +++ mktemp ++ local LAST_OUT=/tmp/tmp.HoUs5Ro1K2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.GNLFhQkjUU ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb-backup --no-headers ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.HoUs5Ro1K2 ++ cat /tmp/tmp.GNLFhQkjUU No resources found in mongod-major-upgrade-sharded-7355 namespace. ++ rm /tmp/tmp.HoUs5Ro1K2 /tmp/tmp.GNLFhQkjUU ++ return 0 + '[' 0 '!=' 0 ']' + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.FMDlVE0FLo ++ mktemp + local LAST_ERR=/tmp/tmp.3VWsuVF8Ww + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.FMDlVE0FLo customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.3VWsuVF8Ww + rm /tmp/tmp.FMDlVE0FLo /tmp/tmp.3VWsuVF8Ww + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/crd.yaml ++ grep -v '\-\-\-' grep: warning: stray \ before - grep: warning: stray \ before - + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.1gZV2KJLuy ++ mktemp + local LAST_ERR=/tmp/tmp.6k5C21hxxc + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.1gZV2KJLuy + cat /tmp/tmp.6k5C21hxxc + rm /tmp/tmp.1gZV2KJLuy /tmp/tmp.6k5C21hxxc + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.jvVBo3pWYE ++ mktemp + local LAST_ERR=/tmp/tmp.RtY4S5HSUw + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.jvVBo3pWYE + cat /tmp/tmp.RtY4S5HSUw + rm /tmp/tmp.jvVBo3pWYE /tmp/tmp.RtY4S5HSUw + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbs" + kubectl patch perconaservermongodbs.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbs" + : + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.DnA37sLCB4 ++ mktemp + local LAST_ERR=/tmp/tmp.nq5DS8Fedx + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.DnA37sLCB4 + cat /tmp/tmp.nq5DS8Fedx + rm /tmp/tmp.DnA37sLCB4 /tmp/tmp.nq5DS8Fedx + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.Hg31pGvqsG ++ mktemp + local LAST_ERR=/tmp/tmp.bIEi0Dthdn + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2266/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Hg31pGvqsG clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.bIEi0Dthdn + rm /tmp/tmp.Hg31pGvqsG /tmp/tmp.bIEi0Dthdn + return 0 + destroy_cert_manager + kubectl_bin delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.46dunCfnxI ++ mktemp + local LAST_ERR=/tmp/tmp.mUVYoVI6SL + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.46dunCfnxI + cat /tmp/tmp.mUVYoVI6SL Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 0 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.46dunCfnxI + cat /tmp/tmp.mUVYoVI6SL Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 4 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.46dunCfnxI + cat /tmp/tmp.mUVYoVI6SL Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 8 + cat /tmp/tmp.46dunCfnxI + cat /tmp/tmp.mUVYoVI6SL Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + rm /tmp/tmp.46dunCfnxI /tmp/tmp.mUVYoVI6SL + return 1 + true + '[' -n '' ']' + '[' -n psmdb-operator ']' + kubectl_bin delete --grace-period=0 --force=true namespace mongod-major-upgrade-sharded-7355 + rm -rf /tmp/tmp.out8E8NpB2 + kubectl_bin delete --grace-period=0 --force=true namespace psmdb-operator ++ mktemp ++ mktemp + desc 'test passed' + set +o xtrace ----------------------------------------------------------------------------------- test passed ----------------------------------------------------------------------------------- + local LAST_OUT=/tmp/tmp.Dm7SnEI9UZ + local LAST_OUT=/tmp/tmp.y6X0sL2JBo ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.S8EUznbCqP + local exit_status=0 + local timeout=4 + local LAST_ERR=/tmp/tmp.wXb1pFeMWk + local exit_status=0 + local timeout=4 ++ seq 0 2 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete --grace-period=0 --force=true namespace mongod-major-upgrade-sharded-7355 + for i in $(seq 0 2) + set +e + kubectl delete --grace-period=0 --force=true namespace psmdb-operator