Log: /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/logs/replset-overrides.log grep: warning: stray \ before - Warning: version difference between client (1.34) and server (1.30) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.34) and server (1.30) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.34) and server (1.30) exceeds the supported minor version skew of +/-1 + cluster=some-name + main + create_infra replset-overrides-31757 + local ns=replset-overrides-31757 + [[ 1 == 1 ]] + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.zU5Q9hXXxw ++ mktemp + local LAST_ERR=/tmp/tmp.CjcJYUbTBm + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.zU5Q9hXXxw customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.CjcJYUbTBm + rm /tmp/tmp.zU5Q9hXXxw /tmp/tmp.CjcJYUbTBm + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/deploy/crd.yaml ++ grep -v '\-\-\-' grep: warning: stray \ before - grep: warning: stray \ before - + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n replset-overrides-6781 backup-minio-physical --type=merge -p '{"metadata":{"finalizers":[]}}' perconaservermongodbbackup.psmdb.percona.com/backup-minio-physical patched + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.NQKHnfggvQ ++ mktemp + local LAST_ERR=/tmp/tmp.CrshHVyPHo + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.NQKHnfggvQ customresourcedefinition.apiextensions.k8s.io/perconaservermongodbbackups.psmdb.percona.com condition met + cat /tmp/tmp.CrshHVyPHo + rm /tmp/tmp.NQKHnfggvQ /tmp/tmp.CrshHVyPHo + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.1mwI8ROofn ++ mktemp + local LAST_ERR=/tmp/tmp.AykJyORnrL + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.1mwI8ROofn + cat /tmp/tmp.AykJyORnrL + rm /tmp/tmp.1mwI8ROofn /tmp/tmp.AykJyORnrL + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbs" + kubectl patch perconaservermongodbs.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbs" + : + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.1kSFPt1qmx ++ mktemp + local LAST_ERR=/tmp/tmp.YfM2pBdsuO + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.1kSFPt1qmx + cat /tmp/tmp.YfM2pBdsuO + rm /tmp/tmp.1kSFPt1qmx /tmp/tmp.YfM2pBdsuO + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.lbaplJ1oOz ++ mktemp + local LAST_ERR=/tmp/tmp.MhjCLMYTd2 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.lbaplJ1oOz clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.MhjCLMYTd2 + rm /tmp/tmp.lbaplJ1oOz /tmp/tmp.MhjCLMYTd2 + return 0 + check_crd_for_deletion PR-2045-4cbffc79 + local git_tag=PR-2045-4cbffc79 ++ curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/PR-2045-4cbffc79/deploy/crd.yaml ++ yq eval .metadata.name ++ /usr/sbin/sed s/---//g ++ /usr/sbin/sed ':a;N;$!ba;s/\n/ /g' + for crd_name in $(curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/${git_tag}/deploy/crd.yaml | yq eval '.metadata.name' | $sed 's/---//g' | $sed ':a;N;$!ba;s/\n/ /g') ++ kubectl_bin get crd/null -o 'jsonpath={.status.conditions[-1].type}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dmv3FawXnN +++ mktemp ++ local LAST_ERR=/tmp/tmp.kTZZG5IvUT ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.dmv3FawXnN ++ cat /tmp/tmp.kTZZG5IvUT Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 0 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.dmv3FawXnN ++ cat /tmp/tmp.kTZZG5IvUT Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 4 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.dmv3FawXnN ++ cat /tmp/tmp.kTZZG5IvUT Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 8 ++ cat /tmp/tmp.dmv3FawXnN ++ cat /tmp/tmp.kTZZG5IvUT Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ rm /tmp/tmp.dmv3FawXnN /tmp/tmp.kTZZG5IvUT ++ return 1 + [[ '' == \T\e\r\m\i\n\a\t\i\n\g ]] + '[' -n psmdb-operator ']' + create_namespace psmdb-operator + local namespace=psmdb-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + egrep -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' + awk '{print$1}' + '[' -n '' ']' + desc 'cleaned up old namespaces psmdb-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace psmdb-operator --ignore-not-found ++ mktemp + xargs kubectl delete ns egrep: warning: egrep is obsolescent; using grep -E + local LAST_OUT=/tmp/tmp.FTjjSZDZ3l ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.sT2z9ysDlq + local LAST_ERR=/tmp/tmp.hB0MNSfbia + local exit_status=0 + local timeout=4 ++ mktemp ++ seq 0 2 + for i in $(seq 0 2) + local LAST_ERR=/tmp/tmp.VDoGk4k9AV + set +e + local exit_status=0 + local timeout=4 + kubectl get ns ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete namespace psmdb-operator --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.FTjjSZDZ3l + cat /tmp/tmp.hB0MNSfbia + rm /tmp/tmp.FTjjSZDZ3l /tmp/tmp.hB0MNSfbia + return 0 namespace "replset-overrides-6781" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.sT2z9ysDlq namespace "psmdb-operator" deleted + cat /tmp/tmp.VDoGk4k9AV + rm /tmp/tmp.sT2z9ysDlq /tmp/tmp.VDoGk4k9AV + return 0 + kubectl_bin wait --for=delete namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.9XUoLkFnVj ++ mktemp + local LAST_ERR=/tmp/tmp.GKjJSowih7 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.9XUoLkFnVj + cat /tmp/tmp.GKjJSowih7 + rm /tmp/tmp.9XUoLkFnVj /tmp/tmp.GKjJSowih7 + return 0 + desc 'create namespace psmdb-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.E96YM5xYKZ ++ mktemp + local LAST_ERR=/tmp/tmp.XwMhJbm8vS + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.E96YM5xYKZ namespace/psmdb-operator created + cat /tmp/tmp.XwMhJbm8vS + rm /tmp/tmp.E96YM5xYKZ /tmp/tmp.XwMhJbm8vS + return 0 + set_kube_ctx psmdb-operator + local namespace=psmdb-operator ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.uTPM2R7Zl2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Y1M5rXeJom ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.uTPM2R7Zl2 ++ cat /tmp/tmp.Y1M5rXeJom ++ rm /tmp/tmp.uTPM2R7Zl2 /tmp/tmp.Y1M5rXeJom ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2045-4cbffc79-4-cluster1 --namespace=psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.h5WqKitxOS ++ mktemp + local LAST_ERR=/tmp/tmp.mbOza1CHdg + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2045-4cbffc79-4-cluster1 --namespace=psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.h5WqKitxOS Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2045-4cbffc79-4-cluster1" modified. + cat /tmp/tmp.mbOza1CHdg + rm /tmp/tmp.h5WqKitxOS /tmp/tmp.mbOza1CHdg + return 0 + deploy_operator + desc 'start PSMDB operator' + set +o xtrace ----------------------------------------------------------------------------------- start PSMDB operator ----------------------------------------------------------------------------------- + local cr_file + '[' -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/conf/crd.yaml ']' + cr_file=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/deploy/crd.yaml + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.ionkoAaEqI ++ mktemp + local LAST_ERR=/tmp/tmp.q1cUSdIIMr + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ionkoAaEqI customresourcedefinition.apiextensions.k8s.io/perconaservermongodbbackups.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbrestores.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com serverside-applied + cat /tmp/tmp.q1cUSdIIMr + rm /tmp/tmp.ionkoAaEqI /tmp/tmp.q1cUSdIIMr + return 0 + '[' -n psmdb-operator ']' + apply_rbac cw-rbac + local operator_namespace=psmdb-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: psmdb-operator^' + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.QiQQ5QnyyD ++ mktemp + local LAST_ERR=/tmp/tmp.c9SuiV6iqU + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.QiQQ5QnyyD clusterrole.rbac.authorization.k8s.io/percona-server-mongodb-operator created serviceaccount/percona-server-mongodb-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-server-mongodb-operator created + cat /tmp/tmp.c9SuiV6iqU + rm /tmp/tmp.QiQQ5QnyyD /tmp/tmp.c9SuiV6iqU + return 0 + yq eval ' (.spec.template.spec.containers[].image = "perconalab/percona-server-mongodb-operator:PR-2045-4cbffc79") | ((.. | select(.[] == "DISABLE_TELEMETRY")) |= .value="true") | ((.. | select(.[] == "LOG_LEVEL")) |= .value="DEBUG")' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/deploy/cw-operator.yaml + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.7Ct0b8Jf30 ++ mktemp + local LAST_ERR=/tmp/tmp.KinvbJySOD + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.7Ct0b8Jf30 deployment.apps/percona-server-mongodb-operator created + cat /tmp/tmp.KinvbJySOD + rm /tmp/tmp.7Ct0b8Jf30 /tmp/tmp.KinvbJySOD + return 0 + sleep 2 ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.FgV2fFAuuO +++ mktemp ++ local LAST_ERR=/tmp/tmp.Dh1kidi2Nj ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.FgV2fFAuuO ++ cat /tmp/tmp.Dh1kidi2Nj ++ rm /tmp/tmp.FgV2fFAuuO /tmp/tmp.Dh1kidi2Nj ++ return 0 + wait_pod percona-server-mongodb-operator-787f9b79bc-8x4rt + local pod=percona-server-mongodb-operator-787f9b79bc-8x4rt + set +o xtrace waiting for pod/percona-server-mongodb-operator-787f9b79bc-8x4rt to be ready..OK + echo 'Print operator info from log' Print operator info from log + grep 'Manager starting up' ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.lVFI9GJ4nL +++ mktemp ++ local LAST_ERR=/tmp/tmp.gHuujaRpjQ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.lVFI9GJ4nL ++ cat /tmp/tmp.gHuujaRpjQ ++ rm /tmp/tmp.lVFI9GJ4nL /tmp/tmp.gHuujaRpjQ ++ return 0 + kubectl_bin logs percona-server-mongodb-operator-787f9b79bc-8x4rt ++ mktemp + local LAST_OUT=/tmp/tmp.tJvoSkM1qM ++ mktemp + local LAST_ERR=/tmp/tmp.O8PuVTr4jp + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl logs percona-server-mongodb-operator-787f9b79bc-8x4rt + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.tJvoSkM1qM + cat /tmp/tmp.O8PuVTr4jp + rm /tmp/tmp.tJvoSkM1qM /tmp/tmp.O8PuVTr4jp + return 0 2025-09-16T15:05:15.796Z INFO setup Manager starting up {"gitCommit": "4cbffc79dc7467dd090ba165bdb1d88c4552480a", "gitBranch": "PR-2045-4cbffc79", "buildTime": "", "goVersion": "go1.25.1", "os": "linux", "arch": "amd64"} + create_namespace replset-overrides-31757 + local namespace=replset-overrides-31757 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + egrep -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' + awk '{print$1}' ++ mktemp + '[' -n '' ']' + desc 'cleaned up old namespaces replset-overrides-31757' + xargs kubectl delete ns + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces replset-overrides-31757 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace replset-overrides-31757 --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.JIAwKsmmLl egrep: warning: egrep is obsolescent; using grep -E ++ mktemp + local LAST_OUT=/tmp/tmp.EVBuWExDlr ++ mktemp + local LAST_ERR=/tmp/tmp.vH9VlgCqAD + local exit_status=0 + local timeout=4 ++ seq 0 2 + local LAST_ERR=/tmp/tmp.jtAumHaAf4 + local exit_status=0 + local timeout=4 + for i in $(seq 0 2) + set +e + kubectl get ns ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete namespace replset-overrides-31757 --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.JIAwKsmmLl + cat /tmp/tmp.vH9VlgCqAD + rm /tmp/tmp.JIAwKsmmLl /tmp/tmp.vH9VlgCqAD + return 0 error: resource(s) were provided, but no name was specified + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.EVBuWExDlr + cat /tmp/tmp.jtAumHaAf4 + rm /tmp/tmp.EVBuWExDlr /tmp/tmp.jtAumHaAf4 + return 0 + kubectl_bin wait --for=delete namespace replset-overrides-31757 ++ mktemp + local LAST_OUT=/tmp/tmp.FtLmtxvnUL ++ mktemp + local LAST_ERR=/tmp/tmp.Mr4oNehbyV + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete namespace replset-overrides-31757 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.FtLmtxvnUL + cat /tmp/tmp.Mr4oNehbyV + rm /tmp/tmp.FtLmtxvnUL /tmp/tmp.Mr4oNehbyV + return 0 + desc 'create namespace replset-overrides-31757' + set +o xtrace ----------------------------------------------------------------------------------- create namespace replset-overrides-31757 ----------------------------------------------------------------------------------- + kubectl_bin create namespace replset-overrides-31757 ++ mktemp + local LAST_OUT=/tmp/tmp.3sAipsBQnl ++ mktemp + local LAST_ERR=/tmp/tmp.dMFWtOPuFS + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace replset-overrides-31757 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.3sAipsBQnl namespace/replset-overrides-31757 created + cat /tmp/tmp.dMFWtOPuFS + rm /tmp/tmp.3sAipsBQnl /tmp/tmp.dMFWtOPuFS + return 0 + set_kube_ctx replset-overrides-31757 + local namespace=replset-overrides-31757 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.z39elq0WxK +++ mktemp ++ local LAST_ERR=/tmp/tmp.qlNLviCjDh ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.z39elq0WxK ++ cat /tmp/tmp.qlNLviCjDh ++ rm /tmp/tmp.z39elq0WxK /tmp/tmp.qlNLviCjDh ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2045-4cbffc79-4-cluster1 --namespace=replset-overrides-31757 ++ mktemp + local LAST_OUT=/tmp/tmp.woAenIfhUE ++ mktemp + local LAST_ERR=/tmp/tmp.sqiBNSy5aT + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2045-4cbffc79-4-cluster1 --namespace=replset-overrides-31757 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.woAenIfhUE Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2045-4cbffc79-4-cluster1" modified. + cat /tmp/tmp.sqiBNSy5aT + rm /tmp/tmp.woAenIfhUE /tmp/tmp.sqiBNSy5aT + return 0 + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.9nLn3wEPCl ++ mktemp + local LAST_ERR=/tmp/tmp.y61qJlwvoq + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/conf/client.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.9nLn3wEPCl deployment.apps/psmdb-client created + cat /tmp/tmp.y61qJlwvoq + rm /tmp/tmp.9nLn3wEPCl /tmp/tmp.y61qJlwvoq + return 0 + deploy_minio + desc 'install Minio' + set +o xtrace ----------------------------------------------------------------------------------- install Minio ----------------------------------------------------------------------------------- + helm uninstall minio-service Error: uninstall: Release not loaded: minio-service: release: not found + : + helm repo remove minio "minio" has been removed from your repositories + helm repo add minio https://charts.min.io/ "minio" has been added to your repositories + retry 10 60 helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/.minio/ --set persistence.size=2G --set securityContext.enabled=false minio/minio + local max=10 + local delay=60 + shift 2 + local n=1 + helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/.minio/ --set persistence.size=2G --set securityContext.enabled=false minio/minio NAME: minio-service LAST DEPLOYED: Tue Sep 16 15:05:37 2025 NAMESPACE: replset-overrides-31757 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: MinIO can be accessed via port 9000 on the following DNS name from within your cluster: minio-service.replset-overrides-31757.cluster.local To access MinIO from localhost, run the below commands: 1. export POD_NAME=$(kubectl get pods --namespace replset-overrides-31757 -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}") 2. kubectl port-forward $POD_NAME 9000 --namespace replset-overrides-31757 Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/ You can now access MinIO server on http://localhost:9000. Follow the below steps to connect to MinIO server with mc client: 1. Download the MinIO mc client - https://min.io/docs/minio/linux/reference/minio-mc.html#quickstart 2. export MC_HOST_minio-service-local=http://$(kubectl get secret --namespace replset-overrides-31757 minio-service -o jsonpath="{.data.rootUser}" | base64 --decode):$(kubectl get secret --namespace replset-overrides-31757 minio-service -o jsonpath="{.data.rootPassword}" | base64 --decode)@localhost:9000 3. mc ls minio-service-local ++ kubectl_bin get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ucCq92mx01 +++ mktemp ++ local LAST_ERR=/tmp/tmp.8c2u8vnQ3t ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ucCq92mx01 ++ cat /tmp/tmp.8c2u8vnQ3t ++ rm /tmp/tmp.ucCq92mx01 /tmp/tmp.8c2u8vnQ3t ++ return 0 + MINIO_POD=minio-service-86dfccd949-nhrw2 + wait_pod minio-service-86dfccd949-nhrw2 + local pod=minio-service-86dfccd949-nhrw2 + set +o xtrace waiting for pod/minio-service-86dfccd949-nhrw2 to be ready.OK + '[' -n psmdb-operator ']' + kubectl_bin create svc -n psmdb-operator externalname minio-service --external-name=minio-service.replset-overrides-31757.svc.cluster.local --tcp=9000 ++ mktemp + local LAST_OUT=/tmp/tmp.ZwAPQcadcc ++ mktemp + local LAST_ERR=/tmp/tmp.pHHKFV7wY8 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create svc -n psmdb-operator externalname minio-service --external-name=minio-service.replset-overrides-31757.svc.cluster.local --tcp=9000 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ZwAPQcadcc service/minio-service created + cat /tmp/tmp.pHHKFV7wY8 + rm /tmp/tmp.ZwAPQcadcc /tmp/tmp.pHHKFV7wY8 + return 0 + create_minio_bucket operator-testing + local bucket=operator-testing + kubectl_bin run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- bash -c 'AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url http://minio-service:9000 s3 mb s3://operator-testing' ++ mktemp + local LAST_OUT=/tmp/tmp.RBbufzjgdl ++ mktemp + local LAST_ERR=/tmp/tmp.6l16TP4jAM + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- bash -c 'AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url http://minio-service:9000 s3 mb s3://operator-testing' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.RBbufzjgdl pod "aws-cli" deleted from replset-overrides-31757 namespace + cat /tmp/tmp.6l16TP4jAM All commands and output from this session will be recorded in container logs, including credentials and sensitive information passed through the command prompt. If you don't see a command prompt, try pressing enter. + rm /tmp/tmp.RBbufzjgdl /tmp/tmp.6l16TP4jAM + return 0 + desc 'Case 1: Deploying a new cluster with hostname overrides' + set +o xtrace ----------------------------------------------------------------------------------- Case 1: Deploying a new cluster with hostname overrides ----------------------------------------------------------------------------------- + test_deploy_with_host_overrides + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/conf/secrets_with_tls.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/conf/minio-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.DlzA36Oexs ++ mktemp + local LAST_ERR=/tmp/tmp.3BzlVlyKcG + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/conf/secrets_with_tls.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/conf/minio-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.DlzA36Oexs secret/some-users created secret/some-name-ssl created secret/some-name-ssl-internal created secret/minio-secret created + cat /tmp/tmp.3BzlVlyKcG + rm /tmp/tmp.DlzA36Oexs /tmp/tmp.3BzlVlyKcG + return 0 + echo 'creating external services' creating external services + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/conf/external-services.yml ++ mktemp + local LAST_OUT=/tmp/tmp.23akolOshE ++ mktemp + local LAST_ERR=/tmp/tmp.6iHcSTy8KQ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/conf/external-services.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.23akolOshE service/external-rs0-0 created service/external-rs0-1 created service/external-rs0-2 created + cat /tmp/tmp.6iHcSTy8KQ + rm /tmp/tmp.23akolOshE /tmp/tmp.6iHcSTy8KQ + return 0 + echo 'creating PSMDB cluster: some-name' creating PSMDB cluster: some-name + sed s/NAMESPACE/replset-overrides-31757/g /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/conf/some-name-overridden.yml + apply_cluster /tmp/tmp.GJOJy85nvw/some-name-overridden.yml + '[' -z '' ']' + cat_config /tmp/tmp.GJOJy85nvw/some-name-overridden.yml + kubectl_bin apply -f - + cat /tmp/tmp.GJOJy85nvw/some-name-overridden.yml + yq eval '(.spec | select(.image == null)).image = "perconalab/percona-server-mongodb-operator:main-mongod7.0"' ++ mktemp + yq eval '(.spec | select(has("backup"))).backup.image = "perconalab/percona-server-mongodb-operator:main-backup"' + yq eval '(.spec | select(has("pmm"))).pmm.image = "percona/pmm-client:2.44.1-1"' + yq eval '(.spec | select(has("initImage"))).initImage = "perconalab/percona-server-mongodb-operator:PR-2045-4cbffc79"' + yq eval '.spec.upgradeOptions.apply="Never"' + local LAST_OUT=/tmp/tmp.6RVLOBNNrz ++ mktemp + local LAST_ERR=/tmp/tmp.jv9NrTSBZ6 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.6RVLOBNNrz perconaservermongodb.psmdb.percona.com/some-name created + cat /tmp/tmp.jv9NrTSBZ6 + rm /tmp/tmp.6RVLOBNNrz /tmp/tmp.jv9NrTSBZ6 + return 0 + wait_for_running some-name-rs0 3 + local name=some-name-rs0 + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=rs0 + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-rs0-0 + local pod=some-name-rs0-0 + set +o xtrace waiting for pod/some-name-rs0-0 to be ready..............OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-rs0-1 + local pod=some-name-rs0-1 + set +o xtrace waiting for pod/some-name-rs0-1 to be ready......OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DUvBjwUu3K +++ mktemp ++ local LAST_ERR=/tmp/tmp.NVowUUCS6t ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.DUvBjwUu3K ++ cat /tmp/tmp.NVowUUCS6t ++ rm /tmp/tmp.DUvBjwUu3K /tmp/tmp.NVowUUCS6t ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-rs0-2 + local pod=some-name-rs0-2 + set +o xtrace waiting for pod/some-name-rs0-2 to be ready.....OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].non_voting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5dkuYZDjah +++ mktemp ++ local LAST_ERR=/tmp/tmp.hxbvLA41ql ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].non_voting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5dkuYZDjah ++ cat /tmp/tmp.hxbvLA41ql ++ rm /tmp/tmp.5dkuYZDjah /tmp/tmp.hxbvLA41ql ++ return 0 + [[ '' == \t\r\u\e ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8VWS1Gs8Fm +++ mktemp ++ local LAST_ERR=/tmp/tmp.388O9M0WoQ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.8VWS1Gs8Fm ++ cat /tmp/tmp.388O9M0WoQ ++ rm /tmp/tmp.8VWS1Gs8Fm /tmp/tmp.388O9M0WoQ ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness...... + echo 'writing some data' writing some data + run_mongo 'use myApp\n db.test.insert({ x: 100500 })' databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 + local 'command=use myApp\n db.test.insert({ x: 100500 })' + local uri=databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 + local driver=mongodb+srv + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LWFjcQEaV8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.85EYIHXnIF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.LWFjcQEaV8 ++ cat /tmp/tmp.85EYIHXnIF ++ rm /tmp/tmp.LWFjcQEaV8 /tmp/tmp.85EYIHXnIF ++ return 0 + local client_container=psmdb-client-66f577db5f-94kg8 + local mongo_flag= + [[ databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 == *cfg* ]] + replica_set=rs0 + kubectl_bin exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb+srv://databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.FZyYNUmM86 ++ mktemp + local LAST_ERR=/tmp/tmp.fDnYuiLP3E + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb+srv://databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.FZyYNUmM86 Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://external-rs0-0.replset-overrides-31757.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb&replicaSet=rs0&ssl=false Implicit session: session { "id" : UUID("caa2b8cd-50a0-4da3-8282-07b4524c0ce4") } Percona Server for MongoDB server version: v8.0.12-4 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.fDnYuiLP3E + rm /tmp/tmp.FZyYNUmM86 /tmp/tmp.fDnYuiLP3E + return 0 + compare_mongo_cmd find databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 + local command=find + local uri=databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 + local postfix= + local suffix= + local database=myApp + local collection=test + local sort= + local tls=false + local 'full_command=db.test.find()' + [[ -n '' ]] + log 'running db.test.find() in myApp' + set +o xtrace [2025-09-16T15:08:28+0000] running db.test.find() in myApp + [[ false == \t\r\u\e ]] + mongo_command=run_mongo + run_mongo 'use myApp\n db.test.find()' databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 mongodb '' + local 'command=use myApp\n db.test.find()' + egrep -v 'I NETWORK|W NETWORK|F NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:|Started a new thread for the timer service' + local uri=databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 + local driver=mongodb + local suffix=.svc.cluster.local + /usr/sbin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp egrep: warning: egrep is obsolescent; using grep -E ++ local LAST_OUT=/tmp/tmp.MvuIsxJrJI +++ mktemp ++ local LAST_ERR=/tmp/tmp.sBxv8OL9De ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.MvuIsxJrJI ++ cat /tmp/tmp.sBxv8OL9De ++ rm /tmp/tmp.MvuIsxJrJI /tmp/tmp.sBxv8OL9De ++ return 0 + local client_container=psmdb-client-66f577db5f-94kg8 + local mongo_flag= + [[ databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 == *cfg* ]] + replica_set=rs0 + kubectl_bin exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.jcxicAh45Q ++ mktemp + local LAST_ERR=/tmp/tmp.s7PrzLcaGU + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.jcxicAh45Q + cat /tmp/tmp.s7PrzLcaGU + rm /tmp/tmp.jcxicAh45Q /tmp/tmp.s7PrzLcaGU + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/compare/find.json /tmp/tmp.GJOJy85nvw/find + run_backup minio backup-minio-physical physical + local storage=minio + local backup_name=backup-minio-physical + local type=physical + desc 'run backup backup-minio-physical' + set +o xtrace ----------------------------------------------------------------------------------- run backup backup-minio-physical ----------------------------------------------------------------------------------- + yq eval '.metadata.name = "backup-minio-physical" | .spec.storageName = "minio" | .spec.type = "physical"' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/conf/backup-minio.yml + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.azbkCIEYsY ++ mktemp + local LAST_ERR=/tmp/tmp.C2DnoNkCRu + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.azbkCIEYsY perconaservermongodbbackup.psmdb.percona.com/backup-minio-physical created + cat /tmp/tmp.C2DnoNkCRu + rm /tmp/tmp.azbkCIEYsY /tmp/tmp.C2DnoNkCRu + return 0 + wait_backup backup-minio-physical + local backup_name=backup-minio-physical + local target_state=ready + set +o xtrace waiting for backup-minio-physical to reach ready state...... + run_recovery_check some-name backup-minio-physical external-rs0-0.replset-overrides-31757 + local cluster=some-name + local backup=backup-minio-physical + local host=external-rs0-0.replset-overrides-31757 + echo 'write data after backup' write data after backup + run_mongo 'use myApp\n db.test.insert({ x: 100501 })' databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 + local 'command=use myApp\n db.test.insert({ x: 100501 })' + local uri=databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 + local driver=mongodb+srv + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DvGMHQesFN +++ mktemp ++ local LAST_ERR=/tmp/tmp.7d0iw5zN4Z ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.DvGMHQesFN ++ cat /tmp/tmp.7d0iw5zN4Z ++ rm /tmp/tmp.DvGMHQesFN /tmp/tmp.7d0iw5zN4Z ++ return 0 + local client_container=psmdb-client-66f577db5f-94kg8 + local mongo_flag= + [[ databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 == *cfg* ]] + replica_set=rs0 + kubectl_bin exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb+srv://databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.yc13ohhEpG ++ mktemp + local LAST_ERR=/tmp/tmp.Fd6r7CXWWg + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb+srv://databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.yc13ohhEpG Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://external-rs0-0.replset-overrides-31757.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb&replicaSet=rs0&ssl=false Implicit session: session { "id" : UUID("86ebfb29-2308-4282-917f-4eed35e46169") } Percona Server for MongoDB server version: v8.0.12-4 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.Fd6r7CXWWg + rm /tmp/tmp.yc13ohhEpG /tmp/tmp.Fd6r7CXWWg + return 0 + echo 'restore to backup backup-minio-physical' restore to backup backup-minio-physical + compare_mongo_cmd find databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 -2nd + local command=find + local uri=databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 + local postfix=-2nd + local suffix= + local database=myApp + local collection=test + local sort= + local tls=false + local 'full_command=db.test.find()' + [[ -n '' ]] + log 'running db.test.find() in myApp' + set +o xtrace [2025-09-16T15:08:47+0000] running db.test.find() in myApp + [[ false == \t\r\u\e ]] + mongo_command=run_mongo + run_mongo 'use myApp\n db.test.find()' databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 mongodb '' + local 'command=use myApp\n db.test.find()' + local uri=databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 + local driver=mongodb + local suffix=.svc.cluster.local + egrep -v 'I NETWORK|W NETWORK|F NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:|Started a new thread for the timer service' + /usr/sbin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp egrep: warning: egrep is obsolescent; using grep -E ++ local LAST_OUT=/tmp/tmp.tnC5zUO4Ub +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZtGWGnQzXy ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.tnC5zUO4Ub ++ cat /tmp/tmp.ZtGWGnQzXy ++ rm /tmp/tmp.tnC5zUO4Ub /tmp/tmp.ZtGWGnQzXy ++ return 0 + local client_container=psmdb-client-66f577db5f-94kg8 + local mongo_flag= + [[ databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 == *cfg* ]] + replica_set=rs0 + kubectl_bin exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.N0hjNSwBrF ++ mktemp + local LAST_ERR=/tmp/tmp.lpvorrcB02 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.N0hjNSwBrF + cat /tmp/tmp.lpvorrcB02 + rm /tmp/tmp.N0hjNSwBrF /tmp/tmp.lpvorrcB02 + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/compare/find-2nd.json /tmp/tmp.GJOJy85nvw/find-2nd + run_restore backup-minio-physical + local backup_name=backup-minio-physical + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/conf/restore.yml + /usr/sbin/sed -e 's/name:/name: restore-backup-minio-physical/' + /usr/sbin/sed -e 's/backupName:/backupName: backup-minio-physical/' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.sPXmYvwKxF ++ mktemp + local LAST_ERR=/tmp/tmp.I2882ul22m + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.sPXmYvwKxF perconaservermongodbrestore.psmdb.percona.com/restore-backup-minio-physical created + cat /tmp/tmp.I2882ul22m + rm /tmp/tmp.sPXmYvwKxF /tmp/tmp.I2882ul22m + return 0 + wait_restore backup-minio-physical some-name + local backup_name=backup-minio-physical + local cluster_name=some-name + local target_state=ready + local wait_cluster_consistency=1 + local wait_time=1780 + local ok_if_ready=0 + set +o xtrace Waiting for the psmdb-restore/restore-backup-minio-physical object to be createdOK Waiting psmdb-restore/restore-backup-minio-physical to reach state "ready" ...OK after 3 minutes + [[ 1 -eq 1 ]] + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HvZRkBzl6I +++ mktemp ++ local LAST_ERR=/tmp/tmp.7O2cc153sU ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.HvZRkBzl6I ++ cat /tmp/tmp.7O2cc153sU ++ rm /tmp/tmp.HvZRkBzl6I /tmp/tmp.7O2cc153sU ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 1 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tjjF8wKWUo +++ mktemp ++ local LAST_ERR=/tmp/tmp.ySTP7NxrPU ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.tjjF8wKWUo ++ cat /tmp/tmp.ySTP7NxrPU ++ rm /tmp/tmp.tjjF8wKWUo /tmp/tmp.ySTP7NxrPU ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 2 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EUWAtdF31z +++ mktemp ++ local LAST_ERR=/tmp/tmp.d6sOAggFzZ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.EUWAtdF31z ++ cat /tmp/tmp.d6sOAggFzZ ++ rm /tmp/tmp.EUWAtdF31z /tmp/tmp.d6sOAggFzZ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 3 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oFmapXcdqW +++ mktemp ++ local LAST_ERR=/tmp/tmp.secfNG2NgX ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.oFmapXcdqW ++ cat /tmp/tmp.secfNG2NgX ++ rm /tmp/tmp.oFmapXcdqW /tmp/tmp.secfNG2NgX ++ return 0 + [[ error == \r\e\a\d\y ]] + let retry+=1 + '[' 4 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4tNEZGLq3H +++ mktemp ++ local LAST_ERR=/tmp/tmp.3E2ALgbkuW ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4tNEZGLq3H ++ cat /tmp/tmp.3E2ALgbkuW ++ rm /tmp/tmp.4tNEZGLq3H /tmp/tmp.3E2ALgbkuW ++ return 0 + [[ error == \r\e\a\d\y ]] + let retry+=1 + '[' 5 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jvjHROpI2Q +++ mktemp ++ local LAST_ERR=/tmp/tmp.7d5F6DtBgS ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jvjHROpI2Q ++ cat /tmp/tmp.7d5F6DtBgS ++ rm /tmp/tmp.jvjHROpI2Q /tmp/tmp.7d5F6DtBgS ++ return 0 + [[ error == \r\e\a\d\y ]] + let retry+=1 + '[' 6 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QParNyhFgx +++ mktemp ++ local LAST_ERR=/tmp/tmp.FIhHpPDAP1 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.QParNyhFgx ++ cat /tmp/tmp.FIhHpPDAP1 ++ rm /tmp/tmp.QParNyhFgx /tmp/tmp.FIhHpPDAP1 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 7 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xitGEnm1HI +++ mktemp ++ local LAST_ERR=/tmp/tmp.lGIbXzKeNB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xitGEnm1HI ++ cat /tmp/tmp.lGIbXzKeNB ++ rm /tmp/tmp.xitGEnm1HI /tmp/tmp.lGIbXzKeNB ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 8 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mlqkkKmr5b +++ mktemp ++ local LAST_ERR=/tmp/tmp.wcnHXkWuKh ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.mlqkkKmr5b ++ cat /tmp/tmp.wcnHXkWuKh ++ rm /tmp/tmp.mlqkkKmr5b /tmp/tmp.wcnHXkWuKh ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 9 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UAzrYAb5zo +++ mktemp ++ local LAST_ERR=/tmp/tmp.CCSq2gfVkV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.UAzrYAb5zo ++ cat /tmp/tmp.CCSq2gfVkV ++ rm /tmp/tmp.UAzrYAb5zo /tmp/tmp.CCSq2gfVkV ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 10 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DZ36EacxC7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.pDTtIUbx4R ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.DZ36EacxC7 ++ cat /tmp/tmp.pDTtIUbx4R ++ rm /tmp/tmp.DZ36EacxC7 /tmp/tmp.pDTtIUbx4R ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 11 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RtH31mhy7R +++ mktemp ++ local LAST_ERR=/tmp/tmp.eeARdsldPj ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.RtH31mhy7R ++ cat /tmp/tmp.eeARdsldPj ++ rm /tmp/tmp.RtH31mhy7R /tmp/tmp.eeARdsldPj ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 12 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rD5w2wv6jj +++ mktemp ++ local LAST_ERR=/tmp/tmp.MOpmrwbXz8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.rD5w2wv6jj ++ cat /tmp/tmp.MOpmrwbXz8 ++ rm /tmp/tmp.rD5w2wv6jj /tmp/tmp.MOpmrwbXz8 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 13 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1F3SsZLxRo +++ mktemp ++ local LAST_ERR=/tmp/tmp.JwaCWOHzGH ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.1F3SsZLxRo ++ cat /tmp/tmp.JwaCWOHzGH ++ rm /tmp/tmp.1F3SsZLxRo /tmp/tmp.JwaCWOHzGH ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 14 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LMG8RUY2sh +++ mktemp ++ local LAST_ERR=/tmp/tmp.8JgDwUIPcz ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.LMG8RUY2sh ++ cat /tmp/tmp.8JgDwUIPcz ++ rm /tmp/tmp.LMG8RUY2sh /tmp/tmp.8JgDwUIPcz ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 15 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oLb4zJ2rzK +++ mktemp ++ local LAST_ERR=/tmp/tmp.ft2t3OFNUD ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.oLb4zJ2rzK ++ cat /tmp/tmp.ft2t3OFNUD ++ rm /tmp/tmp.oLb4zJ2rzK /tmp/tmp.ft2t3OFNUD ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo + compare_mongo_cmd find databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 + local command=find + local uri=databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 + local postfix= + local suffix= + local database=myApp + local collection=test + local sort= + local tls=false + local 'full_command=db.test.find()' + [[ -n '' ]] + log 'running db.test.find() in myApp' + set +o xtrace [2025-09-16T15:15:50+0000] running db.test.find() in myApp + [[ false == \t\r\u\e ]] + mongo_command=run_mongo + run_mongo 'use myApp\n db.test.find()' databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 mongodb '' + local 'command=use myApp\n db.test.find()' + local uri=databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 + local driver=mongodb + local suffix=.svc.cluster.local + egrep -v 'I NETWORK|W NETWORK|F NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:|Started a new thread for the timer service' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp + /usr/sbin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' egrep: warning: egrep is obsolescent; using grep -E ++ local LAST_OUT=/tmp/tmp.KYD4h41gr0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.mUKF66avUt ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.KYD4h41gr0 ++ cat /tmp/tmp.mUKF66avUt ++ rm /tmp/tmp.KYD4h41gr0 /tmp/tmp.mUKF66avUt ++ return 0 + local client_container=psmdb-client-66f577db5f-94kg8 + local mongo_flag= + [[ databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 == *cfg* ]] + replica_set=rs0 + kubectl_bin exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.xe5xm0SfRj ++ mktemp + local LAST_ERR=/tmp/tmp.TW6XDIDgI2 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.xe5xm0SfRj + cat /tmp/tmp.TW6XDIDgI2 + rm /tmp/tmp.xe5xm0SfRj /tmp/tmp.TW6XDIDgI2 + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/compare/find.json /tmp/tmp.GJOJy85nvw/find + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9A9ZC0FSiF +++ mktemp ++ local LAST_ERR=/tmp/tmp.M1aFgG3pey ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9A9ZC0FSiF ++ cat /tmp/tmp.M1aFgG3pey ++ rm /tmp/tmp.9A9ZC0FSiF /tmp/tmp.M1aFgG3pey ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo + delete_cluster some-name + local cluster=some-name + echo 'deleting cluster: some-name' deleting cluster: some-name + kubectl_bin delete psmdb some-name ++ mktemp + local LAST_OUT=/tmp/tmp.EBioY5FKZQ ++ mktemp + local LAST_ERR=/tmp/tmp.sPZFBW3f06 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete psmdb some-name + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.EBioY5FKZQ perconaservermongodb.psmdb.percona.com "some-name" deleted from replset-overrides-31757 namespace + cat /tmp/tmp.sPZFBW3f06 + rm /tmp/tmp.EBioY5FKZQ /tmp/tmp.sPZFBW3f06 + return 0 + wait_for_delete psmdb/some-name + local res=psmdb/some-name + local wait_time=60 + set +o xtrace waiting for psmdb/some-name to be deletedError from server (NotFound): perconaservermongodbs.psmdb.percona.com "some-name" not found Error from server (NotFound): perconaservermongodbs.psmdb.percona.com "some-name" not found Error from server (NotFound): perconaservermongodbs.psmdb.percona.com "some-name" not found Error from server (NotFound): perconaservermongodbs.psmdb.percona.com "some-name" not found + wait_for_delete pod/some-name-rs0-0 + local res=pod/some-name-rs0-0 + local wait_time=60 + set +o xtrace waiting for pod/some-name-rs0-0 to be deletedError from server (NotFound): pods "some-name-rs0-0" not found Error from server (NotFound): pods "some-name-rs0-0" not found Error from server (NotFound): pods "some-name-rs0-0" not found Error from server (NotFound): pods "some-name-rs0-0" not found + kubectl delete secrets --all secret "internal-some-name-users" deleted from replset-overrides-31757 namespace secret "minio-secret" deleted from replset-overrides-31757 namespace secret "minio-service" deleted from replset-overrides-31757 namespace secret "sh.helm.release.v1.minio-service.v1" deleted from replset-overrides-31757 namespace secret "some-name-mongodb-encryption-key" deleted from replset-overrides-31757 namespace secret "some-name-pbm-config" deleted from replset-overrides-31757 namespace secret "some-name-ssl" deleted from replset-overrides-31757 namespace secret "some-name-ssl-internal" deleted from replset-overrides-31757 namespace secret "some-users" deleted from replset-overrides-31757 namespace + desc 'Case 1: PASSED' + set +o xtrace ----------------------------------------------------------------------------------- Case 1: PASSED ----------------------------------------------------------------------------------- + desc 'Case 2: Patching a running cluster to override hostnames' + set +o xtrace ----------------------------------------------------------------------------------- Case 2: Patching a running cluster to override hostnames ----------------------------------------------------------------------------------- + test_override_host_after_deploy + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/conf/secrets_with_tls.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/conf/minio-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.bhnJWJmiOt ++ mktemp + local LAST_ERR=/tmp/tmp.JNueRJIghB + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/conf/secrets_with_tls.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/conf/minio-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.bhnJWJmiOt secret/some-users created secret/some-name-ssl created secret/some-name-ssl-internal created secret/minio-secret created + cat /tmp/tmp.JNueRJIghB + rm /tmp/tmp.bhnJWJmiOt /tmp/tmp.JNueRJIghB + return 0 + echo 'creating PSMDB cluster: some-name' creating PSMDB cluster: some-name + apply_cluster /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/conf/some-name.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/conf/some-name.yml + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/conf/some-name.yml + yq eval '(.spec | select(.image == null)).image = "perconalab/percona-server-mongodb-operator:main-mongod7.0"' + yq eval '(.spec | select(has("pmm"))).pmm.image = "percona/pmm-client:2.44.1-1"' ++ mktemp + yq eval '(.spec | select(has("initImage"))).initImage = "perconalab/percona-server-mongodb-operator:PR-2045-4cbffc79"' + yq eval '(.spec | select(has("backup"))).backup.image = "perconalab/percona-server-mongodb-operator:main-backup"' + yq eval '.spec.upgradeOptions.apply="Never"' + local LAST_OUT=/tmp/tmp.bIrZWk2l8p ++ mktemp + local LAST_ERR=/tmp/tmp.Qt3fakT55m + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.bIrZWk2l8p perconaservermongodb.psmdb.percona.com/some-name created + cat /tmp/tmp.Qt3fakT55m + rm /tmp/tmp.bIrZWk2l8p /tmp/tmp.Qt3fakT55m + return 0 + wait_for_running some-name-rs0 3 + local name=some-name-rs0 + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=rs0 + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-rs0-0 + local pod=some-name-rs0-0 + set +o xtrace waiting for pod/some-name-rs0-0 to be ready...............OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-rs0-1 + local pod=some-name-rs0-1 + set +o xtrace waiting for pod/some-name-rs0-1 to be ready......OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SDrMtUofGW +++ mktemp ++ local LAST_ERR=/tmp/tmp.cdBxaxkNhv ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.SDrMtUofGW ++ cat /tmp/tmp.cdBxaxkNhv ++ rm /tmp/tmp.SDrMtUofGW /tmp/tmp.cdBxaxkNhv ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-rs0-2 + local pod=some-name-rs0-2 + set +o xtrace waiting for pod/some-name-rs0-2 to be ready.........OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].non_voting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Z1qrbn0Lpj +++ mktemp ++ local LAST_ERR=/tmp/tmp.U9SLf9wqlP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].non_voting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Z1qrbn0Lpj ++ cat /tmp/tmp.U9SLf9wqlP ++ rm /tmp/tmp.Z1qrbn0Lpj /tmp/tmp.U9SLf9wqlP ++ return 0 + [[ '' == \t\r\u\e ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TmhULfDs13 +++ mktemp ++ local LAST_ERR=/tmp/tmp.YWbOHkrBmW ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.TmhULfDs13 ++ cat /tmp/tmp.YWbOHkrBmW ++ rm /tmp/tmp.TmhULfDs13 /tmp/tmp.YWbOHkrBmW ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness.. + sleep 10 + echo 'writing some data' writing some data + run_mongo 'use myApp\n db.test.insert({ x: 100500 })' databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757 + local 'command=use myApp\n db.test.insert({ x: 100500 })' + local uri=databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757 + local driver=mongodb+srv + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2LVewS5XpF +++ mktemp ++ local LAST_ERR=/tmp/tmp.zI86QCeZr3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.2LVewS5XpF ++ cat /tmp/tmp.zI86QCeZr3 ++ rm /tmp/tmp.2LVewS5XpF /tmp/tmp.zI86QCeZr3 ++ return 0 + local client_container=psmdb-client-66f577db5f-94kg8 + local mongo_flag= + [[ databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757 == *cfg* ]] + replica_set=rs0 + kubectl_bin exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb+srv://databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.5DGULZNAq7 ++ mktemp + local LAST_ERR=/tmp/tmp.bg76AKA5AU + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb+srv://databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.5DGULZNAq7 Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-rs0-1.some-name-rs0.replset-overrides-31757.svc.cluster.local:27017,some-name-rs0-2.some-name-rs0.replset-overrides-31757.svc.cluster.local:27017,some-name-rs0-0.some-name-rs0.replset-overrides-31757.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb&replicaSet=rs0&ssl=false Implicit session: session { "id" : UUID("25b1a730-c52c-43be-928e-dfd8248302ea") } Percona Server for MongoDB server version: v8.0.12-4 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.bg76AKA5AU + rm /tmp/tmp.5DGULZNAq7 /tmp/tmp.bg76AKA5AU + return 0 + compare_mongo_cmd find databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757 + local command=find + local uri=databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757 + local postfix= + local suffix= + local database=myApp + local collection=test + local sort= + local tls=false + local 'full_command=db.test.find()' + [[ -n '' ]] + log 'running db.test.find() in myApp' + set +o xtrace [2025-09-16T15:18:34+0000] running db.test.find() in myApp + [[ false == \t\r\u\e ]] + mongo_command=run_mongo + run_mongo 'use myApp\n db.test.find()' databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757 mongodb '' + local 'command=use myApp\n db.test.find()' + local uri=databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757 + local driver=mongodb + local suffix=.svc.cluster.local + egrep -v 'I NETWORK|W NETWORK|F NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:|Started a new thread for the timer service' + /usr/sbin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp egrep: warning: egrep is obsolescent; using grep -E ++ local LAST_OUT=/tmp/tmp.shcyG3oqec +++ mktemp ++ local LAST_ERR=/tmp/tmp.spMvzlP4G2 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.shcyG3oqec ++ cat /tmp/tmp.spMvzlP4G2 ++ rm /tmp/tmp.shcyG3oqec /tmp/tmp.spMvzlP4G2 ++ return 0 + local client_container=psmdb-client-66f577db5f-94kg8 + local mongo_flag= + [[ databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757 == *cfg* ]] + replica_set=rs0 + kubectl_bin exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.YxmPGroYyQ ++ mktemp + local LAST_ERR=/tmp/tmp.KeSQG3ZwVp + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.YxmPGroYyQ + cat /tmp/tmp.KeSQG3ZwVp + rm /tmp/tmp.YxmPGroYyQ /tmp/tmp.KeSQG3ZwVp + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/compare/find.json /tmp/tmp.GJOJy85nvw/find + run_backup minio backup-minio-logical + local storage=minio + local backup_name=backup-minio-logical + local type=logical + desc 'run backup backup-minio-logical' + set +o xtrace ----------------------------------------------------------------------------------- run backup backup-minio-logical ----------------------------------------------------------------------------------- + yq eval '.metadata.name = "backup-minio-logical" | .spec.storageName = "minio" | .spec.type = "logical"' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/conf/backup-minio.yml + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.mAjMnTXXVR ++ mktemp + local LAST_ERR=/tmp/tmp.QZjSpOqAUl + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.mAjMnTXXVR perconaservermongodbbackup.psmdb.percona.com/backup-minio-logical created + cat /tmp/tmp.QZjSpOqAUl + rm /tmp/tmp.mAjMnTXXVR /tmp/tmp.QZjSpOqAUl + return 0 + wait_backup backup-minio-logical + local backup_name=backup-minio-logical + local target_state=ready + set +o xtrace waiting for backup-minio-logical to reach ready state...... + echo 'creating external services' creating external services + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/conf/external-services.yml ++ mktemp + local LAST_OUT=/tmp/tmp.snOm8TuPbp ++ mktemp + local LAST_ERR=/tmp/tmp.8TIgQuzf7g + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/conf/external-services.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.snOm8TuPbp service/external-rs0-0 unchanged service/external-rs0-1 unchanged service/external-rs0-2 unchanged + cat /tmp/tmp.8TIgQuzf7g + rm /tmp/tmp.snOm8TuPbp /tmp/tmp.8TIgQuzf7g + return 0 + echo 'patching PSMDB cluster with replsetOverrides: some-name' patching PSMDB cluster with replsetOverrides: some-name + sed s/NAMESPACE/replset-overrides-31757/g /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/conf/some-name-overridden.yml + apply_cluster /tmp/tmp.GJOJy85nvw/some-name-overridden.yml + '[' -z '' ']' + cat_config /tmp/tmp.GJOJy85nvw/some-name-overridden.yml + kubectl_bin apply -f - + cat /tmp/tmp.GJOJy85nvw/some-name-overridden.yml + yq eval '(.spec | select(.image == null)).image = "perconalab/percona-server-mongodb-operator:main-mongod7.0"' ++ mktemp + yq eval '(.spec | select(has("pmm"))).pmm.image = "percona/pmm-client:2.44.1-1"' + yq eval '(.spec | select(has("initImage"))).initImage = "perconalab/percona-server-mongodb-operator:PR-2045-4cbffc79"' + yq eval '(.spec | select(has("backup"))).backup.image = "perconalab/percona-server-mongodb-operator:main-backup"' + yq eval '.spec.upgradeOptions.apply="Never"' + local LAST_OUT=/tmp/tmp.ngVvSiCNkv ++ mktemp + local LAST_ERR=/tmp/tmp.ENWHICdx3R + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ngVvSiCNkv perconaservermongodb.psmdb.percona.com/some-name configured + cat /tmp/tmp.ENWHICdx3R + rm /tmp/tmp.ngVvSiCNkv /tmp/tmp.ENWHICdx3R + return 0 + wait_for_cluster_state some-name ready + local cluster_name=some-name + local target_state=ready + echo -n 'Waiting for cluster to reach ready state' Waiting for cluster to reach ready state+ local timeout=0 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.q5pjZSomRx +++ mktemp ++ local LAST_ERR=/tmp/tmp.LeXL4htdda ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.q5pjZSomRx ++ cat /tmp/tmp.LeXL4htdda ++ rm /tmp/tmp.q5pjZSomRx /tmp/tmp.LeXL4htdda ++ return 0 + [[ ready == ready ]] + echo + echo 'checking if we can read existing data' checking if we can read existing data + compare_mongo_cmd find databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 + local command=find + local uri=databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 + local postfix= + local suffix= + local database=myApp + local collection=test + local sort= + local tls=false + local 'full_command=db.test.find()' + [[ -n '' ]] + log 'running db.test.find() in myApp' + set +o xtrace [2025-09-16T15:18:56+0000] running db.test.find() in myApp + [[ false == \t\r\u\e ]] + mongo_command=run_mongo + run_mongo 'use myApp\n db.test.find()' databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 mongodb '' + local 'command=use myApp\n db.test.find()' + local uri=databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 + local driver=mongodb + local suffix=.svc.cluster.local + egrep -v 'I NETWORK|W NETWORK|F NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:|Started a new thread for the timer service' + /usr/sbin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.U1c3kzhqjO egrep: warning: egrep is obsolescent; using grep -E +++ mktemp ++ local LAST_ERR=/tmp/tmp.HFPmQP2Qem ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.U1c3kzhqjO ++ cat /tmp/tmp.HFPmQP2Qem ++ rm /tmp/tmp.U1c3kzhqjO /tmp/tmp.HFPmQP2Qem ++ return 0 + local client_container=psmdb-client-66f577db5f-94kg8 + local mongo_flag= + [[ databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 == *cfg* ]] + replica_set=rs0 + kubectl_bin exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.mR5Q39REbr ++ mktemp + local LAST_ERR=/tmp/tmp.0pvRHrvkQN + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.mR5Q39REbr + cat /tmp/tmp.0pvRHrvkQN + rm /tmp/tmp.mR5Q39REbr /tmp/tmp.0pvRHrvkQN + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/compare/find.json /tmp/tmp.GJOJy85nvw/find + run_recovery_check some-name backup-minio-logical external-rs0-0.replset-overrides-31757 + local cluster=some-name + local backup=backup-minio-logical + local host=external-rs0-0.replset-overrides-31757 + echo 'write data after backup' write data after backup + run_mongo 'use myApp\n db.test.insert({ x: 100501 })' databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 + local 'command=use myApp\n db.test.insert({ x: 100501 })' + local uri=databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 + local driver=mongodb+srv + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.B7uNU38p29 +++ mktemp ++ local LAST_ERR=/tmp/tmp.TMvgWYrCle ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.B7uNU38p29 ++ cat /tmp/tmp.TMvgWYrCle ++ rm /tmp/tmp.B7uNU38p29 /tmp/tmp.TMvgWYrCle ++ return 0 + local client_container=psmdb-client-66f577db5f-94kg8 + local mongo_flag= + [[ databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 == *cfg* ]] + replica_set=rs0 + kubectl_bin exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb+srv://databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.U3YonQgKl3 ++ mktemp + local LAST_ERR=/tmp/tmp.bN89iG9Mh8 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb+srv://databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.U3YonQgKl3 Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://external-rs0-0.replset-overrides-31757.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb&replicaSet=rs0&ssl=false Implicit session: session { "id" : UUID("88097b0d-ffa8-4104-a4b2-b998f6b90ffd") } Percona Server for MongoDB server version: v8.0.12-4 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.bN89iG9Mh8 + rm /tmp/tmp.U3YonQgKl3 /tmp/tmp.bN89iG9Mh8 + return 0 + echo 'restore to backup backup-minio-logical' restore to backup backup-minio-logical + compare_mongo_cmd find databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 -2nd + local command=find + local uri=databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 + local postfix=-2nd + local suffix= + local database=myApp + local collection=test + local sort= + local tls=false + local 'full_command=db.test.find()' + [[ -n '' ]] + log 'running db.test.find() in myApp' + set +o xtrace [2025-09-16T15:19:01+0000] running db.test.find() in myApp + [[ false == \t\r\u\e ]] + mongo_command=run_mongo + run_mongo 'use myApp\n db.test.find()' databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 mongodb '' + egrep -v 'I NETWORK|W NETWORK|F NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:|Started a new thread for the timer service' + local 'command=use myApp\n db.test.find()' + local uri=databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 + local driver=mongodb + local suffix=.svc.cluster.local + /usr/sbin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp egrep: warning: egrep is obsolescent; using grep -E ++ local LAST_OUT=/tmp/tmp.BfHssyS78N +++ mktemp ++ local LAST_ERR=/tmp/tmp.Z4KiXfOxjp ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.BfHssyS78N ++ cat /tmp/tmp.Z4KiXfOxjp ++ rm /tmp/tmp.BfHssyS78N /tmp/tmp.Z4KiXfOxjp ++ return 0 + local client_container=psmdb-client-66f577db5f-94kg8 + local mongo_flag= + [[ databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 == *cfg* ]] + replica_set=rs0 + kubectl_bin exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.I6rVwFIWub ++ mktemp + local LAST_ERR=/tmp/tmp.AsAIOydNOm + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.I6rVwFIWub + cat /tmp/tmp.AsAIOydNOm + rm /tmp/tmp.I6rVwFIWub /tmp/tmp.AsAIOydNOm + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/compare/find-2nd.json /tmp/tmp.GJOJy85nvw/find-2nd + run_restore backup-minio-logical + local backup_name=backup-minio-logical + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/conf/restore.yml + /usr/sbin/sed -e 's/name:/name: restore-backup-minio-logical/' + /usr/sbin/sed -e 's/backupName:/backupName: backup-minio-logical/' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.jdQBprfZ8t ++ mktemp + local LAST_ERR=/tmp/tmp.jm3lzriwQV + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.jdQBprfZ8t perconaservermongodbrestore.psmdb.percona.com/restore-backup-minio-logical created + cat /tmp/tmp.jm3lzriwQV + rm /tmp/tmp.jdQBprfZ8t /tmp/tmp.jm3lzriwQV + return 0 + wait_restore backup-minio-logical some-name + local backup_name=backup-minio-logical + local cluster_name=some-name + local target_state=ready + local wait_cluster_consistency=1 + local wait_time=1780 + local ok_if_ready=0 + set +o xtrace Waiting for the psmdb-restore/restore-backup-minio-logical object to be createdOK Waiting psmdb-restore/restore-backup-minio-logical to reach state "ready" OK after 0 minutes + [[ 1 -eq 1 ]] + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7CNyN2Mm99 +++ mktemp ++ local LAST_ERR=/tmp/tmp.6P8nExupOG ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.7CNyN2Mm99 ++ cat /tmp/tmp.6P8nExupOG ++ rm /tmp/tmp.7CNyN2Mm99 /tmp/tmp.6P8nExupOG ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo + compare_mongo_cmd find databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 + local command=find + local uri=databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 + local postfix= + local suffix= + local database=myApp + local collection=test + local sort= + local tls=false + local 'full_command=db.test.find()' + [[ -n '' ]] + log 'running db.test.find() in myApp' + set +o xtrace [2025-09-16T15:19:25+0000] running db.test.find() in myApp + [[ false == \t\r\u\e ]] + mongo_command=run_mongo + run_mongo 'use myApp\n db.test.find()' databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 mongodb '' + local 'command=use myApp\n db.test.find()' + local uri=databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 + local driver=mongodb + local suffix=.svc.cluster.local + egrep -v 'I NETWORK|W NETWORK|F NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:|Started a new thread for the timer service' + /usr/sbin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp egrep: warning: egrep is obsolescent; using grep -E ++ local LAST_OUT=/tmp/tmp.j5BFccLJmr +++ mktemp ++ local LAST_ERR=/tmp/tmp.gZ93Pk4WVq ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.j5BFccLJmr ++ cat /tmp/tmp.gZ93Pk4WVq ++ rm /tmp/tmp.j5BFccLJmr /tmp/tmp.gZ93Pk4WVq ++ return 0 + local client_container=psmdb-client-66f577db5f-94kg8 + local mongo_flag= + [[ databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757 == *cfg* ]] + replica_set=rs0 + kubectl_bin exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.RqYsXyWzF7 ++ mktemp + local LAST_ERR=/tmp/tmp.Dwdunth3eu + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://databaseAdmin:databaseAdmin123456@external-rs0-0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.RqYsXyWzF7 + cat /tmp/tmp.Dwdunth3eu + rm /tmp/tmp.RqYsXyWzF7 /tmp/tmp.Dwdunth3eu + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/compare/find.json /tmp/tmp.GJOJy85nvw/find + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7fM78CrSrL +++ mktemp ++ local LAST_ERR=/tmp/tmp.ePyICn9IaO ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.7fM78CrSrL ++ cat /tmp/tmp.ePyICn9IaO ++ rm /tmp/tmp.7fM78CrSrL /tmp/tmp.ePyICn9IaO ++ return 0 + [[ ready == \r\e\a\d\y ]] + echo + delete_cluster some-name + local cluster=some-name + echo 'deleting cluster: some-name' deleting cluster: some-name + kubectl_bin delete psmdb some-name ++ mktemp + local LAST_OUT=/tmp/tmp.BDAi9HaUBM ++ mktemp + local LAST_ERR=/tmp/tmp.OrB0KkH1BN + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete psmdb some-name + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.BDAi9HaUBM perconaservermongodb.psmdb.percona.com "some-name" deleted from replset-overrides-31757 namespace + cat /tmp/tmp.OrB0KkH1BN + rm /tmp/tmp.BDAi9HaUBM /tmp/tmp.OrB0KkH1BN + return 0 + wait_for_delete psmdb/some-name + local res=psmdb/some-name + local wait_time=60 + set +o xtrace waiting for psmdb/some-name to be deletedError from server (NotFound): perconaservermongodbs.psmdb.percona.com "some-name" not found Error from server (NotFound): perconaservermongodbs.psmdb.percona.com "some-name" not found Error from server (NotFound): perconaservermongodbs.psmdb.percona.com "some-name" not found Error from server (NotFound): perconaservermongodbs.psmdb.percona.com "some-name" not found + wait_for_delete pod/some-name-rs0-0 + local res=pod/some-name-rs0-0 + local wait_time=60 + set +o xtrace waiting for pod/some-name-rs0-0 to be deletedError from server (NotFound): pods "some-name-rs0-0" not found Error from server (NotFound): pods "some-name-rs0-0" not found Error from server (NotFound): pods "some-name-rs0-0" not found Error from server (NotFound): pods "some-name-rs0-0" not found + kubectl delete secrets --all secret "internal-some-name-users" deleted from replset-overrides-31757 namespace secret "minio-secret" deleted from replset-overrides-31757 namespace secret "some-name-mongodb-encryption-key" deleted from replset-overrides-31757 namespace secret "some-name-ssl" deleted from replset-overrides-31757 namespace secret "some-name-ssl-internal" deleted from replset-overrides-31757 namespace secret "some-users" deleted from replset-overrides-31757 namespace + desc 'Case 2: PASSED' + set +o xtrace ----------------------------------------------------------------------------------- Case 2: PASSED ----------------------------------------------------------------------------------- + desc 'Case 3: Overriding member priorities' + set +o xtrace ----------------------------------------------------------------------------------- Case 3: Overriding member priorities ----------------------------------------------------------------------------------- + test_override_priority + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/conf/secrets_with_tls.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/conf/minio-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.kbHGFOGXDg ++ mktemp + local LAST_ERR=/tmp/tmp.XriKW343gL + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/conf/secrets_with_tls.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/conf/minio-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.kbHGFOGXDg secret/some-users created secret/some-name-ssl created secret/some-name-ssl-internal created secret/minio-secret created + cat /tmp/tmp.XriKW343gL + rm /tmp/tmp.kbHGFOGXDg /tmp/tmp.XriKW343gL + return 0 + echo 'creating PSMDB cluster: some-name' creating PSMDB cluster: some-name + apply_cluster /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/conf/some-name-override-priority.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/conf/some-name-override-priority.yml + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/conf/some-name-override-priority.yml + yq eval '(.spec | select(.image == null)).image = "perconalab/percona-server-mongodb-operator:main-mongod7.0"' ++ mktemp + yq eval '(.spec | select(has("backup"))).backup.image = "perconalab/percona-server-mongodb-operator:main-backup"' + yq eval '(.spec | select(has("pmm"))).pmm.image = "percona/pmm-client:2.44.1-1"' + yq eval '(.spec | select(has("initImage"))).initImage = "perconalab/percona-server-mongodb-operator:PR-2045-4cbffc79"' + local LAST_OUT=/tmp/tmp.rpcbiCCNAF ++ mktemp + yq eval '.spec.upgradeOptions.apply="Never"' + local LAST_ERR=/tmp/tmp.1rZk6dG6P9 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.rpcbiCCNAF perconaservermongodb.psmdb.percona.com/some-name created + cat /tmp/tmp.1rZk6dG6P9 + rm /tmp/tmp.rpcbiCCNAF /tmp/tmp.1rZk6dG6P9 + return 0 + wait_for_running some-name-rs0 3 + local name=some-name-rs0 + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=rs0 + local cluster_name=some-name ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod some-name-rs0-0 + local pod=some-name-rs0-0 + set +o xtrace waiting for pod/some-name-rs0-0 to be ready.........OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod some-name-rs0-1 + local pod=some-name-rs0-1 + set +o xtrace waiting for pod/some-name-rs0-1 to be ready.....OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.07gWlIaU4x +++ mktemp ++ local LAST_ERR=/tmp/tmp.blTD6dkPMb ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.07gWlIaU4x ++ cat /tmp/tmp.blTD6dkPMb ++ rm /tmp/tmp.07gWlIaU4x /tmp/tmp.blTD6dkPMb ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-rs0-2 + local pod=some-name-rs0-2 + set +o xtrace waiting for pod/some-name-rs0-2 to be ready........OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].non_voting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.a1gQNCS134 +++ mktemp ++ local LAST_ERR=/tmp/tmp.B01Ul19nlm ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].non_voting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.a1gQNCS134 ++ cat /tmp/tmp.B01Ul19nlm ++ rm /tmp/tmp.a1gQNCS134 /tmp/tmp.B01Ul19nlm ++ return 0 + [[ '' == \t\r\u\e ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kzTxmNMylK +++ mktemp ++ local LAST_ERR=/tmp/tmp.E0AsoyE754 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.kzTxmNMylK ++ cat /tmp/tmp.E0AsoyE754 ++ rm /tmp/tmp.kzTxmNMylK /tmp/tmp.E0AsoyE754 ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness... + echo 'writing some data' writing some data + run_mongo 'use myApp\n db.test.insert({ x: 100500 })' databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757 + local 'command=use myApp\n db.test.insert({ x: 100500 })' + local uri=databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757 + local driver=mongodb+srv + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Bsd8iGId59 +++ mktemp ++ local LAST_ERR=/tmp/tmp.7eIV6dSBS7 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Bsd8iGId59 ++ cat /tmp/tmp.7eIV6dSBS7 ++ rm /tmp/tmp.Bsd8iGId59 /tmp/tmp.7eIV6dSBS7 ++ return 0 + local client_container=psmdb-client-66f577db5f-94kg8 + local mongo_flag= + [[ databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757 == *cfg* ]] + replica_set=rs0 + kubectl_bin exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb+srv://databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.Q5rqDgxgor ++ mktemp + local LAST_ERR=/tmp/tmp.s6Rj99KEWr + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb+srv://databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Q5rqDgxgor Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-rs0-0.some-name-rs0.replset-overrides-31757.svc.cluster.local:27017,some-name-rs0-2.some-name-rs0.replset-overrides-31757.svc.cluster.local:27017,some-name-rs0-1.some-name-rs0.replset-overrides-31757.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb&replicaSet=rs0&ssl=false Implicit session: session { "id" : UUID("23a4b0d6-b094-49ab-bf19-a9a7d03e40d6") } Percona Server for MongoDB server version: v8.0.12-4 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.s6Rj99KEWr + rm /tmp/tmp.Q5rqDgxgor /tmp/tmp.s6Rj99KEWr + return 0 + compare_mongo_cmd find databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757 + local command=find + local uri=databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757 + local postfix= + local suffix= + local database=myApp + local collection=test + local sort= + local tls=false + local 'full_command=db.test.find()' + [[ -n '' ]] + log 'running db.test.find() in myApp' + set +o xtrace [2025-09-16T15:21:46+0000] running db.test.find() in myApp + [[ false == \t\r\u\e ]] + mongo_command=run_mongo + run_mongo 'use myApp\n db.test.find()' databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757 mongodb '' + local 'command=use myApp\n db.test.find()' + local uri=databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757 + local driver=mongodb + local suffix=.svc.cluster.local + egrep -v 'I NETWORK|W NETWORK|F NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:|Started a new thread for the timer service' + /usr/sbin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sXPseCKWvs +++ mktemp egrep: warning: egrep is obsolescent; using grep -E ++ local LAST_ERR=/tmp/tmp.BVGNbhu8U3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.sXPseCKWvs ++ cat /tmp/tmp.BVGNbhu8U3 ++ rm /tmp/tmp.sXPseCKWvs /tmp/tmp.BVGNbhu8U3 ++ return 0 + local client_container=psmdb-client-66f577db5f-94kg8 + local mongo_flag= + [[ databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757 == *cfg* ]] + replica_set=rs0 + kubectl_bin exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.lJjWj6yisK ++ mktemp + local LAST_ERR=/tmp/tmp.ctIE0LJ3nI + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.lJjWj6yisK + cat /tmp/tmp.ctIE0LJ3nI + rm /tmp/tmp.lJjWj6yisK /tmp/tmp.ctIE0LJ3nI + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/compare/find.json /tmp/tmp.GJOJy85nvw/find + echo 'checking member priorities' checking member priorities + run_mongo 'rs.conf().members.map(m => m.priority)' databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757 + local 'command=rs.conf().members.map(m => m.priority)' + local uri=databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757 + local driver=mongodb+srv + local suffix=.svc.cluster.local + egrep -v 'I NETWORK|W NETWORK|F NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp egrep: warning: egrep is obsolescent; using grep -E ++ local LAST_OUT=/tmp/tmp.trUS7c0k8U +++ mktemp ++ local LAST_ERR=/tmp/tmp.bt8FGPLHmr ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.trUS7c0k8U ++ cat /tmp/tmp.bt8FGPLHmr ++ rm /tmp/tmp.trUS7c0k8U /tmp/tmp.bt8FGPLHmr ++ return 0 + local client_container=psmdb-client-66f577db5f-94kg8 + local mongo_flag= + [[ databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757 == *cfg* ]] + replica_set=rs0 + kubectl_bin exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''rs.conf().members.map(m => m.priority)\n'\'' | mongo mongodb+srv://databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' ++ mktemp + local LAST_OUT=/tmp/tmp.UFToqcRDbM ++ mktemp + local LAST_ERR=/tmp/tmp.DCNDzLrkD0 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-66f577db5f-94kg8 -- bash -c 'printf '\''rs.conf().members.map(m => m.priority)\n'\'' | mongo mongodb+srv://databaseAdmin:databaseAdmin123456@some-name-rs0.replset-overrides-31757.svc.cluster.local/admin?ssl=false\&replicaSet=rs0 ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.UFToqcRDbM + cat /tmp/tmp.DCNDzLrkD0 + rm /tmp/tmp.UFToqcRDbM /tmp/tmp.DCNDzLrkD0 + return 0 + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2045/e2e-tests/replset-overrides/compare/priorities.json /tmp/tmp.GJOJy85nvw/priorities.json + echo 'member priorities are OK' member priorities are OK + delete_cluster some-name + local cluster=some-name + echo 'deleting cluster: some-name' deleting cluster: some-name + kubectl_bin delete psmdb some-name ++ mktemp + local LAST_OUT=/tmp/tmp.WllkVqq0VS ++ mktemp + local LAST_ERR=/tmp/tmp.B7ok7rvaYn + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete psmdb some-name + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.WllkVqq0VS perconaservermongodb.psmdb.percona.com "some-name" deleted from replset-overrides-31757 namespace + cat /tmp/tmp.B7ok7rvaYn + rm /tmp/tmp.WllkVqq0VS /tmp/tmp.B7ok7rvaYn + return 0 + wait_for_delete psmdb/some-name + local res=psmdb/some-name + local wait_time=60 + set +o xtrace waiting for psmdb/some-name to be deletedError from server (NotFound): perconaservermongodbs.psmdb.percona.com "some-name" not found Error from server (NotFound): perconaservermongodbs.psmdb.percona.com "some-name" not found Error from server (NotFound): perconaservermongodbs.psmdb.percona.com "some-name" not found Error from server (NotFound): perconaservermongodbs.psmdb.percona.com "some-name" not found + wait_for_delete pod/some-name-rs0-0 + local res=pod/some-name-rs0-0 + local wait_time=60 + set +o xtrace waiting for pod/some-name-rs0-0 to be deletedError from server (NotFound): pods "some-name-rs0-0" not found Error from server (NotFound): pods "some-name-rs0-0" not found Error from server (NotFound): pods "some-name-rs0-0" not found Error from server (NotFound): pods "some-name-rs0-0" not found + kubectl delete secrets --all secret "internal-some-name-users" deleted from replset-overrides-31757 namespace secret "minio-secret" deleted from replset-overrides-31757 namespace secret "some-name-mongodb-encryption-key" deleted from replset-overrides-31757 namespace secret "some-name-ssl" deleted from replset-overrides-31757 namespace secret "some-name-ssl-internal" deleted from replset-overrides-31757 namespace secret "some-users" deleted from replset-overrides-31757 namespace + desc 'Case 3: PASSED' + set +o xtrace ----------------------------------------------------------------------------------- Case 3: PASSED ----------------------------------------------------------------------------------- + desc 'All cases PASSED' + set +o xtrace ----------------------------------------------------------------------------------- All cases PASSED -----------------------------------------------------------------------------------